Figures
Abstract
Environmental challenges are rarely confined to national, disciplinary, or linguistic domains. Convergent solutions require international collaboration and equitable access to new technologies and practices. The ability of international, multidisciplinary and multilingual research teams to work effectively can be challenging. A major impediment to innovation in diverse teams often stems from different understandings of the terminology used. These can vary greatly according to the cultural and disciplinary backgrounds of the team members. In this paper we take an empirical approach to examine sources of terminological confusion and their effect in a technically innovative, multidisciplinary, multinational, and multilingual research project, adhering to Open Science principles. We use guided reflection of participant experience in two contrasting teams—one applying Deep Learning (Artificial Intelligence) techniques, the other developing guidance for Open Science practices—to identify and classify the terminological obstacles encountered and reflect on their impact. Several types of terminological incongruities were identified, including fuzziness in language, disciplinary differences and multiple terms for a single meaning. A novel or technical term did not always exist in all domains, or if known, was not fully understood or adopted. Practical matters of international data collection and comparison included an unanticipated need to incorporate different types of data labels from country to country, authority to authority. Sometimes these incongruities could be solved quickly, sometimes they stopped the workflow. Active collaboration and mutual trust across the team enhanced workflows, as incompatibilities were resolved more speedily than otherwise. Based on the research experience described in this paper, we make six recommendations accompanied by suggestions for their implementation to improve the success of similar multinational, multilingual and multidisciplinary projects. These recommendations are conceptual drawing on a singular experience and remain to be sources for discussion and testing by others embarking on their research journey.
Citation: Specht A, Stall S, Machicao J, Catry T, Chaumont M, David R, et al. (2024) Managing linguistic obstacles in multidisciplinary, multinational, and multilingual research projects. PLoS ONE 19(12): e0311967. https://doi.org/10.1371/journal.pone.0311967
Editor: Muhammad Afzaal, Shanghai International Studies University - Songjiang Campus, CHINA
Received: May 7, 2024; Accepted: September 27, 2024; Published: December 5, 2024
Copyright: © 2024 Specht et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All data are in the manuscript. There are no additional data.
Funding: This project was conducted as part of the PARSEC project, funded under a Belmont Forum Collaborative Research Action (CRA) on Science-Driven e-Infrastructures Innovation (SEI), with funding from the French Agence Nationale de Recherche (grant number ANR-18-BELM-0002), the Japan Science and Technology Agency (grant number 21-191029671), the National Science Foundation (NSF) of the USA (grant number 1929464), and the São Paulo Research Foundation (FAPESP: grant number 2018/24017–3), with support from the synthesis centre CESAB of the French Foundation for Research on Biodiversity (FRB). J.M. is grateful for support from FAPESP (grant number 2020/03514–9).
Competing interests: The authors have declared that no competing interests exist.
1 Introduction
Environmental challenges such as biodiversity decline, climate change, and viral pandemics rarely stop at national, disciplinary, or linguistic borders. Instead they usually demand international collaboration to find convergent solutions [1, 2]. To discover such solutions it is necessary to bring together experts with different, but pertinent, disciplines and skill sets, and to enable access to and sharing of data and information across geographic and cultural boundaries. These requirements were advocated in the UNESCO Recommendation on Open Science in 2021 [3], which also emphasised the need to ensure that multilingual scientific knowledge is openly available, accessible, and reusable. International collaborative research benefits the researcher as well as being for the common good [4], but the lack of a common frame of reference can greatly impede collaboration and often requires the development of new frameworks based on a common language or ontology [5, 6].
Groups of experts are commonly brought together (or come together of their own volition) to pool their expertise and knowledge to generate innovative, novel insights and potentially to achieve solutions [7]. Such diverse, often multidisciplinary, research teams are usually created in response to a problem: members are deliberately chosen with relevant skills and expertise to contribute to a solution, while keeping in mind other factors, such as organisation, country, and life situation [8]. Teamwork of any sort, however, is not simple; effective teamwork does not magically happen and can be impeded by lack of communication among parties for several reasons. These reasons, if identified, can be negotiated and their effects reduced. If not, they can be a breakpoint for the group. The physical separation of group members creates logistic problems, and if collaboration is remote only, time zone challenges can be prohibitive [9, 10].
The information generated to analyse and derive solutions is still siloed in different languages and locations throughout the world [11, 12]. Although English is the most commonly used international scientific language, and having such a language certainly facilitates communication [13], much information of value in biodiversity conservation, for example, is published in other languages and consequently overlooked [13–16]. Taking into consideration the variety of terms and concepts across languages and cultures has been argued to lead to better conservation decisions [16, 17]. Equally, not conveying results in a multiplicity of languages can impede efforts to conserve and protect a species. A lack of recognition of the semantic diversity involved within and across languages undoubtedly affects practice [12–18]. As a means to mitigate such knowledge gaps, it has been suggested that researchers and the scholarly literature should aim to be linguistically inclusive by, for example, providing non-English language abstracts [19, 20] and improving translation tools [11]. Good translations, of course, occur at the concept level, not as a simple word-for-word translation, a limitation of most on-line translation tools [21]. Fundamentally, language is not a precise tool that can be used to accurately express all concepts, but rather a collection of words and phrases that can be used to refer to a variety of different concepts, which can lead to confusion and misinterpretation [22]. Words and phrases have potentially different meanings in the minds of the communicator and the receiver, and their interpretation will depend on the context in which they are expressed. Paying attention to, and asking questions about, the words people use can point to repeated patterns, or touchstones, as well as sticking points that shape collaborations [23].
As multidisciplinarity is becoming the norm in research projects, its emergence is conditioned by the willingness to overcome disciplinary and linguistic barriers. A desire to learn from other fields of expertise is key to shared (convergent) data-driven ways of working, and transformative practices [24, 25]. Communication among team members, and understanding the different practices of each member, is modified by the prism through which each views the world. This can be affected by nationality, disciplinary background or sphere of expertise within that discipline. For effective multinational and multidisciplinary collaboration to occur, each party needs to have some cognisance and understanding of the different ’languages’ used by their fellow researchers, and they often need to engage in active co-interpretation.
2 Potential terminological challenges
Researchers from different domains often have their own discipline-specific terminology, consider different types of data more important than others, or deem different kinds of analyses as being more valid [26]. Such barriers can be syntactic (different formats of information), semantic (different groups assign different meanings to information), or pragmatic (different groups have different practices or interests) [27]. Disciplinarily defined language often resides exclusively within a community and is disconnected from upper-level knowledge. If semantic inconsistencies exist but are poorly identified or unknown, they can lead to misunderstanding and misuse, and ultimately impair data management itself [10]. High levels of disciplinary diversity can reduce the functionality of a research group [8, 28], especially if there is significant cognitive distance between the disciplines and insufficient time or willingness among group members to bridge the gaps [24, 29].
As Earth observation data and products increase in availability, it is common for researchers from the computer sciences (data, model sciences or Artificial Intelligence ‐ AI) to use these products in their work. Without appropriate expertise, however, the input data and the interpretation of any analysis can be challenging. The opposite pattern can also occur as more and more computer models are made available and used by people lacking expertise in AI. Modelling in multidisciplinary space requires input from all participants to ensure model validity, relevance, transparency and acceptability [30]. The rewards of overcoming disciplinary boundaries are well recognised with enhanced intellectual stimulation and novel outcomes occurring when interdisciplinary teams function well [31–33].
It has been found that increasing the number of countries represented in a research group results in the degree of innovation in the group being reduced [8]. Participants from different countries with different languages and ways of working can have communication and cultural challenges that take time to overcome [8, 34]. An often overlooked component resulting in difficulties in international collaborations is the different terminology used for in-country administrative units such as provinces versus states, and villages versus towns. Cross-country comparisons depend on the quality of aggregation which can affect the granularity and accuracy of information [35].
Globalisation of research requires interoperability of observations and experimentation systems [22, 36]. Programming and data languages present a challenge, especially for machine discoverability. Developers are often experts in only one programming language (e.g. Python, C++, or R) and it can be challenging to use code that has been developed by another researcher, even in a known language. It is even more difficult to use code developed in an unfamiliar programming language. Increasingly researchers are using (with attribution) code that is openly available and converting it to a more familiar programming language for the purpose of reproducing experiments [37]. The diversity of programming languages, programming styles, and frameworks, even in a familiar computer language, can be very limiting and is certainly opaque to non-experts.
To improve sharing and interoperability of data, software and workflows across research teams, common FAIR (Findable, Accessible, Interoperable and Reusable) vocabularies that are both human and machine-readable have been proposed [38, 39]. These are intended to enable data interoperability and meta-analyses, even when data or software have different origins and are based on multiple vocabularies (as per FAIR Principle I2) [38].
3 Rationale
Learning from experience (experiential learning) is a fundamental way teams improve their practice. Observation and reflection on the root causes of problems and dysfunctional workflows can result in adjustment of practice, and new implications for action can be drawn [40, 41]. Individuals in software development teams learn during the work by reflecting on the process (reflection in action), but also after the work is either fully or partially completed (reflection on action) [42]. Continual adjustment of practice in response to internal and external feedback is fundamental to Agile Project Management, a well-accepted method of experience-based modification of practices in the software community [43]. Too often, however, work practices are adjusted ‘on the run’ and little is learnt from the experience for future reference [42]. The experiential learning process for a research team is illustrated in Fig 1. This is in contrast to the pattern for industry, for example, where the process becomes a repeated cycle, in each case the next iteration benefitting from the learning of its predecessor. Research projects usually finish when the funding runs out, so it remains for others to take advantage of the learning gained.
In this paper we use participant observation and reflection of their research experience as our research strategy [24, 40, 46]. We use a case study strategy, ideally suited to in-depth exploration of phenomena and for constructions of theory in its real-life (natural) context [4, 46, 47]. Our exploratory case study (sensu [48]) consists of two workflows within a multi-disciplinary and multinational research project. The research questions we posed are: (a) what are the terminological obstacles encountered, (b) are mismatches in language concepts due to disciplinary differences, native language, or country-specific nomenclature, and (c) what is the impact of these obstacles on the two workflows. From this structured experiential learning we propose ways to anticipate and overcome similar terminological stumbling blocks for future testing.
4 Case study
We examine the experience of two teams within a technically innovative, multidisciplinary, multinational, and multilingual research project aiming to be an exemplar of Open Science practices, the PARSEC project, ‘Building new tools for data sharing and reuse through a transnational investigation of the socioeconomic impacts of protected areas’ [49, 50]. PARSEC was an international Belmont Forum-awarded research project under a Coordinated Research Action on Science-driven e-Infrastructure Innovation. The 38 team members and six associate members were drawn from Brazil, France, Japan, the United States of America, and Australia. Two postdoctoral researchers and several short-term researchers were funded through the project. PARSEC ran from 2019 until 2024—notably encompassing the Covid19 travel restrictions—and was divided into two interrelated components: one pursuing a scientific objective, the other pursuing e-infrastructure innovation according to Open Science principles.
The core scientific investigation (carried out by 25 ’Synthesis Science Strand’ members) examined the socioeconomic effects of creating marine and terrestrial protected areas. This approach required combining and analysing existing remote sensing data with existing socioeconomic data using AI and other tools. The goal was to develop robust methodologies capable of estimating trends in socioeconomic indicators from remote sensing imagery. The 13 ’Data Science Strand’ members were charged with developing recommendations for data and code curation through interaction with the Synthesis Science Strand team and the wider research community. The goal was to remove barriers to data reuse by promoting best practices throughout the research data life cycle compliant with Open Science principles.
The disciplinary range in PARSEC was considerable, with specialists in data science, research data management, remote sensing data analysis, AI, machine learning, Deep Learning (DL), spatial systems, socioeconomics, wildlife biology, and ecology. The majority of team members had experience working with a variety of data types, from observational, spatial, remotely sensed to socioeconomic data. Most team members had experience working on more than one type of data. Half of the team members at the beginning of the project had limited experience in Open Science data practices [51].
The work of the Synthesis Science Strand started with an assessment of the use of Convolutional Neural Networks (CNN) to estimate poverty in a sample of east African villages [52], and an exploration of using Google Street View to detect socioeconomic conditions in an area of Brazil using DL methodology [53]. Code developed for the project has been made openly available [54, 55]. The Data Science Strand started work with a general introduction to the FAIR Principles and the steps required to account for them throughout a project [56] and examining best practices for reproducibility of DL experiments [37]. The Data Science Strand established various tracksheets for data and software use and for outputs and a project ’dictionary’ that was shared in a common on-line workspace.
Challenges due to the international and multidisciplinary nature of the work occurred (a) in the development and application of the DL model, and (b) in the preparation of guidelines for our multilingual researchers. The experience of the teams facing these two challenges are the focus of this paper and they will be referred to as the ’DL’ and ’Checklist’ teams.
4.1 DL team
The DL team was dedicated to leveraging advanced machine learning techniques, particularly DL models, for the analysis of remote sensing imagery in Brazil, Japan, and several countries in east Africa. The six DL team members were situated across France, Brazil, and North Africa. None of the DL team members were native English speakers, but all were computer scientists or engineers with high-level expertise in the use of AI to elucidate patterns across different types of data. The team used the English language to collaborate. Five of the members are authors of this paper (MC, PPC, RJ, JM, GS).
At the start of the project there were various meetings and workshops across the whole PARSEC team to understand the main goals. The most important task was to collect data so that a first data exploration could be conducted. Given the complexity of DL techniques, a generic workflow was adopted by the DL team (Fig 2). The stages of the workflow were:
- Data collection in which remote sensing and socioeconomic data are gathered for training and validation;
- Pre-processing and cross-validation tasks, such as data cleaning, integrating poverty indicators, and splitting the data for validation;
- DL model experiments tailored to our specific objectives; and
- Evaluation to assess the performance of the DL models, ensuring robustness and efficacy in analysing remote sensing data for socioeconomic estimations.
The general components are in bold while the steps of the DL workflow are shown in the boxes (adapted from [37]).
4.2 Checklist team
The Checklist team was composed of a PARSEC Data Strand team member from each funding country (Brazil, France, Japan, and the United States: SSantos, LM, NM and SStall respectively). This team comprised members with considerable expertise in journal publication (i.e. SCIELO, AGU), methods of attribution and provenance (e.g. ORCID), all underpinned by data management expertise. As the PARSEC common language was English, the Checklist team used English as its primary language.
At the time the Checklist team was forming, and the format of the Open Science guidance was being considered, the team converged on the concept of a concise list or set of tasks—a Checklist—that could be quickly understood by a researcher. The team had collectively found that this technique was effective in providing guidance and helping researchers learn and implement new practices efficiently. Checklists are recognised as useful support tools in the implementation of complex processes, helping to summarise information in a concise manner by breaking them down into a series of clear and actionable steps [57]. In PARSEC they needed to be as generic as possible for the scientific and technical elements to be relevant to the wider scientific community.
Working iteratively, and informed by experiences with the PARSEC Synthesis Science Strand, the team organised the elements of Open Science specific to data and software management into two levels and produced three checklists for each (Fig 3):
- Elements that a researcher can control directly (1: data presence, 2: data documentation, and 3: software documentation)
- Elements needed by a research team and their leader (4: open science practices, 5: resources and guidance, and 6: digital objects preservation).
This workflow shows the main steps to develop and translate each of the checklists for both level A and level B: determine team needs; iterate on the conceptualization; develop each checklist; translate each checklist; publish, promote, and receive feedback.
The expertise of one of the post-doctoral team members, a native Spanish speaker, was harnessed to provide initial input. Updates were made in consequence to this and feedback from other PARSEC members. The initial English version of each checklist was published in the PARSEC Zenodo community [58–64]. The completed checklists were promoted through the Data Science Strand’s networks and the feedback received informed revisions of each checklist which were then re-published in Zenodo under version control (Fig 3).
After strong positive feedback was received about the English version of the checklists, it was clear that translations into other languages would be optimal (consistent with [12]). As the Checklist team had native speakers of French, Japanese, Portuguese and Spanish, these languages were our starting point. Each team member leading the translation consulted with someone local to them who could validate it. Questions about context were brought back to the project team and clarifications made, including tracking updates to the English version (Fig 3).
Each checklist was reviewed by the larger PARSEC team and feedback was incorporated. As the translations were completed, they were published on Zenodo [58], and promoted on social media and at relevant conferences. The team collected any additional feedback for further updates.
The French [64–70], Japanese [71–73], Portuguese [74–79] and Spanish [80–85] checklists at the time of writing this paper are shown in Fig 4.
Full citations for the English versions are in the references. TBC is an abbreviation for ‘to be created’.
5 Methods
Each team was asked to reflect on its practices, (a) identifying terminological obstacles it encountered in its workflows, (b) to classify them in one of the nine ways shown in Table 1 (inspired by David et al. [21]), and (c) to categorise them on a five-point scale of importance, from 5, indicating a challenge that stopped progress, to 1, a challenge that was easily overcome. The teams did not include a professional linguist but the terminology made sense to us.
The DL team reflected on its practice, guided by its workflow (Fig 2). Feedback was achieved by completion of a table, supported by a series of virtual meetings. The results were analysed by team members, feedback was obtained, and the table updated. The Checklist team examined the steps in its workflow (Fig 3) through a series of meetings, identifying and describing the stumbling blocks encountered.
6 Reflective observations
The two teams presented their observations in different ways, reflective of their different task-sets.
6.1 DL team
From initial discussions it became clear that misunderstandings and difficulties in the use of terminology specific to DL methodologies and processes were important. Team members felt strongly that their work was hampered by poor inter-domain understanding of the aims of the project, the scientific question, the hypothesis, and the DL methodology. This led to confusion and lack of useful data provision, causing difficulties in DL model training, estimation, and validation. Team members from different countries had different understandings and norms.
The data collection phase (Fig 2) required cross-team collaboration, both for the acquisition of satellite imagery and socioeconomic data and to optimally align the two. The two relevant disciplinary groups (spatial and socioeconomic) had to work with the DL team to select appropriate data for the analysis, including the time range and the frequency of repeat observations, while ensuring the specificity of each country’s organisational structures was accommodated. This required more effort and time than was initially expected. Often available data were in a non-English language with no translation available, which prevented their use and stopped the investigation completely. As Demographic and Health Survey data were not available in all the study countries, a robust and comparable poverty (or wealth) index had to be created, requiring advice and participation from across the team. Once these matters were in place, the actual DL modelling phase was straightforward. The evaluation stage, as the last step of the workflow, again required an integration of ideas and communication with the full team to understand the meaning of obtained results (the spatio-temporal mapping estimations).
When reviewing the occurrence of terminological impediments, four workflow stages were noted: (i) definition of the work, (ii) collection of both remotely-sensed imagery and socioeconomic data, (iii) pre-processing of imagery and (iv) data splitting for cross validation. These workflow stages highlighted the lack of explicit methodological description in many publications (Table 2). Overall, eight obstacles were identified, with most clustered around challenges concerning the whole team’s understanding of the DL work, the acquisition of population data for the project, and incomplete or inaccurate terminology and metadata in resources (e.g. exemplar articles and data sources).
Some of the linguistic (or terminological) confusion arose during the conceptualisation and establishment of the experiments as a result of disciplinary differences, such as the use of the word ’ecoregion’ (1a, Table 2). This term was assumed by the DL team to refer to an abstract model for the work, the DL team being unaware of its well-accepted use in environmental science, and its basis for the Synthesis Science Strand’s work [95, 96]. Once this was explained, the methodology for selecting sites had to be revised. Similarly, the term ’experiment’ proved to be quite a stumbling block as it had a different meaning for computer scientists than experimental scientists (1d, Table 2). In both cases discrepancies were not realised until after some of the work had been completed and were exposed when sharing with interdisciplinary team members from other parts of the PARSEC project.
At a more practical level, the enumeration area was often not only named differently from country to country but also differently defined spatially (2a, Table 2). For example, population size as a social clustering tool was used for collection and delivery of data in many countries, which could mean different cluster types from one location to another. This created some difficulties for the remote sensing experts on the team when comparing physically different and administratively (geopolitically) different areas. In addition, names and designated areas were sometimes dynamic over the study period.
Several terms necessary for incorporating data into the DL model were different from country to country (2b, Table 2); for example, the methods and terminology used to obtain socioeconomic data. In some methods, data from a sample of people (a panel) were used and re-surveyed over repeated instances (waves), and in others, data were obtained from whole country surveys (a census). The frequency and duration of data collection was also different across countries (e.g. every 5 years versus 10 years). An additional issue was that, in several cases, terms were used somewhat casually by team members, and differences between such terms were outside the DL team’s knowledge. Realising this required active and constant dialogue among team members, which in a distributed team across multiple time zones was not always easy to achieve. The use of varying terminology for the same concept (linguistic confusion [I]), not just within the team but more broadly in the referenced literature, was an obstacle that also had to be overcome (3a, Table 2).
Two additional confusions arose when explaining the constraints of the DL method to non-DL specialists in the team (4a, Table 2). In training a DL model, the test sets and the training sets must not overlap, else one will get a positive bias. For instance, X is a dataset split into X_train and X_test. The model is trained on X_train and evaluated on X_test, and a performance score is computed. This is referred to as a ’test (or regular test)’. Then, it is often argued that the model can be used on another dataset Y. It can be, but sometimes the underlying message is that the performance score on the dataset Y and X_test will be similar. To prove the last statement one could use an ’out of domain test’. In a related confusion, an exemplar paper used the terms ’within/in-country’ and ’out-of-country’ without explanation.
Impediments to progress were particularly acute at the beginning of the project and they returned somewhat at the end. An assumption was made that ’everyone knows this’ and it was clearly not the case for the DL team. Similarly, casualness and variability in terminological use in important and relevant literature produced significant delays; some could be solved by speaking with the authors, but that was not always possible (4a, Table 1). Basic across-team misunderstanding generally produced challenges (1d, 2a and 2b, Table 1). Developing a common cross-country suite of terminology and definitions for the DL team to enable its model to be applied was time consuming (2a, Table 1).
6.2 Checklist team
The goal of the Checklist team was to ensure that researchers reading in their native language would easily understand the concepts and recommendations included in each checklist. In short, to make Open Science easy to adopt in daily practice. Because the concepts and practice around Open Science are still new, identification of language-specific references was difficult. Many sources were used to support the translations with the expectation that on-line translation tools would be helpful, but they were not completely so.
Concepts without existing equivalent words.
The English language has been historically influenced by Latin languages, borrowing many of its words (e.g. ’Data’ is Latin, ’Management’ comes from old French, as does ‘Plan’). This linguistic proximity and the format of the scripts used facilitate translation between English and the Latin-based languages of French, Portuguese and Spanish. For languages based on other roots or where the basic form is an ideogram, the translations from English are less direct. For example, the phonetic (katakana) transcription for the term ‘Data Management Plan’ (データマネジメントプラン, deetamanijimentopuran) was used, as there was no Japanese equivalent term with this meaning. Specifically, the word is directly ‘borrowed’ from the English. If the equivalent kanji was used (which in theory it could be), it would be a lengthy and confusing term. In contrast, the concept of ’Data Management Plan’ can be directly mapped in French to ’Plan de gestion des données’. Another example encountered was the English term ’funder’. The verb suffix ‘-er’ indicates someone or an organisation that ‘does something’, but in Japanese ‘funder’ is expanded as 研究資金配分機関, meaning ‘organisation allocating research funding’. It must be explained literally.
Nature of issue: [H] lexical gap; [C] fuzziness in language. A minor interruption [1].
Concepts that exist but do not have a routine translation.
In French the strict translation of the word ‘checklist’ would be ’liste de vérification / contrôle’ which is not usually used in practice. As the English term ‘checklist’ is known in the French language, and to be consistent with the easy-to-use and evocative character of the checklist terms, we determined that we should choose a distinctive name; namely, ‘check-liste’.
Similarly, in Japanese, words derived from foreign languages usually make sense if they are written in katakana with a close pronunciation. For example, the phonetic translation for ‘computer’ is コンピュータ (konpyuuta) and ‘notebook’ is ノートブック (nootobukku). However, when new terms are generated by simply combining existing words, such as ‘computation(al) notebook’ to describe something specific (in this case applications such as Jupyter Notebook and R Markdown), doing the same in Japanese would not work. For this reason, the translation of ‘computation(al) notebook’ into Japanese had to use the redundant phrase 計算機上のノートブック環境, a ‘notebook environment on computers’.
Nature of issue: [C] fuzziness in language; [F] country of practice; [H] lexical gap; [I] adoption challenges. A significant interruption [3] and clarifications had to be made.
Translations exist, but they need social endorsement for adoption.
When practising Open Science, being inclusive includes taking the time to ensure that a new concept is fully understood and used by others in their language. This applied to both the checklists being developed, as well as the concepts being used. If a language did not have the words broadly accepted in a relevant community to describe a concept, care had to be taken in the translation process. If the concept needed a concise word or phrase that was not yet widely recognised, the word would be introduced in the language along with a description.
An example is the expression, ‘Digital Presence’, which is emerging alongside increasing online availability of research products. Researchers need to adjust their practices to take advantage of this opportunity, and thus create a ‘Digital Presence’. In French a ‘Digital Presence’ can be translated as ‘Présence numérique’ which is broadly understood but not yet a common expression in all disciplines of science. Until a research community endorses the new term it will not be widely accepted or adopted.
The Japanese translation of ‘Digital Presence’ fell into this same category (vaguely known but not widely adopted), and we decided to use the katakana デジタルプレゼンス (dejitarupurezensu) to simply mimic the pronunciation as it would be more likely to attract the reader’s attention and be remembered as a concept.
Nature of issue: [H] lexical gap; [I] adoption challenges. A significant interruption [3] and clarifications had to be made.
A compendium of the words and phrases used in the checklists is shown in Table 3. The table illustrates that different methods for translation need to be used to ensure the best comprehension of important words. This may include a decision to create new words or ‘borrow’ words from another language. In Japanese a phonetic transcription is used when words are borrowed. The acronym for Data Management Plan, ‘PDG’, is the same for the Latin-based languages (Portuguese, Spanish and French) while ‘DMP’ is used for English and Japanese. In the latter case this is due to the English words being borrowed as shown by the phonetic transcription.
7 Conceptualization
As mentioned in the Introduction, multinational, multilingual, and multidisciplinary collaborations require that each party understands the different ’languages’ used by their fellow researchers. We posed the following research questions: (a) what are the terminological obstacles encountered, (b) are mismatches in language concepts due to disciplinary differences, native language, or country-specific nomenclature, and (c) what is the impact of these obstacles on the two workflows. There were indeed several unfamiliar concepts within and outside each of the teams that needed to be translated or explained with variable success. This affected team communication, the acquisition and interpretation of data and information, communication with the wider community, and ultimately in generating outputs and products. The retrospective linguistic classification of the terminological obstacles (A-I) assisted our insight into the actual nature of the difficulties we faced and provided us with a pathway to solutions. Being aware of these potential confusions at the start of the project might have enabled us to anticipate them and to respond more efficiently when they occurred.
The two sub-teams (’DL’ and ’Checklist’) studied in this paper identified different terminological obstacles, albeit with some commonalities. Fuzziness in language, disciplinary differences, linguistic confusion, lexical gaps and adoption challenges were most often cited as the cause of terminological mismatches. Some of these obstacles significantly impeded progress, such as the lack of terminological equivalents across disciplinary domains or languages. Some obstacles were less significant, but required attention if project goals were to be achieved, such as terminology used in one domain that was not familiar to those from another, or adoption challenges when communicating new concepts in a country. Some terms were ’known’, but not yet sufficiently accepted to be translated smoothly. The technical nature of many of the terms and concepts made translation challenging unless the translator had disciplinary insight. Terminology and approaches to nomenclature unique to one country were on occasions surprising and required adjustment, taking time and effort in addition to the main work. Disappointingly, given the expected high standards of scholarly publishing, obstacles occurred due to key terms on which research was based not being defined in the source literature.
Often obstacles took some time to overcome depending on the cognitive, linguistic or developmental distance involved. The cognitive distance across the team was not insignificant; for example DL programmers worked directly with human geographers, a considerable disciplinary difference. There was a joke within the team that for one particular concept, where the word used was the same in two disciplines but had a different meaning in each, that it took ten attempts to explain the difference. It was discussed that the number of times an explanation had to be made could form a scale of difficulty that could be applied to all. As mentioned in the description of the case study, the Data Science Strand team created an on-line ’dictionary’ in the shared workspace at the beginning of the project that was intended as a place where everyone could record terms and definitions, but it was not used. Retrospectively this resource could have alleviated some confusion if team members were (a) more aware of its existence, and (b) contributed to it.
While engaged in the reflective observation process illustrated in this paper, we noted that the use of English as our common working language was essential, but team members had varying levels of English proficiency. This can affect the degree of collaboration and the development of trust [34, 97]. It was observed that although all communications were made in English, it may not have been ‘standard’ English in each case. A ‘common’ version of English was often established that had elements of the mother tongue of each interlocutor. This adjustment was unique to the task at hand, dependent on the team members involved, and not meant to be shared outside PARSEC. This lack of standardisation in language also occurred across disciplinary boundaries, as parties listening to one another interpreted meanings according to their own perspective (Enryo-Sasshi: [98]). The translations by the Checklist team were designed to alleviate this effect for the external research community.
The complexity of translating research concepts into the five main languages in this paper shows how difficult concept-based translations can be. The illustrations provided by Ducarme et al. [17] and Droz et al. [16, 18] for the different meanings conveyed by the concept of ‘Nature’ in a range of languages are evidence of this difficulty. Good translations occur at the concept level, not as a simple word-for-word translation [21], a limitation of most on-line translation tools, and possibly also a limitation of many online domain-specific translation tools. Vanderbilt et al. [6] tested an automated translation of metadata using Google translate, Bing and World Lingo Version, none of which were close to being completely correct. Back translation showed a 60% accuracy for Japanese or Chinese to English, but 90% for Swedish to English. Tools have improved since then, certainly, but caution still needs to be applied.
One of the most important realisations from this work was the need to ensure that all team members had a strong common understanding of the project purpose, good knowledge of the tasks to be done, the requirements for these tasks, and their contribution to them. This should be established at the beginning of the project and revisited throughout the project (the ’redundancy’ of Podestá et al. [99]). In PARSEC full team meetings were held every six months, but more regular conversations were clearly beneficial, especially around particular tasks (as practised within the Data Science Strand). The importance of these meetings was not appreciated by all team members at the time, but not only did it enable the development of trust among team members but also kept conversations alive. It is, as is often the case, only in retrospect that the value of such diligence occurs.
Although co-design of research is commonly a practice mentioned in projects where members of a wider community are involved (i.e. stakeholders who are non-scientists, such as citizen scientists and indigenous knowledge-holders) [3, 100], it is clearly an appropriate model here. In both case studies—the e-informatics research of the DL team, and the communications work of the Checklist team—members from other disciplines or native speakers of other languages, were able to provide checks and clarifications in a timely manner, even if not involved in the core work of a particular team, and thus avoid delays.
8 Learning outcomes
The objective of this paper was to highlight, through an empirical approach, the pitfalls of ignoring sources of terminological confusion in technically innovative, multidisciplinary, multinational, and multilingual research projects. Our case study provides unique insight into the effect of terminological obstacles on project work in real time. It confirms much previous understanding but adds a perspective from lived experience. Guided by a classification of terminological obstacles (A-I), our two contrasting research teams reflected on the nature of the terminological obstacles encountered, and the effect on their workflows. This paper not only reports their reflection but also distills their observations into guidelines that could prove useful to others attempting such a journey.
Three different situations emerged from our research experience: (1) concepts without existing equivalent words, (2) concepts that existed but did not have a routine translation, and (3) cases where translations existed but needed social endorsement for adoption. Fuzziness in language, linguistic and disciplinary confusion, lexical gaps and adoption challenges were most often cited as obstacles across both workflow-types. Linguistic distance, when translating new concepts associated with open science across countries and languages, was noted. Practical matters of international data collection and comparison included an unanticipated need to incorporate different types of data labels from country to country, authority to authority.
We thus make the following simple but practical recommendations to improve the success of multinational, multilingual, and multidisciplinary projects (Table 4). These recommendations are conceptual drawing on experience through a reflexive approach (Fig 1) and remain to be sources for discussion and testing by others embarking on their research journey.
Acknowledgments
We appreciate the constructive comments of the reviewers, the helpful advice of Dr Glyn Rimmington, USA and Australia, and Prof. Yasuhisa Kondo of the Research Institute for Humanity and Nature, Japan.
References
- 1. Ernakovich JG, Eklund N, Varner RK, Kirchner N, Jeuring J, Duderstadt K, et al. Is A Common Goal A False Hope in Convergence Research?: Opportunities and Challenges of International Convergence Research to Address Arctic Change. Earth’s Future [Internet]. 2021 [cited 2024 Jan 19];9(5):e2020EF001865. Available from: https://agupubs.onlinelibrary.wiley.com/doi/10.1029/2020EF001865
- 2. Büntgen U, Rees G. Global change research needs international collaboration. Science of The Total Environment [Internet]. 2023 [cited 2024 Jan 19];902:166054. Available from: https://linkinghub.elsevier.com/retrieve/pii/S004896972304679X pmid:37543344
- 3.
UNESCO. UNESCO Recommendation on Open Science. 2021 Nov 23 [cited 2024 Apr 20]; Available from: https://zenodo.org/record/5834767
- 4. Dusdal J, Powell JJW. Benefits, Motivations, and Challenges of International Collaborative Research: A Sociology of Science Case Study. Science and Public Policy [Internet]. 2021 Jun 23 [cited 2022 Jan 1];48(2):235–45. Available from: https://academic.oup.com/spp/article/48/2/235/6135106
- 5. Sidlauskas B, Ganapathy G, Hazkani-Covo E, Jenkins KP, Lapp H, McCall LW, et al. Linking Big: the continuing promise of evolutionary synthesis. Evolution [Internet]. 2009 Oct 23 [cited 2024 Jan 19];64(4):871–80. Available from: https://academic.oup.com/evolut/article/64/4/871/6853098 pmid:19895550
- 6. Vanderbilt K, Porter JH, Lu SS, Bertrand N, Blankman D, Guo X, et al. A prototype system for multilingual data discovery of International Long-Term Ecological Research (ILTER) Network data. Ecological Informatics [Internet]. 2017 [cited 2023 Oct 10];40:93–10 Available from: https://linkinghub.elsevier.com/retrieve/pii/S1574954116301297
- 7. Wagner CS, Whetsell TA, Mukherjee S. International research collaboration: Novelty, conventionality, and atypicality in knowledge recombination. Research Policy [Internet]. 2019 [cited 2021 Jul 6];48(5):1260–70. Available from: https://linkinghub.elsevier.com/retrieve/pii/S0048733319300046
- 8. Specht A, Crowston K. Interdisciplinary collaboration from diverse science teams can produce significant outcomes. Lozano S, editor. PLoS ONE [Internet]. 2022 Nov 29 [cited 2022 Dec 1];17(11):e0278043. Available from: pmid:36445918
- 9. Baker B. The Science of Team Science. BioScience [Internet]. 2015 Jul 1 [cited 2020 Sep 9];65(7):639–44. Available from: http://academic.oup.com/bioscience/article/65/7/639/258550/The-Science-of-Team-ScienceAn-emerging-field
- 10. Crowston K, Specht A, Hoover C, Chudoba KM, Watson-Manheim MB. Perceived discontinuities and continuities in transdisciplinary scientific working groups. Science of The Total Environment [Internet]. 2015 [cited 2021 Jul 5];534:159–72. Available from: https://linkinghub.elsevier.com/retrieve/pii/S0048969715300553 pmid:25957788
- 11. Prieto D. Make research-paper databases multilingual. Nature [Internet]. 2018 [cited 2024 Jan 16];560(7716):29–29. Available from: https://www.nature.com/articles/d41586-018-05844-0 pmid:30065332
- 12. Amano T, Rios Rojas C, Boum Ii Y, Calvo M, Misra BB. Ten tips for overcoming language barriers in science. Nat Hum Behav [Internet]. 2021 Jul 8 [cited 2023 Sep 20];5(9):1119–22. Available from: https://www.nature.com/articles/s41562-021-01137-1 pmid:34239079
- 13. Lynch AJ, Fernández-Llamazares Á, Palomo I, Jaureguiberry P, Amano T, Basher Z, et al. Culturally diverse expert teams have yet to bring comprehensive linguistic diversity to intergovernmental ecosystem assessments. One Earth [Internet]. 2021 [cited 2024 Jan 19];4(2):269–78. Available from: https://linkinghub.elsevier.com/retrieve/pii/S2590332221000488
- 14. Amano T, Berdejo-Espinola V, Christie AP, Willott K, Akasaka M, Báldi A, et al. Tapping into non-English-language science for the conservation of global biodiversity. PLOS Biology [Internet]. 2021 Oct 7 [cited 2022 Mar 29];19(10):e3001296. Available from: https://journals.plos.org/plosbiology/article?id=10.1371/journal.pbio.3001296 pmid:34618803
- 15. Chowdhury S, Gonzalez K, Aytekin MÇK, Baek S, Bełcik M, Bertolino S, et al. Growth of non‐English‐language literature on biodiversity conservation. Conservation Biology [Internet]. 2022 Mar 24 [cited 2022 Mar 28]; Available from: https://onlinelibrary.wiley.com/doi/10.1111/cobi.13883 pmid:34981574
- 16. Droz L, Chen HM, Chu HT, Fajrini R, Imbong J, Jannel R, et al. Exploring the diversity of conceptualizations of nature in East and South-East Asia. Humanit Soc Sci Commun [Internet]. 2022 May 31 [cited 2024 May 3];9(1):186. Available from: https://www.nature.com/articles/s41599-022-01186-5
- 17. Ducarme F, Flipo F, Couvet D. How the diversity of human concepts of nature affects conservation of biodiversity. Conservation Biology [Internet]. 2021 [cited 2024 May 3];35(3):1019–28. Available from: https://conbio.onlinelibrary.wiley.com/doi/10.1111/cobi.13639 pmid:32996235
- 18. Droz L, Brugnach M, Pascual U. Multilingualism for pluralising knowledge and decision making about people and nature relationships. People and Nature [Internet]. 2023 [cited 2024 May 2];5(3):874–84. Available from: https://besjournals.onlinelibrary.wiley.com/doi/10.1002/pan3.10468
- 19. Amano T, Berdejo-Espinola V, Akasaka M, De Andrade Junior MAU, Blaise N, Checco J, et al. The role of non-English-language science in informing national biodiversity assessments. Nat Sustain [Internet]. 2023 Mar 16 [cited 2023 Sep 21];6(7):845–54. Available from: https://www.nature.com/articles/s41893-023-01087-8
- 20. Arenas-Castro H, Berdejo-Espinola V, Chowdhury S, Rodríguez-Contreras A, James ARM, Raja NB, et al. Academic publishing requires linguistically inclusive policies. Proc R Soc B [Internet]. 2024 Mar 13 [cited 2024 Mar 27];291(2018):20232840. Available from: https://royalsocietypublishing.org/doi/10.1098/rspb.2023.2840 pmid:38471557
- 21. David R, Specht A, O’Brien M, Wyborn L, Drummond C, Edmunds R, et al. Multilingual Data Challenges in Professionalizing Data Stewardship worldwide. 2022 May 27 [cited 2023 Sep 20]; Available from: https://zenodo.org/record/6588167
- 22. Rimmington GM, Alagic M. Third place learning: reflective inquiry into intercultural and global cage painting. Charlotte, N.C: IAP; 2008. 162 p. (Teaching-learning indigenous, intercultural worldviews).
- 23. McGreavy B, Haynal K, Smith-Mayo J, Reilly-Moman J, Kinnison MT, Ranco D, et al. How Does Strategic Communication Shape Transdisciplinary Collaboration? A Focus on Definitions, Audience, Expertise, and Ethical Praxis. Front Commun [Internet]. 2022 Feb 14 [cited 2024 Jan 8];7:831727. Available from: https://www.frontiersin.org/articles/10.3389/fcomm.2022.831727/full
- 24. Pennington DD, Simpson GL, McConnell MS, Fair JM, Baker RJ. Transdisciplinary Research, Transformative Learning, and Transformative Science. BioScience [Internet]. 2013 [cited 2024 Jan 19];63(7):564–73. Available from: https://academic.oup.com/bioscience/article-lookup/doi/10.1525/bio.2013.63.7.9
- 25.
Someh I, The University of Queensland, Wixom B, Massachusetts University of Technology, Davern M, The University of Melbourne, et al. Configuring Relationships between Analytics and Business Domain Groups for Knowledge Integration. JAIS [Internet]. 2023 [cited 2024 Mar 15];24(2):592–618. Available from: https://aisel.aisnet.org/jais/vol24/iss2/1/
- 26. Henson VR, Cobourn KM, Weathers KC, Carey CC, Farrell KJ, Klug JL, et al. A Practical Guide for Managing Interdisciplinary Teams: Lessons Learned from Coupled Natural and Human Systems Research. Social Sciences [Internet]. 2020 Jul 9 [cited 2024 Mar 21];9(7):119. Available from: https://www.mdpi.com/2076-0760/9/7/119
- 27. Carlile PR. A Pragmatic View of Knowledge and Boundaries: Boundary Objects in New Product Development. Organization Science [Internet]. 2002 [cited 2024 Jan 8];13(4):442–55. Available from: https://pubsonline.informs.org/doi/10.1287/orsc.13.4.442.2953
- 28. Yegros-Yegros A, Rafols I, D’Este P. Does Interdisciplinary Research Lead to Higher Citation Impact? The Different Effect of Proximal and Distal Interdisciplinarity. Glanzel W, editor. PLoS ONE [Internet]. 2015 Aug 12 [cited 2021 Jul 6];10(8):e0135095. Available from: https://dx.plos.org/10.1371/journal.pone.0135095 pmid:26266805
- 29. Qi F, Zhou H, Sun B, Huang Y, Zhang L. Facilitating interdisciplinarity: the contributions of boundary-crossing activities among disciplines. Scientometrics [Internet]. 2024 Jan 6 [cited 2024 May 6]; Available from: https://link.springer.com/10.1007/s11192-023-04924-x
- 30. Iwanaga T, Zare F, Croke B, Fu B, Merritt W, Partington D, et al. Development of an integrated model for the Campaspe catchment: a tool to help improve understanding of the interaction between society, policy, farming decision, ecology, hydrology and climate. Proc IAHS [Internet]. 2018 Jun 5 [cited 2024 Mar 15];379:1–12. Available from: https://piahs.copernicus.org/articles/379/1/2018/
- 31. Pennington D. A conceptual model for knowledge integration in interdisciplinary teams: orchestrating individual learning and group processes. J Environ Stud Sci [Internet]. 2016 [cited 2024 Apr 13];6(2):300–12. Available from: http://link.springer.com/10.1007/s13412-015-0354-5
- 32. Leahey E, Beckman CM, Stanko TL. Prominent but Less Productive: The Impact of Interdisciplinarity on Scientists’ Research. Administrative Science Quarterly [Internet]. 2017 [cited 2021 Jul 5];62(1):105–39. Available from: http://journals.sagepub.com/doi/10.1177/0001839216665364
- 33. Chen S, Qiu J, Arsenault C, Larivière V. Exploring the interdisciplinarity patterns of highly cited papers. Journal of Informetrics [Internet]. 2021 [cited 2021 Dec 15];15(1):101124. Available from: https://linkinghub.elsevier.com/retrieve/pii/S1751157720306416
- 34. Li H (Jessica), Yuan , Bazarova NN, Bell BS. Talk and Let Talk: The Effects of Language Proficiency on Speaking Up and Competence Perceptions in Multinational Teams. Group & Organization Management [Internet]. 2019 [cited 2024 Apr 22];44(5):953–89. Available from: http://journals.sagepub.com/doi/10.1177/1059601118756734
- 35. Yin X, Li P, Feng Z, Yang Y, You Z, Xiao C. Which Gridded Population Data Product Is Better? Evidences from Mainland Southeast Asia (MSEA). ISPRS International Journal of Geo-Information [Internet]. 2021 Oct [cited 2024 Jan 19];10(10):68 Available from: https://www.mdpi.com/2220-9964/10/10/681
- 36. David R, Ohmann C, Boiten JW, Abadía MC, Bietrix F, Canham S, et al. An iterative and interdisciplinary categorisation process towards FAIRer digital resources for sensitive life-sciences data. Sci Rep [Internet]. 2022 Dec 5 [cited 2024 May 6];12(1):20989. Available from: https://www.nature.com/articles/s41598-022-25278-z pmid:36470968
- 37. Machicao J, Ben Abbes A, Meneguzzi L, Corrêa PLP, Specht A, David R, et al. Mitigation Strategies to Improve Reproducibility of Poverty Estimations From Remote Sensing Images Using Deep Learning. Earth and Space Science [Internet]. 2022 [cited 2022 Nov 8];9(8). Available from: https://onlinelibrary.wiley.com/doi/10.1029/2022EA002379
- 38. Wilkinson MD, Dumontier M, Aalbersberg IjJ, Appleton G, Axton M, Baak A, et al. The FAIR Guiding Principles for scientific data management and stewardship. Sci Data [Internet]. 2016 Mar 15 [cited 2024 Jan 2];3(1):160018. Available from: https://www.nature.com/articles/sdata201618 pmid:26978244
- 39. Goble C, Cohen-Boulakia S, Soiland-Reyes S, Garijo D, Gil Y, Crusoe MR, et al. FAIR Computational Workflows. Data Intellegence [Internet]. 2020 [cited 2024 May 6];2(1–2):108–2 Available from: https://direct.mit.edu/dint/article/2/1-2/108-121/10003
- 40. Wood Daudelin M. Learning from experience through reflection. Organizational Dynamics [Internet]. 1996 [cited 2024 Sep 18];24(3):36–48. Available from: https://linkinghub.elsevier.com/retrieve/pii/S0090261696900042
- 41. Kayes AB, Kayes DC, Kolb DA. Experiential learning in teams. Simulation & Gaming [Internet]. 2005 [cited 2024 Sep 17];36(3):330–54. Available from: http://journals.sagepub.com/doi/10.1177/1046878105279012
- 42. Rupčić N. How does learning facilitate development? TLO [Internet]. 2024 May 9 [cited 2024 Sep 19];31(3):449–57. Available from: https://www.emerald.com/insight/content/doi/10.1108/TLO-04-2024-306/full/html
- 43. Dong H, Dacre N, Baxter D, Ceylan S. What is Agile Project Management? Developing a New Definition Following a Systematic Literature Review. Project Management Journal [Internet]. 2024 May 20 [cited 2024 Sep 18];87569728241254100. Available from: https://journals.sagepub.com/doi/10.1177/87569728241254095
- 44.
Kolb DA. Experiential learning: experience as the source of learning and development. Second edition. Upper Saddle River, New Jersey: Pearson Education, Inc; 2015. 390 p.
- 45.
Khanna D, Wang X. How Software Startup Teams Reflect: Approaches, Triggers and Challenges. In: Hyrynsalmi S, Suoranta M, Nguyen-Duc A, Tyrväinen P, Abrahamsson P, editors. Software Business [Internet]. Cham: Springer International Publishing; 2019 [cited 2024 Sep 19]. p. 353–68. Available from: https://link.springer.com/10.1007/978-3-030-33742-1_28
- 46. Kohlbacher F. The Use of Qualitative Content Analysis in Case Study Research. Forum Qualitative Sozialforschung / Forum: Qualitative Social Research [Internet]. 2006 Jan 31 [cited 2024 Sep 19];Vol 7:No 1 (2006): Learning About Risk. Available from: http://www.qualitative-research.net/index.php/fqs/article/view/75
- 47.
Yin RK. Case study research: design and methods [Internet]. 4th ed. Los Angeles, Calif: Sage Publications; 2009 [cited 2024 Sep 2]. 219 p. (Applied social research methods). Available from: https://www.google.com.au/books/edition/Case_Study_Research/FzawIAdilHkC?hl=en&gbpv=1&dq=Yin+2009&pg=PR1&printsec=frontcover
- 48. Priya A. Case Study Methodology of Qualitative Research: Key Attributes and Navigating the Conundrums in Its Application. Sociological Bulletin [Internet]. 2021 [cited 2024 Sep 19];70(1):94–110. Available from: http://journals.sagepub.com/doi/10.1177/0038022920970318
- 49. Specht A, O’Brien M, Edmunds R, Corrêa P, David R, Mabile L, et al. The Value of a Data and Digital Object Management Plan (D(DO)MP) in Fostering Sharing Practices in a Multidisciplinary Multinational Project. Data Science Journal [Internet]. 2023 Oct 6 [cited 2023 Oct 9]; 22:38. Available from: http://datascience.codata.org/articles/10.5334/dsj-2023-038/
- 50. Specht A, Stall S, Mouquet N, Mouillot D, Corrêa P, Murayama Y. PARSEC: Building New Tools for Data Sharing and Reuse through a Transnational Investigation of the Socioeconomic Impacts of Protected Areas. 2019 Nov 28 [cited 2024 Jul 30]; Available from: https://zenodo.org/record/3890876
- 51. Specht A, Stall S. PARSEC entry data management survey. 2022 Dec 11 [cited 2024 May 6]; Available from: https://zenodo.org/record/7425596
- 52.
Jarry R, Chaumont M, Berti-Équille L, Subsol G. Assessment of CNN-Based Methods for Poverty Estimation from Satellite Images. In: Del Bimbo A, Cucchiara R, Sclaroff S, Farinella GM, Mei T, Bertini M, et al., editors. Pattern Recognition ICPR International Workshops and Challenges [Internet]. Cham: Springer International Publishing; 2021 [cited 2024 May 6]. p. 550–65. Available from: http://link.springer.com/10.1007/978-3-030-68787-8_40
- 53. Machicao J, Specht A, Vellenich D, Meneguzzi L, David R, Stall S, et al. A Deep-Learning Method for the Prediction of Socio-Economic Indicators from Street-View Imagery Using a Case Study from Brazil. Data Science Journal [Internet]. 2022 Feb 11 [cited 2022 Feb 14];21:6. Available from: http://datascience.codata.org/articles/10.5334/dsj-2022-006/
- 54. Ben Abbes A, Machicao J, Corrêa PLP, Specht A, Devillers R, Ometto JP, et al. DeepWealth: A generalizable open-source deep learning framework using satellite images for well-being estimation. SoftwareX [Internet]. 2024 Sep 1 [cited 2024 Jun 21];27:101785. Available from: https://www.sciencedirect.com/science/article/pii/S2352711024001560
- 55. Ben Abbes A, Machicao J, Correa PLP, Specht A, Devillers R, Ometto JP, et al. Source code: DeepWealth: A Generalizable Open-Source Deep Learning Framework using Satellite Images for Well-Being Estimation [Internet]. 2024 [cited 2024 Jun 23]. Available from: https://zenodo.org/doi/10.5281/zenodo.12189299
- 56. David R, Mabile L, Specht A, Stryeck S, Thomsen M, Yahia M, et al. FAIRness Literacy: The Achilles’ Heel of Applying FAIR Principles. Data Science Journal [Internet]. 2020 Aug 11 [cited 2024 May 6];19:32. Available from: http://datascience.codata.org/articles/10.5334/dsj-2020-032/
- 57. Gehlbach H, Artino AR. The Survey Checklist (Manifesto). Academic Medicine [Internet]. 2018 [cited 2024 Mar 4];93(3):360–6. Available from: https://journals.lww.com/00001888-201803000-00018 pmid:29210756
- 58.
Zenodo [Internet]. [cited 2024 May 6]. Available from: https://zenodo.org/communities/parsec/about
- 59. Stall S, Specht A, Amato JG, Corrêa PLP, Curivil FAL, David R, et al. Digital Presence Checklist. 2023 Apr 18 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.4706118
- 60. Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, David R, et al. Data Documentation and Citation Checklist. 2023 Apr 18 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7062402
- 61. Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, David R, et al. Software Documentation and Citation Checklist. 2023 Apr 18 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7062413
- 62. Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, David R, et al. Open Science Practices for Teams (Prepare). 2023 Jun 15 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7402075
- 63. Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, David R, et al. Open Science Resources and Guidance for Teams (Equip). 2023 Jun 15 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7402270
- 64. Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, David R, et al. Digital Objects Preservation Checklist For Teams (Preserve). 2023 Jun 15 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7402540
- 65. David R, Santos S, Stall S, Specht A, Corrêa PLP, Machicao J, et al. Check-liste pour la Présence Numérique. 2023 May 9 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7847715
- 66. David R, Santos S, Stall S, Specht A, Corrêa PLP, Machicao J, et al. Check-liste pour la Documentation et la Citation des Données. 2023 May 9 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7847718
- 67. David R, Santos S, Stall S, Specht A, Corrêa PLP, Machicao J, et al. Check-liste pour la Documentation et la Citation des Logiciels. 2023 May 9 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7847721
- 68. David R, Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, et al. Check-liste «Préparez votre équipe à la Science Ouverte». 2023 May 30 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7985782
- 69. David R, Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, et al. Check-liste «Outillez votre équipe pour la Science Ouverte». 2023 May 30 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7985792
- 70. David R, Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, et al. Check-liste «Développez des pratiques de sauvegarde dans l’équipe». 2023 May 30 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7985798
- 71. Miyairi N, Murayama Y, Stall S, Specht A, Corrêa PLP, David R, et al. 研究の認知度向上のために: 研究者の「デジタルプレゼンス」チェックリスト. 2023 May 25 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7944747
- 72. Miyairi N, Murayama Y, Stall S, Specht A, Corrêa PLP, David R, et al. 研究データ共有のために: データのドキュメンテーションと引用のチェックリスト. 2023 May 25 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7944764
- 73. Miyairi N, Murayama Y, Stall S, Specht A, Corrêa PLP, David R, et al. ソフトウェア共有のために: ソフトウェアのドキュメンテーションと引用のチェックリスト. 2023 May 25 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7944760
- 74. Santos S, Stall S, Specht A, Amato JG, Corrêa PLP, Curivil FAL, et al. Checklist de Presença Digital. 2023 Apr 19 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7845372
- 75. Santos S, Stall S, Specht A, Corrêa PLP, David R, Machicao J, et al. Checklist de Documentação e Citação de Dados. 2023 Apr 19 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7845376
- 76. Santos S, Stall S, Specht A, Corrêa PLP, David R, Machicao J, et al. Checklist de Documentação e Citação de Software. 2023 Apr 19 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7845378
- 77. Santos S, Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, et al. Práticas de Ciência Aberta para equipes de pesquisa/projeto. 2023 Jun 1 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7995395
- 78. Santos S, Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, et al. Recursos e diretrizes para laboratórios/equipes. 2023 Jun 1 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7995397
- 79. Santos S, Stall S, Specht A, O’Brien M, Machicao J, Corrêa PLP, et al. Checklist para Preservação de Objetos Digitais da Equipe. 2023 Jun 1 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7995399
- 80. Santos S, Stall S, Specht A, Amato JG, Corrêa PLP, Curivil FAL, et al. Lista de verificación de presencia digital. 2023 Apr 19 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7841946
- 81. Santos S, Stall S, Specht A, Corrêa PLP, David R, Machicao J, et al. Lista de verificación de citas de datos y documentación. 2023 Apr 19 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7841992
- 82. Santos S, Stall S, Specht A, Corrêa PLP, David R, Machicao J, et al. Lista de verificación de citas y documentación de software. 2023 Apr 19 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7841994
- 83. Machicao J, Santos S, Stall S, Specht A, O’Brien M, Corrêa PLP, et al. Prácticas de Ciencia Abierta del equipo. 2023 Jun 13 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7995781
- 84. Machicao J, Santos S, Stall S, Specht A, O’Brien M, Corrêa PLP, et al. Prepare a su Equipo para la Ciencia Abierta–Recursos y guías del equipo. 2023 Jun 13 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7995783
- 85. Machicao J, Santos S, Stall S, Specht A, O’Brien M, Corrêa PLP, et al. Lista de verificación para la preservación de objetos digitales del equipo. 2023 Jun 13 [cited 2024 May 6]; Available from: https://zenodo.org/doi/10.5281/zenodo.7995787
- 86.
Kroeger PR. Analyzing Meaning: an introduction to semantics and pragmatics: Second corrected and slightly revised edition / Volume 0. [Place of publication not identified]: Language Science Press; 2019.
- 87. Zhang Q. Fuzziness ‐ vagueness ‐ generality ‐ ambiguity. Journal of Pragmatics [Internet]. 1998 [cited 2024 Jan 14];29(1):13–3 Available from: https://linkinghub.elsevier.com/retrieve/pii/S0378216697000143
- 88. Madin J, Bowers S, Schildhauer M, Krivov S, Pennington D, Villa F. An ontology for describing and synthesizing ecological observation data. Ecological Informatics [Internet]. 2007 [cited 2024 May 6];2(3):279–96. Available from: https://linkinghub.elsevier.com/retrieve/pii/S1574954107000362
- 89. Vanderbilt KL, Lin CC, Lu SS, Kassim AR, He H, Guo X, et al. Fostering ecological data sharing: collaborations in the International Long Term Ecological Research Network. Ecosphere [Internet]. 2015 [cited 2022 Mar 29];6(10):art204. Available from: http://doi.wiley.com/10.1890/ES14-00281
- 90. Vanderbilt KL, Blankman D, Guo X, He H, Lin CC, Lu SS, et al. A multilingual metadata catalog for the ILTER: Issues and approaches. Ecological Informatics [Internet]. 2010 [cited 2023 Oct 11];5(3):187–93. Available from: https://linkinghub.elsevier.com/retrieve/pii/S1574954110000336
- 91.
Kennedy C. 23. Ambiguity and vagueness: An overview. In: Maienborn C, Heusinger KV, Portner P, editors. Semantics [Internet]. Berlin, Boston: DE GRUYTER MOUTON; 2011 [cited 2024 Jan 15]. Available from: https://www.degruyter.com/document/doi/10.1515/9783110226614.507/html
- 92. Tabrizi HH, Pezeshki M. Strategies Used in Translation of Scientific Texts to Cope with Lexical Gaps (Case of Biomass Gasification and Pyrolysis Book). TPLS [Internet]. 2015 May 31 [cited 2024 Mar 15];5(6):1173. Available from: http://www.academypublication.com/issues2/tpls/vol05/06/07.pdf
- 93.
Ullman RE, Enloe Y. Accelerating Technology Adoption Through Community Endorsement. In: Di L, Ramapriyan HK, editors. Standard-Based Data and Information Systems for Earth Observation [Internet]. Berlin, Heidelberg: Springer Berlin Heidelberg; 2010 [cited 2024 Jan 19]. p. 227–48. Available from: http://link.springer.com/10.1007/978-3-540-88264-0_13
- 94. Borycz J, Olendorf R, Specht A, Grant B, Crowston K, Tenopir C, et al. Perceived benefits of open data are improving but scientists still lack resources, skills, and rewards. Humanit Soc Sci Commun [Internet]. 2023 Jun 20 [cited 2023 Jun 22];10(1):339. Available from: https://www.nature.com/articles/s41599-023-01831-7
- 95. Olson DM, Dinerstein E, Wikramanayake ED, Burgess ND, Powell GVN, Underwood EC, et al. Terrestrial Ecoregions of the World: A New Map of Life on Earth: A new global map of terrestrial ecoregions provides an innovative tool for conserving biodiversity. BioScience [Internet]. 2001 Nov 1 [cited 2021 Jan 24];51(11):933–8. Available from: https://doi.org/10.1641/0006-3568(2001)051[0933:TEOTWA]2.0.CO;2
- 96. Spalding MD, Fox HE, Allen GR, Davidson N, Ferdaña ZA, Finlayson M, et al. Marine Ecoregions of the World: A Bioregionalization of Coastal and Shelf Areas. BioScience [Internet]. 2007 Jul 1 [cited 2021 Jan 24];57(7):573–83. Available from: https://doi.org/10.1641/B570707
- 97. Tenzer H, Pudelko M, Harzing AW. The impact of language barriers on trust formation in multinational teams. J Int Bus Stud [Internet]. 2014 [cited 2024 Apr 26];45(5):508–35. Available from: http://link.springer.com/10.1057/jibs.2013.64
- 98. Miike KMC. 慶應義塾大学学術情報リポジトリ (KOARA) [Internet]. [cited 2024 Mar 23]. Available from: https://koara.lib.keio.ac.jp/xoonips
- 99. Podestá GP, Natenzon CE, Hidalgo C, Ruiz Toranzo F. Interdisciplinary production of knowledge with participation of stakeholders: A case study of a collaborative project on climate variability, human decisions and agricultural ecosystems in the Argentine Pampas. Environmental Science & Policy [Internet]. 2013 [cited 2024 Apr 22];26:40–8. Available from: https://linkinghub.elsevier.com/retrieve/pii/S1462901112001025
- 100. Bammer G, O’Rourke M, O’Connell D, Neuhauser L, Midgley G, Klein JT, et al. Expertise in research integration and implementation for tackling complex problems: when is it needed, where can it be found and how can it be strengthened? Palgrave Commun [Internet]. 2020 [cited 2021 Dec 14];6(1):5. Available from: http://www.nature.com/articles/s41599-019-0380-0.