Attitudes and practices of open data, preprinting, and peer-review—A cross sectional study on Croatian scientists

Attitudes towards open peer review, open data and use of preprints influence scientists’ engagement with those practices. Yet there is a lack of validated questionnaires that measure these attitudes. The goal of our study was to construct and validate such a questionnaire and use it to assess attitudes of Croatian scientists. We first developed a 21-item questionnaire called Attitudes towards Open data sharing, preprinting, and peer-review (ATOPP), which had a reliable four-factor structure, and measured attitudes towards open data, preprint servers, open peer-review and open peer-review in small scientific communities. We then used the ATOPP to explore attitudes of Croatian scientists (n = 541) towards these topics, and to assess the association of their attitudes with their open science practices and demographic information. Overall, Croatian scientists’ attitudes towards these topics were generally neutral, with a median (Md) score of 3.3 out of max 5 on the scale score. We also found no gender (P = 0.995) or field differences (P = 0.523) in their attitudes. However, attitudes of scientist who previously engaged in open peer-review or preprinting were higher than of scientists that did not (Md 3.5 vs. 3.3, P<0.001, and Md 3.6 vs 3.3, P<0.001, respectively). Further research is needed to determine optimal ways of increasing scientists’ attitudes and their open science practices.


Introduction
Open science, despite lacking an universally accepted definition, is widely recognized as a global phenomenon and an initiative emerging from the philosophical concept of scholarly "openness". With the principles and values of openness rooted in the idea of scientific knowledge being a common good [1]. The term open science was coined in 2001 by Recep  i.e., preprints, before those manuscripts were peer reviewed and published in journals (or other venues, such as books or conference proceedings). However, it took a while for prepint servers to become the go to place for researchers. For example, it took arXiv 8 years to become a major player in the dissemination of results in physics and mathematics [21]. Other scholarly fields have been even slower to adapt to the preprint culture: with bioRxiv, a preprint server dedicated to the biological sciences, originating in 2013, SocArXiv a server for preprints in social sciences in 2016, and MedRxiv, a server for clinical research preprints, in June of 2019 [19]. Further actualization and discussions surrounding preprint servers, also rose after Chalmers and Glasziou estimated that 85% of research is wasted due to inadequate research design or methodology, poor reporting, publication bias, and lack of scholarly openness [22], with some viewing preprint serves as a way to address some of these issues. Today, there are more than 60 preprint servers in the world covering all scholarly fields [23], and the number of preprints is rising, fuelled additionally by the COVID-19 pandemic [24]. Preprints are seen as a step toward greater openness of science, and in 2019, Fu and Hughey estimated that manuscripts first published as preprints received 36% more citations and had a 49% higher Altmetric score [25]. Increasing number of journals and funders today encourage preprinting [26]. Furthermore, many scholarly engines have started indexing preprints, e.g. Europe PMC [20], Scopus [18], and Dimensions [27].

Open peer-review
Peer-review is a quality control mechanism for scholarly research or funding proposals. Traditionally, journal peer-review was most commonly blind (single, double or triple blind) and it was often criticised for being slow, expensive, subjective, not able to detect errors, non-reliable, prone to bias and easily abused [28]. This lead to growing need for a more open peer-review process [29]. Open peer-review as a term, however, lacks a universal definition [7]. Most often it used to describe one of the following practices: open identities of the authors and reviewers, open review reports published alongside the article, open interaction and discussion between author(s) and reviewers, or open platforms where a review is facilitated by a different entity than the one where the paper is published [7,24,30]. In our study, we consider open peer review to be open (public) sharing of review reports (with or without reviewers' names) as part of the journal or grant peer review processes. Practice and uptake of open peer review, however has been low, with less than 1% of journals today practicing it [31], and, to the best of our knowledge, no known estimates of its use by funders is available. We are not aware of any studies, that analysed attitudes towards open data, preprinting and peer-review with a validated questionnaire. It was, therefore, our goal to construct and validate such a questionnaire; and use it to report attitudes towards open data, preprinting and open peer-review of Croatian scientists, as well as on the association between their attitudes and open science practices or their demographic information.

Attitudes measurement
Attitudes can be defined as an individual's positive, neutral or negative feelings (evaluative affect) about a certain behaviour or a value [32,33]. Attitudes are often measured with either one-item questions or with multi-item questionnaires (psychometric scales, whose answers are then often summarized to create a scale or an attitude score). While one-item questions can be a useful method for "snapshot measuring" [34], measuring attitudes with only a single question is generally not considered an optimal approach. On the other hand, creation of scales requires rigorous methodological approaches for questionnaire construction and validation [35][36][37][38].
This process often includes steps that include item/question generation, face validity checks, testing scales validity and reliability, and evaluating responsiveness and scale interpretability [35,[39][40][41]. A known fallacy of many attitude assessments is the difficulty to compare research findings, as the same questions or scales are rarely used multiple times or for different populations, and differences between studies can turn out to be a consequence of different wording of questions, emphasizing the need for creation of standardized questionnaires.

Attitudes towards open data, preprinting and open peer-review
We present below our literature review of studies analysing attitudes towards open data, preprinting and open peer-review.

Attitudes towards open data
In a recent (2020) systematic review, Zuiderwijk, Shinde and Jeng summarized results of 32 quantitative and qualitative studies on open data, of which 15 were surveys [42]. They found that "scholars refer to personal drivers and a positive attitude toward data sharing as vital individual drivers for openly sharing research data" and that they see a negative attitude as an inhibitor of data sharing [42]. In those summarized studies participants were mostly from the United States and Europe, and only a small number of studies were focused on multiple scientific disciplines. Most participants also had generally positive attitudes towards open data. An interesting finding was that in half of those studies (which assessed data sharing), there was no reference to studies own data availability [42].
An earlier 1988 study by Ceci described attitudes of 790 researchers from three US universities using a case scenario approach followed by 3 (snap-shot) questions, finding that researchers have a positive attitude towards data sharing, but acknowledging that those attitudes might have been influenced by giving socially desirable answers [43].
Two large studies of data sharing practices, and barriers of data reuse were authored by Tenopir et al. (2011 and 2015) using a multi-question approach, but without validating an attitude scale or reporting its reliability [44,45]. In the first study they surveyed approximately 1200 scientists, of which 900 were followed up in the second study. Most scientists were from North America (68%), and from the fields of environmental sciences and ecology (32%). Their results showed an increase in data sharing attitudes over time, but also an increase in the number of perceived barriers for data sharing [45]. Building on their questions, Curty et al. [13] later validated a scale for measuring attitudes towards data reuse on a sample of 570 scientist. They tested construct validity, and reported a 3 factor construct of their scale: perceived efficiency of data reuse (5 items), perception of data re-use (2 items) and concern about trustworthiness of data (4 items), with subscale reliability scores (Cronbach's alpha) ranging from 0.73 to 0.81 [13].
Yoon and Kim, in 2017, constructed and validated a scale using structural equation modelling (a combination of factor analysis and multiple regression), on a sample of 292 social scientists. Their questionnaire had 20 items and 7 factors (with Cronbach's alpha values ranging from 0.76 to 0.97). They also concluded that attitudes towards data reuse was a strong predictor of data reuse intention [46].
Zenk-Möltgen et al., in 2018, investigated attitudes towards data sharing of 446 political and sociology scientists using a theory of planned behaviour, but they did not report on their scale's validity or reliability. Overall, they found generally positive attitudes toward data and code sharing, and a strong association between previous sharing behaviour and intention to share [47].
Abele-Brehm et al., in 2019, investigated attitudes towards open data and data sharing of 337 psychological society members and reported a 2 factor scale (positive expectations-10 items with Cronbach's alpha of 0.90, and negative expectations-4 items with Cronbach's alpha of 0.67). They found that respondents attitudes were generally positive [48].
Finally, Zhu [12] in 2020, measured attitude of UK researchers towards data reuse with one question ("How important do you think it is, in general, to make research data available online for reuse?") and 1459 out of 1695 (86%) respondents found it to be very or fairly important.  , and although they used multiple questions, they did not report on their questionnaires validity or reliability. In total they collected approximately 3000 responses, mostly of researchers from Europe (61%), and from science, technology and medical (STM) fields (90%). Overall, respondents reported generally positive attitudes towards open peer-review. In 2018, Segado-Boj, Martín-Quevedo and Prieto-Gutiérrez, surveyed authors of Spanish journals (n = 295), mostly from social sciences (63%) with 7 questions, but they did not report on the questionnaires validity or reliability [54]. Overall, participants were found to be cautious towards open peer review. Lastly, in 2020, Besacon et al., using a small sample (N = 30) of researchers in the computer science field, and eight questions, reported that more than half of the respondents were in favour of open peer review, but not of displaying their reviewer names. They did not report the constructs validity or reliability [55].

Materials & methods
We conducted a cross-sectional study with psychometrical validation of a questionnaire, which we named, the Attitudes towards Open data sharing, preprinting, and peer-review (ATOPP).

Participants
In 2018, Croatia had 17,706 scientists [56]. In order to reach as most of them as we could, we sent invitations through 2 different channels: through the mailing list of Croatian scientists (approximately 17,000 members) compiled by the Rudjer Boskovic Institute (Zagreb, Croatia), and the Dean's secretaries of University of Rijeka (the University of the first author, with 1,256 scientists).

Procedure
Participants were invited to fulfil an anonymous online questionnaire (through Google forms). The survey was open from 12 May 2020 to 7 July 2020, and we sent two reminders 14 days apart.
Constructing the questionnaire. The questionnaire was constructed as a result of three focus groups we held at the University of Rijeka in 2019 and 2020 with a total of 24 participants. The first focus group was held with participants from Biomedical Sceinces (N = 12), second with the participants from Social Sciences (N = 7) and the last with participants from Natural Sciences (N = 5). Participants were asked 5 questions: (1) What is open science to you? Answers to attitude statements were offered on a five-point Likert-type scale, where 1 indicated "strongly disagree;" 2 -"disagree;" 3 -"neither agree nor disagree;" 4 -"agree;" and 5 -"strongly agree." Open science practices questions were of mixed type (yes/no and multiplechoice questions). Demographic questions included questions on gender, age, scientific filed, roles in science, and the total number of published papers.
Our initial exploration (factor analysis) of the 45 attitudes questions showed that questions on open access (8 items) and open science tools (6 items) explained less than 5% of the variance of the total score and were not internally consistent (with Cronbach alpha scores <0.65) [35]. We then re-examined them (face validity), and hypothesized this is most likely due to the fact that these two aspects of open science dealt with concepts outside of direct researcher's influence (i.e. they were built by other actors), while data sharing, open peer review, and selfarchiving through preprints were under direct (self-) agency of the researchers. The psychometrical validation of the remaining questions (31 items) is presented in the results.

Statistical analysis
Validation of the ATOPP questionnaire. Construct validity of the scale was tested with exploratory factor analysis after the suitability of the item correlation matrix was checked with the Kaiser-Meyer-Olkin (KMO) measure of sampling adequacy and Bartlett's test of sphericity. In Exploratory Factor Analysis, we used Principal Axis Factoring (PAF) as the factor extraction method and Oblimin as the rotation method. We included the extracted factors with the eigenvalue >1, more than 5% of the construct variance and those which passed visual inspection on the scree plot. Factor loadings <0.30 are not presented [35]. The factor analysis procedure uses the pattern of correlation between questionnaire items, which represent directly measured manifest variables, grouping them by the variance they share which is captured by factors that are interpreted as latent dimensions, inferred constructs that are not directly measured. Consequently, each extracted factor or dimension is defined only by questionnaire items to which it relates [35,58]. Correlations of factors were calculated with Pearson's coefficient of correlation.
Internal consistency of the scale and subscales were determined with Cronbach alpha. Total score. Before calculating the total score we have recorded 4 items: item 6 and 8 in Open data and items 10 and 11 in Open peer-review (S1 Appendix). The total score of whole scale and factors were constructed as a linear composite of all items divided by the 21 (number of items) with the score range being from 1 to 5. Lower results (<2.6) were considered as negative attitude, average (2.6-3.39) as neutral attitude and higher results (>3.39) as positive.
Analysis of answers, based on the ATOPP survey. Qualitative data are presented with frequency and relative frequency. Comparison of qualitative data is done with χ2 test and test of proportion.
Quantitative Sample size calculation. We based our calculation on the number of initial survey attitude questions (n = 45) and the fact that it is considered sufficient for scale validation to have 10 times more participants than the number of items [39].

Ethics
The study was approved by the Ethical committee of the University of Rijeka, Rijeka, Croatia (KLASA: 003-08/19-01/l; URBROJ: 217 0-24-04-3-19-7). In the invite letter we also presented the informed consent form, which the participants had to approve in the online form before starting to fulfil the questionnaire.

Validation of the ATOPP questionnaire
Thirty-one item related to open peer-review (12 items), open data (10 items) and preprinting (9 items) were entered into the exploratory factor analysis after exclusion of 14 items related to open access and open science tools (see Methods above). Acceptability of the construct was then assessed by analysing the floor and ceiling effects of the individual items on the score distribution and no floor and ceiling effects were observed.
Kaiser-Mayer Olkin test (KMO = 0.79) and the Bartlett's test of sphericity (P < 0.001) have satisfied the condition for principal axis factoring (PAF) of the 31 ATOPP questionnaire item. The inspection of the scree plot, Eigenvalues >1 and more than 5% of variance explained yielded 4 factors with 40% of the construct variance explained. We then repeated the factor analysis with 22 items (S1 Fig) that had factor loadings higher than 0.30 [35]. The second PAF analysis was more suitable (KMO = 0.80; Bartlett's test P<0.001) and it resulted with 4 factors (21 items)-Open Data, Preprinting, Open peer review in small scientific communities, and Open Peer-review that accounted with 51% of the construct variance (Table 1).
Structure matrix (correlations of each item with the extracted dimensions) is presented in S1 Appendix- Table 1, indicating 21 items were left in the model with a simple factorial structure (loadings are distributed on one factor exclusively). The reliability of the whole scale was very good (Cronbach's alpha of 0.815).

Participants' characteristics
We have collected 546 responses, 196 (36%) from University in Rijeka and 350 (64%) from the Rudjer Boskovic Institute list of Croatian scientists. There was no overlap between the respondents of the two sources, and 5 responses were not valid (not completed), leaving a total of 541 3. If I have the opportunity to sign a review report I will always sign it. 0.472 3 (3)(4)(5) 4. Reviews of papers that have been rejected should be available to all journals so that reviewers do not repeat the work. responses. The response rate for the University of Rijeka was 15.6% and for the Croatian scientist list it was 2%. Factorial structure of the attitude scales was the same for both samples and therefore we present them together. Median age of the participants was 45 (38 to 53), with equal percentage of both males and females (43% vs 54%, P = 0.082) Majority of the respondents were from Biomedicine and Health (26%), Social (25%) or Natural Sciences (17%). They were most commonly Assistant Professors (29%), Full Professors (27%) or Associate Professor (19%). Most respondents (n = 529, 98%) published at least one article, with a median of 23 (IQR 10-45). More than two thirds (n = 371, 69%) were also reviewers, 16% (n = 87) acted as reviewers for funding agencies, and 11% (n = 62) as members of the editorial board, finally 3% (n = 18) were editors. Detailed demographic and scholarly information of respondents is presented in Table 2.

Open science practices
Respondents' open science practices are presented in Table 3. Around half (47%, n = 240) of the respondents participated in open peer-review and most of them were happy to sign the review reports (n = 225, 95%).
Nearly half of the authors (46%, n = 249) published a paper in a journal in which research data could be deposited, and one third (29.9%, n = 162) published an article based on public data from other researchers. Most respondents shared their data (as supplementary files) via journals (54%, n = 285). Minority of the respondents posted a preprint (12%, n = 64), mostly on Arxiv (n = 38), BiorXiv (n = 12) or SocarXiv (n = 4).

Attitudes towards open data, preprinting, and peer-review
The total score for all participants on the ATOPP scale was neutral with median of 3.3 (3.0-3.7). The neutral score was also found for their attitudes towards preprinting  Table 4). Differences in attitudes were tested regarding gender, field, open science practices and education (Table 4).
We found no gender differences (all P>0.05) except for the open peer-review in the small scientific communities, where female respondents had a more negative attitude than male respondents [2.0(1.0-3.0) vs 2.0(2.0-3.0), P = 0.032].
We also found no differences in the overall ATOPP score between scientific fields (P = 0.523). However, attitudes toward open peer review in small scientific communities were higher in Natural sciences and Technical sciences than in Social Sciences, Humanities and Interdisciplinary fields Participants who had open peer review experience had higher total ATOPP score (P<0.001), as well as attitudes towards open data (P = 0.008) and open peer-review (P<0.001). Similarly, those who previously shared their data had higher attitudes towards Open data (P = 0.007), Preprinting (P = 0.005) and Open peer review in small scientific communities (P = 0.021). Participants with experience in preprinting had more positive attitudes for all subscales (all P<0.05) except for the Open peer review in small scientific communities (P = 0.140). Finally, participants who had education in open science had a more positive ATOPP score then those that did not (<0.001) and they also had higher attitudes for preprinting and open peer review.

Discussion
In this study we developed the ATOPP questionnaire for measuring attitudes toward open data, preprinting and open peer-review. Using the ATOPP questionnaire, we then explored Croatian scientists' attitudes towards those topics and the association of those attitudes with their open science practices and socio-demographic information. To the best of our knowledge, this is the first psychometrically validated (multiple-item) questionnaire for measuring

Position in academia/science (n = 538)
Research Fellow 35 (7) Post Doc researcher 47 (9) Assistant Professor/Scientific associate 156 (29) Associate professor/Higher scientific associate 105 (20) Full professor/Scientific advisor 148 (28) Other 47 (9) Published an article in a scientific journal (n = 541) attitudes towards all these three topics with one questionnaire. The ATOPP scale, consisting of 21 items, demonstrated good internal consistency and validity. Because of its good psychometrical characteristics and relatively small number of questions, we believe that it represents a fast measurement that can be used in assessing or monitoring attitudes towards open science, thus allowing cross-cultural validation. During ATOPP development, attitudes towards open peer review in small scientific communities turned out to be a separate factor (subscale) from attitudes toward open peer review. This could be a product of both the fact that Croatian scientific community for centuries had a higher number of specialized journals per capita compared to its neighbouring countries, and the fact that open peer review in small (national) fields or subfields has higher likelihood of reviewers being direct competitors for funding or job positions [58]. Additionally, smaller communities may experience greater fear of negative consequences of open peer review, i.e., fear of a potential (vindictive) backlash of their colleagues if they criticize their work, or if due to their review, they negatively affected funding or publication opportunities of their colleagues. And these fears will likely remain until external evaluations or strong protective mechanisms are put into practice (however, for small communities these approaches likely face significant language barriers and high costs). Based on the ATOPP questionnaire, we then found that the Croatian scientists had generally neutral attitudes toward open science. Their most positive attitudes were towards open data, while their attitude towards preprinting and towards open peer-review were neutral, and those towards open peer review in smaller scientific communities were negative. We also found no gender or scholarly field differences in respondents' overall attitude scores. However, scientists who already had experience with open science practices, i.e., shared data, provided open peer review reports in the past, or posted preprints, had generally more positive attitudes previous open science education are in accordance with Bem's self-perception theory that confirms effect of past behavior on internal attitude [59]. Past behaviour influence on attitudes was also confirmed by many researchers since then [60,61]. We have also found that participants who had taken open science courses had more positive ATOPP scale score, preprinting score and open peer-review score. which is a confirmation of a model that positive attitudes are related to intention to behave and behavior [32]. The positive attitudes towards open data in our study were associated with the high prevalence of researchers in our sample (46%) that shared their data in the past. However, in the recent survey by Zhu (2020) in the United Kingdom on 1724 participants from various scholarly fields, there were less participants (21%) who had deposited primary data in online repositories, although majority (86%) had a very positive attitude towards data sharing [12]. However, in that survey, attitude was only measured by a single question "How important do you think it is, in general, to make research data available online for reuse?".
Positive attitude towards open data in our study can be compared with the data in a recent survey among members of the German psychological society (N = 303). Abele-Brehm et al. constructed a scale measuring hopes (10 items, Cronbach α = 0.90) and fears (4 items, Cronbach α = 0.67) towards data sharing. The positive attitude-"hopes" of respondents were neutral, but their experience with data sharing was not measured [48]. Yoon and Kim (2017) investigated data reuse behaviour by measuring beliefs, attitudes, and norms. Based on their theoretical framework attitude was a strong positive predictor of data reuse [46].
Attitudes of Croatian scientist towards preprinting were neutral in our study, except of those scientists who preprinted in the past (12%). These results differ from attitudes participants in South Korea [50], China [49], and Latin America [62] whose attitudes were overall found to be positive; but those surveys included more editors, and all (China), or many respondents who previously posted a preprint (Korea (32% editors, 15% past preprint users; and Latin America, 40% past preprint users), while in our sample only 12% of scientists did so and we had only 11% of editorial board members in the sample [49,50]. Additionally, China introduced a country preprint server ChinaXiv in 2006, and Latin America in 2020 (ScIELO preprints), which most likely further promoted already strong open access culture in those countries (Croatia does not have a national preprint server). Croatian scientist preprint use in our study, is in the line with a large analysis of Biorxiv preprints (n = 67,885, in the period from 2013 to 2019), which found that senior authors of preprints are often researchers from the United States (39.2%) and the United Kingdom (10.5%), while Croatia was described as a "contributor country" whose authors were rarely on senior authorship positions [63]. More studies are, however, needed to determine the main factors that drive researchers to start preprinting manuscripts or project proposals (protocols), as well as inviting or choosing to wait for public comments before deciding to submit the preprint for scholarly journal peer review. Additionally, previous survey has shown that scientist choices toward posting a preprint are influenced by the policies of the journals in which they plan to publish those studies [62,64]. Although Croatia has approximately 400 active scholarly journals [65] of which less than half are indexed in WoS or Scopus [66], preprint policies are listed for only 24 in Sherpa website, and so further research is needed to determine the influence of Croatia's editorial, funder and publishing milieu on the preprint attitudes of its researchers.
In our study, Croatian scientists' overall attitudes towards open peer review were neutral. We also found differences among scientific fields with scientists from Biomedicine and Health and Biotechnical sciences having higher attitude score (albeit still neutral), as did those with previous experience with open peer review. In Ross-Hellauer, Deppe and Schmidt 2017 study on 3062 participants, most thought that open peer-review should be common practice, with the strongest support coming from social science researchers (e.g. economics, psychology and philosophy). Also, the majority of researchers agreed that the obligatory signing of review reports is strongly associated with rejecting peer review requests [53]. That study however did not report on the validity and reliability of its questionnaire.
Attitude towards open peer-review in small scientific communities of Croatian scientist were low and even lower than attitudes towards open peer-review, indicating that Croatian scientific community might not be ready for this aspect of open science yet. Female scientists also had a lower attitude than male scientists, which could be the results of the gender disbalance in academia, and greater fear of retaliation or promotion obstruction [67]. Slightly higher attitudes towards open peer-review in small scientific communities was found among scientists who shared data before, preprinted, or were from natural and technical sciences; which are fields that have been sharing preprints the longest.
Despite presenting the first psychometrically validated scale for measuring attitudes towards open data, preprinting and open peer-review, our study is not without limitations. As all questionnaires, are data are based on self-declared attitudes and open science practices and does not capture independently confirmed practices. Furthermore, as in many recent online surveys, our response rates were low, and that rate might have also been influenced by the fact that the questionnaire was sent during the early months of the COVID-19 pandemic. Additionally, we might have captured opinions only of those interested in these topics, which, if true, could mean that the attitudes of a representative sample of Croatian scientists would be even lower towards these topics. While we did provide definitions of open science practices in our questionnaire, as most of our respondents (78%) did not have education in open science, it is possible some held different ideas of those practices. Finally, while our study, showed a strong association between open sciences attitudes and previous open science practices, and in that way provides further credibility to the ATOPP questionnaire, our study was cross-sectional and was not designed to look at the possible consequences of these attitudes on promotion and implementation of open science practices. Further interventional research is needed to explore the most efficient interventions in increasing positive researchers' attitudes toward open science practices, as well as if those interventions would lead to greater uptake of such practices [68]. Furthermore, with the recent changes in the EU funding schemes for the period of 2021 to 2027, and the requirement for open peer review and data sharing [69]; announcement of the journal eLife for only accepting submissions if they have been posted as preprints [70], as well as dedicated calls for research into ways to increase open sciences practices of those who have not embraced them yet [71] it will be interesting to compare if approaches aimed at rewarding open science practices vs mandating of those practices by several key stakeholders, will have a different impact on researchers attitudes. And, ultimately, which of those approaches will be more effective in inducing changes in the wider scholarly community.

Conclusions
In conclusion, our study presents the validation of a multi-item questionnaire for measuring open science attitudes, specifically open data, preprinting and open peer review. Additionally, using the questionnaire we found that attitudes Croatian researchers towards these topics were neutral, and that more positive attitudes were found among those that participated in open science practices before or had an education in open science. Further studies are needed to assess attitudes of researchers on these topics in other countries, as well as to track changes of these attitudes over time. With more and more funders and institutions encouraging or mandating open science practices, we believe that validated tools, such as this one, could help assess and monitor researchers' attitudes and their associations with open science practices.