Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

A global exploratory comparison of country self-citations 1996-2019

  • Alberto Baccini ,

    Contributed equally to this work with: Alberto Baccini, Eugenio Petrovich

    Roles Conceptualization, Data curation, Formal analysis, Funding acquisition, Investigation, Methodology, Supervision, Validation, Visualization, Writing – original draft, Writing – review & editing

    alberto.baccini@unisi.it

    Affiliation Department of Economics and Statistics, University of Siena, Siena, Italy

  • Eugenio Petrovich

    Contributed equally to this work with: Alberto Baccini, Eugenio Petrovich

    Roles Conceptualization, Data curation, Formal analysis, Investigation, Methodology, Validation, Visualization, Writing – original draft, Writing – review & editing

    Affiliation Department of Philosophy and Education Sciences, University of Turin, Turin, Italy

Abstract

Self-citations are a key topic in evaluative bibliometrics because they can artificially inflate citation-related performance indicators. Recently, self-citations defined at the largest scale, i.e., country self-citations, have started to attract the attention of researchers and policymakers. According to a recent research, in fact, the anomalous trends in the country self-citation rates of some countries, such as Italy, have been induced by the distorting effect of citation metrics-centered science policies. In the present study, we investigate the trends of country self-citations in 50 countries over the world in the period 1996-2019 using Scopus data. Results show that for most countries country self-citations have decreased over time. 12 countries (Colombia, Egypt, Indonesia, Iran, Italy, Malaysia, Pakistan, Romania, Russian Federation, Saudi Arabia, Thailand, and Ukraine), however, exhibit different behavior, with anomalous trends of self-citations. We argue that these anomalies should be attributed to the aggressive science policies adopted by these countries in recent years, which are all characterized by direct or indirect incentives for citations. Our analysis confirms that when bibliometric indicators are integrated into systems of incentives, they are capable of affecting rapidly and visibly the citation behavior of entire countries.

Introduction

Since the early times of citation indexes, self-citations have attracted the attention of bibliometricians [1, 2]. In evaluative bibliometrics, the main concern with self-citations is that they can potentially inflate impact metrics or distort their meaning [36]. It has thus been debated whether they should be removed from citation indicators [710]. In descriptive bibliometrics, on the other hand, self-citations have been studied from the point of view of scholarly communication. The motivations for self-citing have been classified [11, 12] and self-citations have been used to investigate how scientific authors relate with their own production [13, 14].

In the most general sense, self-citation occurs when an entity (e.g., an author, journal, institution, or country) receives a citation from a publication produced by the same entity [15]. Even if all self-citations derive from the act of self-referencing, self-citations and self-references should be distinguished [16]. For technical reasons shortly described below, the results derived from observing self-citations or self-references are different and so is their interpretation. This paper focuses on country self-citations.

Depending on the entity considered, self-citation can be classified into different types [6]. The most basic is the author self-citation, which occurs when the publications written by an author are cited in the following publications by the same author. For multi-authored publications, author self-citation can be defined narrowly or broadly, i.e., including or not citations generated by co-authors. Author self-citations intended in the extensive sense are sometimes called co-author self-citations [17] or all-author to all-author self-citations [6]. Journal self-citation occurs when an article published in a certain journal is cited by a subsequent publication in the same journal [18]; this form of self-citation has been mainly studied to understand how it can influence or even manipulate the journal Impact Factor [1921]. Institution self-citation happens when the authors of the cited and of the citing publications share the same affiliation [22, 23]. By extending rather inappropriately the notion of self-citation, language self-citation refers to citations occurring among publications in the same language [24], and field self-citation for citations between publications belonging to the same academic field [25].

In the context of the study of the scientific performance of countries, the self-citations defined at the highest level of aggregation, i.e., country self-citations, have recently started to attract some attention, both from researchers [2631] and in science policy reports [32]. A country self-citation, also called sometimes domestic citation (e.g., [33]), occurs when the publications produced by the researchers of a country are subsequently cited by researchers of the same country.

In this study, the trends of country self-citations in 50 countries worldwide are investigated in order to reveal groups of countries characterized by similar self-citation behavior in time. The main interest consists in individuating countries that deviate from standard trends because these anomalies may signal perturbations in the scientific development possibly induced by science policies. Italy is a case in point in this sense, as previous research has revealed an anomalous rise in the country’s self-citations after the introduction of pervasive bibliometric evaluation [26, 3438].

The paper is structured as follows. In the next section, the bibliometric indicators based on country self-citations are reviewed. Then, the indicators, data, and analytical methods used in this study (time-series analysis) are presented. The main findings are shown in the Results section, whereas, in the Discussion section, we focus on countries characterized by anomalous self-citation trends. Based on the detailed reconstruction of the research policies adopted by these countries, it is argued that the anomalous trends are most likely explained by the adaptive response of scientists to the systems of incentives established by the policies themselves. Accordingly, Conclusions suggest managing bibliometric indicators in research policy contexts with extreme caution.

Review of the main indicators based on country self-citations

As anticipated, self-citations and self-references are defined, computed, and interpreted in different ways. To highlight these differences [39], for example, called the former diachronous or prospective self-citations and the latter synchronous or retrospective self-citations.

Self-citations can be defined as citations from citing sources to cited items that are both produced by (at least) the same entity E, i.e. the same author, journal, institution, or country. Different ways of defining these entities entail the use of different algorithms, which will be discussed in the next paragraph.

The computation of self-citations requires the definition of: (i) a publication window delimiting the cited items as the ones published in it. For the sake of simplicity, in what follows, the publication window is set at one year, hence the cited items considered are only the ones published in the year y; and (ii) a citation window delimiting the citing sources as the ones published in it. Let y be the reference year for calculation and also the publication window; T is the length of the observation period, expressed in years. Self-citations of an entity E in the year y can be defined as: (1) where sE(y, y − 1 + i) is the number of self-citations received by the set of cited items published by the entity E in the publication window y, from citing sources produced by the same entity E in the citation window (y − 1 + i). The citation window includes the years (y − 1 + i) for i = k, …, T, where k ∈ {1, …, T} is chosen by the user for setting the desired time interval for observing citations. If k > 1 citation window and publication window are disjoint; for k = 1 and T > 1 they are partially overlapped, and for k = T = 1 they are completely overlapped. In the first case, citing sources are published in years following the publication windows; in the second case, the citation window includes the publication window and the following years; in the third one, citation window and publication window coincide. Analogously, the total citations received by the set of cited items published by the entity E in the year y can be defined as: (2) where cE(y, y − 1 + i) is the number of citations received by the set of cited items published by E in the publication window y, from citing sources published in the citation window y − 1 + i.

The basic indicator is the self-citation rate (SRE, y) for entity E in the year y, defined as the ratio between the self-citations and the total number of citations received by E [26, 28, 40, 41]: (3) where SRE,y ∈ [0, 1]. CE,y is usually interpreted as a proxy of the academic impact of the entity E. SE,y is an indicator of academic impact generated by self-citations of E. Hence, a self-citation rate can be interpreted as the proportion of the academic impact of E due to its self-citation activity. It should be noted, that the denominator of the ratio is concretely produced by the citing choices of the whole scholarly community, and the ratio relates the self-citing choices of E with respect to the citing choices of the whole scholarly community.

Self-references indicate as well citations from citing sources to cited items that are both produced by (at least) the same entity, but the sets of citing sources and cited items are defined in different ways with respect to self-citations. Moreover, the computation of self-reference requires only a publication window for the set of citing sources. If the publication window is set also at one year, the cited sources are the publications produced by E in the year y, and cited items are (all or part of) the items listed in the bibliographies of the citing sources. SBE,y, i.e. the number of self-references in the year y, is computed by summing up the citations from citing sources published in the year y by E to cited items previously produced by E. Analogously, RE,y indicates the total number of citations given (i.e., references) by E in the year y to documents published before. The self-reference rate SBRE,y for the year y is thus defined as: (4)

Note that a rigorous definition of this indicator would require the specification of a reference window, analogous to the citation window used in the self-citation rate. The studies based on the self-reference rate (e.g., [16]) usually assume a reference window that is extended to the oldest publication year of the references considered, but, in principle, other reference windows are possible.

When references are observed, the focus of the analysis is on the use of previous knowledge by the entity E, and not on the academic impact of the scientific production of E. It should be noted that the denominator of the ratio is, in this case, produced also by the citing choices of E. SBRE should be interpreted as an indicator of how much E uses in its current work the knowledge that it had previously produced. For instance, for an author, a relatively high value of SBRE,y may indicate that her/his work is largely based on her/his previous work. This in turn may be due to scarce attention to the work of other researchers in the field or to the fact that the author works in a little and specialized niche [39]. A relatively high value of SRE,y, instead, indicates that a relatively high share of her/his academic impact is due to her/his self-citation activity.

Note that, in general, authors have more control over their self-reference rate than their self-citation rate, as they can manage more easily the citations they give (i.e., the denominator of the self-reference rate), than the citations that they receive (i.e., the denominator of the self-citation rate). In fact, authors interested in artificially inflating their citation metrics can effectively conceal strategic self-references within long reference lists and keep their self-reference rate low. Still, if they are not able to attract enough external citations to offset the impact of their strategic self-references, their self-citation rate will inevitably rise. This shows that the self-citation rate is more effective in revealing strategic self-citation behavior than the self-reference rate.

The basic indicator in the literature on country self-citations is the country self-citation rate. Hereafter N denotes the entity “country”; the self-citation rate of country N in the year y is the ratio between its country self-citations and the total number of citations received by that country [26, 28, 40, 41]: (5) where SN,y is the raw number of self-citations of country N in the year y and CN,y is the total number of citations received by N in the year y.

Computing the total of country self-citations for a country SN,y is, however, less obvious than it may seem. As detailed in Methods and data, country self-citations can be in fact computed either narrowly or broadly, depending on how citations to international publications are considered. In [26], a variant of the self-citation rate indicator called “inwardness” is developed, characterized by a citation window variable in length. A broad definition of country self-citation is adopted, according to which a citation is considered a country self-citation when the intersection between the set of the countries of affiliation of the author(s) of the cited publication, and the set of the countries of affiliation of the author(s) of the citing publication, is not empty. This broad definition of country self-citation has the desirable property of ascribing to the world an inwardness value of inwardness 1, which makes therefore the inwardness an indicator normalized for the size of the country in terms of publications (see [26] for details). This property, however, does not hold for the SRN,y when the country self-citations are computed narrowly (see Sec. Methods and data).

Inwardness is interpreted as a measure of the self-referentiality of a country: a higher level of inwardness suggests that the scientific publications produced by a country attract mainly the interest of the national community, whereas a lower level of inwardness suggests that the scientific production is cited mainly abroad. In the same sense, [42] related country self-citations to the degree of scientific insularity of nations. [42] also suggested several factors that may explain why developing countries show higher self-citation rates, among which a focus on applied scientific issues that respond to the perceived needs of national development, poor referencing practice, insufficient training of graduate students, preference for literature in national language than in English, and the proliferation of low-quality national journals. As [26] have shown, however, also developed countries may show anomalous raises in self-citation rates induced by research evaluation policies that reward raw citation metrics.

A variant of the self-citation rate is the citation domesticity indicator [43], in which citations coming from international publications are evenly distributed among the citing countries. Apart from [43], however, fractional counting of country self-citations has never been considered in the following literature. This may be explained by the fact that most more recent studies collect data on country self-citations from Elsevier’s SCImago or SciVal platforms, which do not apply fractional counting [44].

The complement of the self-citation rate is the foreign citation rate FRN,y, also called “international scholarly impact of scientific research” by [45]: (6)

The country self-citation rate results to be positively correlated with the publication output of a country. In particular, [30] proposed a model where the country self-citation rate increases with the logarithm of the output: (7) where PN,y is the number of publications of country N in the year y. This occurs because bigger countries have more domestic papers to cite and, hence, are more likely to attract citations from their own researchers than smaller countries [40, 41, 46]. By contrast, the average number of citations per document of a country is negatively correlated with self-citation rates [29]. Self-citation rates have increased over time: according to [29] estimates, the average self-citation rate of 62 countries raised of 28.9% from 1996 to 2009. [26] noted as well that the Inwardness of G10 countries increased during the period 2000–2016 with a mean increase of 5.2 percentage points.

To correct from the size-dependency of self-citation rates, [40] proposed to compare the self-citation rates with the world-share of publication of a country, based on the idea that if the publications from a country are cited as expected, then its share of country self-citations is proportional to its share of world publications. More recently, this indicator has been called “over-citation ratio” (OCRN,y) by [27] and is defined as: (8) where Pw,y is the total number of publications in the world and αN,y the proportion of publications of country N in the world. An over-citation ratio higher than 1 means that the country receives more citations from its own publications than expected based on its relative weight in the world scientific production. At the field level, [27] found that, as expected, over-citation ratio is higher for scientific fields of more national interest.

However, the over-citation ratio results to be size-dependent, as it is highly influenced by the denominator in the formula, i.e., the fraction of papers published by a country. For countries with high weight in the world publication output, such as the USA, the OCRN,y will be always smaller than for small countries. For instance, for a country publishing one-third of the papers in the world (αN,y = 33%), the OCRN,y can never exceed a value of 1/0.33 ≈ 3, whereas a small country that published αN,y = 3.3% of world papers, the maximum value of the OCRN,y raises to 30 [24, 41]. In fact, [27] found that there is a negative power correlation between OCRN,y and αN,y: (9)

A further indicator, based on probability ratio, has thus been proposed, first in the study of language self-citations by [46] and then adapted to country self-citations by [41]. This indicator, called “odds-ratio” (ON,y), relates two ratios: the numerator is the ratio of country self-citations to foreign citations, and the denominator is the ratio of domestic publication proportion to foreign publication proportion: (10)

The odds-ratio ON,y measures to what extent the country relative preference to cite its own publications is greater or smaller than the existing ratio of its domestic publications to publications from other countries. Note that for small values of SRN,y and αN,y, the odds-ratio approaches the over-citation ratio.

The odds-ratio has three drawbacks as well, however [24]. First, if a country cites only its own publication, the measure is infinite. Second, it is oversensitive to small variations in the SRN,y. Third, it is not normalized between 0 and 1. To fix these issues, [24] proposed the following indicator of relative self-citation rate: (11)

According to their interpretation, this formula considers the publication proportion αN,y as a stimulus to the publication-citation system and the relative self-citation rate EN,y as the subjective reaction of the system, which depends logarithmically on the intensity of the stimulus, as in Weber-Fechner equation. The function also expresses the law of diminishing returns: the larger αN,y, the less important the changes in the relative self-citation rate. The relative self-citation rate has the advantages of being normalized and size-independent. Besides its mathematical merits, however, its meaning is less transparent compared to all the previous alternatives.

Methods and data

As anticipated, this work handles time series analysis of country self-citations: time-series clustering techniques are used for detecting countries whose self-citation behavior is similar [47, 48]. To build the time-series, a preliminary distinction between extensive (or broad) and restrictive (or narrow) country self-citations is introduced. The two counts are conceptually different and generate different estimates of country self-citations. Hence, two self-citation indicators are defined.

The relevant data from a citation database for the countries of interest are next retrieved and the time-series generated. The distance between the time-series is then calculated using a suitable distance measure and the structure of the distance matrix thus obtained is explored using multi-dimensional scaling (MDS). In the resulting MDS maps, countries characterized by similar trends will be placed close to each other, whereas countries with different trends far away [49].

Fig 1 sums up the phases of the present study. In the next paragraphs, the methodological and technical choices taken in each step are presented in detail and justified.

Country self-citations of Type I and II

As said above, country self-citations can be computed in different ways depending on how self-citations to international publications, i.e. publications with authors from different countries, are considered. Analogously to author self-citations, the count of country self-citations for international publications can be done by adopting a publication-based or a author-based perspective [34].

The extensive publication-based perspective, adopted in the Inwardness indicator and implemented in the SciVal database, considers as country self-citations all citations coming from the collaborating countries. In the following, these extensively-intended country self-citations are referred to as country self-citation of Type I (SRI), by omitting for simplicity the indexes of the publication window y and of the country N. The second way of counting country self-citations is author-based and more restrictive. Country self-citations of Type II (SRII) are computed by considering only national author self-citations: a publication produced by (at least) an author of a given country receives a country self-citation of Type II if the citing publication is authored by one of the authors of the cited one, and this author is affiliated with the considered country.

The example in Fig 2 clarifies the computation of the two types. The Figure shows a citation network comprising publications labeled as 1, 2, …, 8, of which 3 are cited items and 5 are citing sources, and 8 citations indicated by the arrows linking the pairs of cited and the citing publications: (1, 4), (1, 5), …, (3, 8). The publications are authored by 5 authors labeled A, B, C, D, E from 3 countries: Italy (IT), Netherlands (NL) and Canada (CA). Inside each node, the authors of the relative publication and their countries of affiliation are reported by letters and acronyms. Three statistics for the two cited countries (Italy and Netherlands) are then calculated (Table 1): the total number of country citations, the total number of country self-citations Type I, and the total number of country self-citations Type II. For country citations, it is sufficient to count the citations landing on the publications produced by each country: the 3 Italian publications get 8 citations and the 1 Dutch publication 5 citations. For country self-citations Type I, we must compare for each citation the set of countries of the citing publications with that of the cited publications: when the intersection is not empty, the citation counts as a country self-citation Type I. Thus, Italy collects 7 country self-citations Type I, the Netherlands 4. Note, however, that the Dutch-Italian publication 1 receives only 1 citation from a publication with a Dutch author (publication 5), whereas the other three citing publications 6, 7, and 8 are in fact from Italian authors. This happens because country self-citations Type I of international publications include the citations coming from any of the collaborating country. In fact, a paper resulting from the collaboration of numerous countries is more likely to attract self-citations of Type I in comparison to a paper resulting from the collaboration of only a few countries. This is because there is a higher likelihood that the countries of the citing papers will overlap with the countries of the former rather than with the countries of the latter (see [26], sec. 3). In the extreme hypothetical scenario of a paper resulting from the collaboration of all countries in the world, any citation to this paper would be considered a self-citation of Type I, as the country of the citing papers will always overlap with at least one of the countries of the cited paper.

thumbnail
Fig 2. Toy citation network.

Nodes represent publications and arrows citation links. The authors of each publication, represented by capital letters, and their affiliation country, represented by acronyms, are shown inside each publication node.

https://doi.org/10.1371/journal.pone.0294669.g002

thumbnail
Table 1. Classification of citations in Fig 2 based on the two definitions of country self-citation.

No fractional counting is applied.

https://doi.org/10.1371/journal.pone.0294669.t001

Country self-citations of Type II are introduced precisely to correct this somehow counter-intuitive property of country self-citations Type I. Country self-citations of Type II in fact include only the citations coming from the focal country. Then, if the focal country is the Netherlands, only the citation-link (1, 5) in the example will count as a country self-citation Type II for the Netherlands because it links two publications sharing the same Dutch author B. Citations (1, 6) and (1, 8), by contrast, will not count because they are generated by the Italian co-author of B. Symmetrically, citation (1, 5) will not count for Italy, whereas (1, 6) and (1, 8) will. Table 1 shows analytically in which relationship each citation-link stands with the two countries and reports the three statistics for each country. Note that no fractional counting was applied: citations from multi-author or international publications or landing on multi-author or international publications are not divided among the cited or citing countries.

Based on the previous definitions, the difference between the country self-citations of Type I and Type II of a country should be attributed mainly to the publications the country produces with other countries, i.e., to its international publications. A wider difference means that these international collaborations are widely cited not only by the focal country but also by the collaborating countries, increasing Type I but not Type II self-citations. A shorter difference, by contrast, means that the country has few international collaborations and/or that they do not attract citations from the collaborating countries. The difference between Type I and Type II is therefore related to the international citation impact of a country.

Part of the difference between the country self-citations of Type I and Type II of a country can be attributed also to citations generated by country publications that are not also author self-citations. In particular, the country self-citations of Type II do not capture citation exchanges between groups of national authors who cite each other but do not collaborate directly on writing papers.

Indicator design

The review of the literature above presented several indicators based on country self-citations. However, only the self-citation rate SR is a pure citation indicator. The others integrate in different ways publication counts to correct for the size-dependency of the SR.

In static analyses of country self-citations, indicators encapsulating both publications and country self-citations are useful, but they become problematic when the temporal dimension is considered. In fact, the trend of the publications-plus-self-citations indicators results to be affected both by the publication trend and the self-citation trend. The indicators, however, cannot say which of them is the driver. Consider, for instance, the over-citation ratio: an increasing over-citation ratio over time may be caused both by a raise in the self-citation rate with the share of country publications remaining stable or by a decrease in the share of country publications with the self-citation rate remaining stable. The indicator as such does not say which of the two dynamics has taken place. Moreover, the share of country publications depends on the publication activities of other countries as well: a country experiencing a rise in scientific productivity, like China, causes an automatic reduction in the publication shares of the other countries, even if their productivity has remained steady. Therefore, too many dynamics affect the trends of the over-citation ratio. An analogous reasoning applies to the other mixed indicators.

Since the focus of this paper is on country self-citation dynamics, the most suitable indicator is the self-citation ratio only. Specifically, two variants of the indicator will be computed: one based on country self-citations Type I (extensive self-citation ratio, SRI) and one based on country self-citations Type II (restrictive self-citation ratio, SRII). Even if country self-citations of Type I are affected by citations coming from other countries, as seen above, they remain interesting because they allow capturing, among other things, the effect of citation clubs at the country level, i.e., the strategic exchanges of citations between researchers in the same country that are not co-authors [26]. The SRI, like the Inwardness, is a size-independent measure of country self-citation, since it is normalized for the size of country publications; SRII is instead not normalized.

The last step in the design of the indicator is the definition of an appropriate citation window. A citation window is required to correct for the fact that older publications have more time to accumulate citations. In evaluative bibliometrics, longer citation windows are sometimes recommended to capture the delayed impact of publications [50]. For our study, however, a long citation window has two shortcomings. The first one is that the larger the citation window, the more observations are lost in the time series since, for the more recent years, the citation window is not complete.

The second one is a smoothing effect on perturbations. Imagine an anomalous peak in country self-citations due to a nationwide policy change, which determines an amount of 80 country self-citations Type II at year y0. Imagine that the policy is afterward dismissed so that the country self-citations return to “normal” values of 40 in the following years y1 and y2. Lastly, stipulate that the country total citations have remained stable at 100 for all three years. Now, if we use a 1-year citation window, i.e., we count only the citations coming from publications that appeared in the same year of the cited publications, the SRII for the three years will result to be, respectively, 0.8, 0.4, and 0.4. With this citation window, the peak at year yo is clearly visible. With a 2-year citation window, which sums together the self-citations of the cited years and those of the next year, by contrast, the values will be 0.6 and 0.4 (note the value for cited year y2 can no longer be calculated). Finally, with a 3-year citation window, the SRII at year y0 will result to be 0.53 (no further values can be computed). The aggregation of the anomalous year with the normal ones has reduced the visibility of the anomaly. With a sufficiently long citation window, the anomaly might even disappear, being absorbed by the average trend.

The yearly Type I or Type II self-citation ratios for country N in the year y are computed by setting a 2-year citation window as follows: (12)

Therefore, for a country N, the number of citations and self-citations Type I/Type II collected in a given year y are computed by considering citations and self-citations Type I/Type II received in the year y by the cited items published in the year y from the citing sources published in the years y and y + 1. This choice permits reducing at a minimum of one year the loss of observations in each time-series; moreover, it permits limiting the smoothing effect and highlighting anomalies in trends. To check for the effect of the citation window, i.e., to check whether the final results were affected by the choice of the citation window, all the analyses were repeated by using also 1-year and 5-year citation windows, by mimicking the usual citation windows of journal impact factors [51]. These results, which are qualitatively similar to those reported in the paper, can be found in the Supporting Information.

Data

Scopus data were used for the computation of the indicators and data were provided by Elsevier through its ICSR Lab. The identification of author self-citations used in the calculation of Type II country self-citations relies on the Scopus author profiling procedure described in [52].

Countries with at least 100,000 publications indexed in the Scopus databases in the period 1996–2019 were considered (n = 50). The indicators were computed for each country on a yearly basis from 1996 to 2019 using, as said above, a 2-year citation window. Publications from all Scopus fields were aggregated, i.e., for each country, the entire scientific output was considered, with no distinction of the research area. Thus, for each indicator, 50 time-series (1 per country) with 24 observations each (1 per year) were computed. In the Supporting Information, the trends of the over-citation ratio, odds-ratio, and relative self-citation rate are available as well. All were calculated using both Types of country self-citations.

Tables 2 and 3 provides the descriptive statistics of the countries in the dataset.

thumbnail
Table 2. Descriptive statistics 1996–2019.

The overline indicates the mean. Publications include only research articles, reviews, and conference papers. Citations () and self-citations ( and ) are computed on a 2-year citation window. Citing sources include all types of documents, cited publications include research articles, reviews, and conference papers.

https://doi.org/10.1371/journal.pone.0294669.t002

thumbnail
Table 3. Average country self-citations of type I and type II and their average difference.

All values are multiplied by 100.

https://doi.org/10.1371/journal.pone.0294669.t003

Comparing time-series

In the literature on time-series analysis, numerous measures for comparing time-series have been developed (reviews can be found in [47, 48, 53]). The various measures encapsulate different senses in which two time-series may be similar or dissimilar. Choosing the suitable measure depends both on the nature of the data and the purposes of the analysis.

In the present setting, the measure should satisfy three conditions. First, it should not be sensible to the mere magnitude of the difference between the self-citation ratios, since self-citation ratios are partly affected by the size of the country: bigger countries tend to have higher self-citation ratios. This excludes all measures based on the point-wise distance between the time-series. Second, the measure should be sensitive to trends and changes in trends of self-citation ratios, as these events may be associated with external perturbations, such as policy changes, useful in explaining the phenomenon. Third, the measure should not assume any underlying statistical model for self-citation ratios, in order to avoid unjustified assumptions on the dynamics of the self-citations over time.

The following dissimilarity measure based on Pearson’s correlation satisfies all the three requirements specified above: (13) where X = (X1, …, Xn), Y = (Y1, …, Yn) are the time-series considered and ρ(X, Y) is the Pearson correlation index.

This measure was proposed originally by [54] and implemented in the function diss.COR of the package TSclusts [53] for [55]. The measure is bounded between 0, when there is a perfect correlation between the time-series, and 2, where there is perfect anti-correlation between them. When the two series show no correlation at all, the value is .

All the dissimilarities between pairs of country trends can be arranged in a dissimilarity matrix. This in turn can be visualized using Kruskal’s Non-metric Multidimensional Scaling, one form of non-metric MDS which respects the ranking of dissimilarities rather than their absolute values [49]. The function isoMDS of the package MASS [56] in can be used to produce MDS maps.

Results

Fig 3 shows the trends of the 50 countries for the two indicators SRI and SRII. Three main observations can be made on these trends. First, the proportion of country self-citations of both types has decreased over time in most countries, following a linear pattern with more or less pronounced oscillations depending on the country. Three countries, however, deviate from this general behavior: in the case of Indonesia, Ukraine, and the Russian Federation, in fact, the trends of both indicators show an inversion from descending to ascending.

thumbnail
Fig 3. Country self-citation rate SRI and SRII by country.

Yearly data 1996–2019.

https://doi.org/10.1371/journal.pone.0294669.g003

Second, the SRI and SRII trends are highly correlated for most of the countries, with both indicators following similar trajectories over time. Again, there is a notable exception represented by China, which is the only country where the share of country self-citations of Type I has surged (China SRI has increased by 24.1 p.p. between 1996 and 2018) while that of self-citations of Type II has substantially contracted (SRII has decreased by 22 p.p. in the same period).

Third, the difference between SRI and SRII varies over time differently depending on the country (Fig 4 and Table 2).

thumbnail
Fig 4. Simple difference between SRI and SRII over time by country.

The solid red line represents yearly data (1996–2019), and the dotted black line is the average difference in the whole period.

https://doi.org/10.1371/journal.pone.0294669.g004

In particular, we can distinguish three groups of countries based on how the difference develops over time. The first group is characterized by an increasing difference, with the two curves of SRI and SRII progressively diverging. It includes Brazil, Chile, China, Colombia, Egypt, Hungary, India, Indonesia, Iran, Malaysia, Pakistan, Romania, Russian Federation, Slovakia, Turkey, and Ukraine. The second group shows a stable difference, meaning that the two curves follow parallel directions. It includes Argentina, Australia, Austria, Belgium, Canada, Croatia, Czech Republic, Denmark, Finland, France, Germany, Greece, Hong Kong, Ireland, Israel, Italy, Japan, Mexico, Netherlands, New Zealand, Norway, Poland, Portugal, Saudi Arabia, Singapore, South Africa, South Korea, Spain, Sweden, Switzerland, Taiwan, Thailand, and the United Kingdom. The last group includes only one country, the United States, where the difference reduces over time, i.e., the curves tend to converge. Note that all G10 countries, apart from the United States, belong to the second group and are all characterized by a significant difference between SRI and SRII, with the United States showing the highest difference of all countries considered (mean difference = 34.4).

Both indicators result to be highly correlated with the logarithm of the publication output for most of the countries (Fig 5). Notably, the relationship between the self-citation rate (of both Types) and scientific production is of inverse proportionality: across most of the countries observed, an increase in scientific output is associated with a decrease in the self-citation rate. This contrasts with the direct proportionality found in prior studies [30, 40, 41, 46]. Interestingly, however, a few countries deviate from this general pattern, exhibiting either positive correlation (China for SRI, Indonesia on both SRI and SRII) or mild to low correlation coefficients (Colombia, Egypt, India, Iran, Italy, Malaysia, Pakistan, Romania, Russian Federation, Saudi Arabia, Thailand, Ukraine). The self-citation rate trend within this latter group of anomalous countries cannot be effectively predicted based solely on the trend of scientific production. (The scientific output trends of the 50 countries are reported in the Supporting Information).

thumbnail
Fig 5. SRI and SRII versus publication output by country.

Each data point represents data for a year between 1996 and 2019. Pearson correlation coefficients (by country and indicator) are overlaid.

https://doi.org/10.1371/journal.pone.0294669.g005

To better investigate differences between groups of countries, the correlation-based distance was used to produce two matrices MI and MII of order 50 × 50, which represent the distances between the 50 countries’ trends respectively on the SRI and SRII.

The structures of the two matrices were visualized using Kruskal’s Non-metric Multidimensional Scaling in the two MDS maps in Figs 6 and 7, based respectively on SRI and SRII trends. As explained in Methods and data, the distance between the dots representing the countries on the MDS maps is inversely proportional to the similarity of their SR trends, so that countries characterized by similar trends will be placed closer and countries characterized by dissimilar trends far away. Note that, in both maps, the distances of the points on the 2-D map distort the original distances between the time-series only slightly, as shown by the low values of the stress of the MDS solutions (respectively, 9.18% and 7.45%). Note that the countries in the two latter zones are identical to those showing anomalous correlations with trends in scientific output.

thumbnail
Fig 6. MDS solution for SRI.

G10 countries are in bold. Stress = 9.18%.

https://doi.org/10.1371/journal.pone.0294669.g006

thumbnail
Fig 7. MDS solution for SRII.

G10 countries are in bold. Stress = 7.45%. Note that the points of Malaysia and Indonesia are outside the plot.

https://doi.org/10.1371/journal.pone.0294669.g007

In the MDS map for SRI (Fig 6), three zones can be distinguished. The first is the big cluster in the left area, where most of the countries are concentrated. The second is the belt that surrounds the cluster and includes Iran, Romania, Pakistan, Thailand, Colombia, Saudi Arabia, Ukraine, Egypt, India, and, most notably, Italy, the only G10 country that is not placed inside the big cluster. The third zone comprises the rest of the map, where countries characterized by very specific trends are scattered: Malaysia, China, Indonesia, and the Russian Federation.

The MDS map for SRII (Fig 7) largely confirms this picture, showing significant overlap with the structural features of the previous map. Again, we find a big cluster including most of the countries, surrounded by a belt and, in the distant zones of the map, a scattering of anomalous countries. The belt includes the same countries as the belt in the previous map: Italy’s anomalous position with respect to G10 countries is confirmed. The most important difference between the two maps is China: under the profile of the SRI trend, the country was placed in the distant zone, whereas under the profile of the SRII trend, it is placed within the big cluster.

Discussion

The decreasing trend in both SRI and SRII which characterizes most of the countries shows that, in most of the cases, the overall citation impact of countries has grown more than the proportion of citation impact generated by domestic authors, i.e., that the denominator of both indicators has increased more than their numerator (see Section Indicator design). The faster increase in citations may be related in turn to the overall growth of scientific production and how it impacts the length of the reference lists of scientific publications. According to [57], in fact, the world’s scientific production exhibits 4% annual growth in publications and 1.8% annual growth in the number of references per publication. Combined, these dynamics produce a 12-year doubling period in the total amount of references, which results in turn in a generalized increase in citations [58]. The decreasing trends, thus, may be simply due to the different rates of growth of the numerator and the denominator of the indicators used here: country self-citations of Type I or Type II grow less than bibliographic references.

The observed decreasing trend in the SRI, however, contradicts previous studies of the development of this indicator over time: [26], in fact, report an average increase of +5.2 p.p. in the SRI of G10 countries between 2000 and 2016. Namely, the discrepancy between the present results and [26] depends on the different way of computing the SRI indicator. Indeed, [26] used a non-fixed citation window, which included all the years from the publication year to 2016. For example, for the cited items published in the year 2000, a 17-year citation window was used, by summing up all citations from 2000 to 2016; whereas, for the year 2006, the citation window was 11 years long, including citations from 2006 to 2016; for 2016, the citation window included only 1 year, i.e., only citations from 2016 itself were counted. Since self-citations are in general younger than external citations [25, 59], they tend to represent a higher proportion of total citations for the years when the citation window is shorter. Hence, [26] registered inflation of SRI for more recent years, as the citation window shortens. The present study, by contrast, does not suffer from this problem as it is based on a fixed citation window, i.e., only a fixed number of years after the target year is considered (see Section Indicator design).

Turning now to a more substantial interpretation of the results, there are two patterns that are worth highlighting. First, the emerging giant of science, China, is characterized by a unique behavior: China SRI and SRII show almost opposite trends, with the former significantly increasing and the latter significantly decreasing. If the increasing SRI trend shows that Chinese scientists heavily rely on the scientific production of their own country, the decreasing SRII trend indicates that author self-citations are diminishing, in line with Western countries. Interestingly, this divergence results in a growing difference over time between the two indicators (Fig 4), which can be interpreted as a sign of the rising international impact of Chinese publications. As noted above, in fact, the difference between SRI and SRII depends on the citations of the international publications of a country. A wider difference means that these international collaborations are widely cited not only by the focal country (China, in our case) but also by the collaborating countries. Results, therefore, seem to show that international collaborations of Chinese authors are increasingly cited by other countries as well, another sign of the new status of China as a scientific superpower [60]. Notably, India shows an increasing difference as well, which may be interpreted analogously as a sign of the rising scientific impact of this country.

The second key pattern, emerging from the MDS maps, is that there are several countries whose self-citation behavior stands out from that of the majority of countries. With a couple of exceptions (China and India), these countries are the same when the two indicators are considered: Colombia, Egypt, Indonesia, Iran, Italy, Malaysia, Pakistan, Romania, Russian Federation, Saudi Arabia, Thailand, and Ukraine. Note that the anomalous self-citation rates of these countries cannot be explained by the trend of their scientific production, as evidenced by the low correlation coefficients found (Fig 5). This suggests searching for specific perturbations of the citation behavior.

Interestingly, all these anomalous countries have adopted, in the recent past, specific research policies aiming at increasing publication output and citation impact of their national scientific community. In the following sections, the recent history of the research policies in each group of anomalous countries is reconstructed. It will be shown that all these policies are characterized by direct or indirect incentives that may create room for the strategic use of self-citations. SRI and SRII, therefore, seem to be sensitive to policy-induced perturbations of the citation habits.

Post-socialist countries: Russian Federation, Ukraine, and Romania

Since 2007, Russian Federation adopted measures aimed at boosting research productivity, in the form of performance-based funding and individual payment for publications [61]. In 2012, Putin’s May decrees N. 599 introduced various incentives for stimulating “the development of science in Russia and an increase in the number of articles by Russian scientists in the Web of Science Journals” [62, 63]. In particular, the project “5top100” aimed to push at least five Russian universities to enter the top hundred of leading international universities according to the global universities rank [64]. The project council paid attention to bibliometric indicators, including the number of publications and citations in Web of Science and Scopus [64]. At the local level, Russian universities introduced new promotion criteria and financial incentives for faculty. After the policy intervention, the research productivity of the country significantly increased [6367]. However, a contemporary rise in country self-citations in conference proceedings has been noted [63] and the spread of unethical practices after the policy change, including “predatory journals”, plagiarism, and paper mills, has been repeatedly denounced [68, 69].

As to Ukraine, the new law “On scientific and Technical Activity” was enacted on January 16, 2016 (http://iht.univ.kiev.ua/ncst2016/index-en.html). It established a National Council of Ukraine on Science and Technology directly controlled by the ministers. The following year, the European Commission organized a Peer review of the Ukrainian Research and Innovation System largely supporting the new law. One of the key recommendations issued by the Commission was to identify research universities after a period of 5 years by taking into consideration also “the number of international publications and citations”. In the meantime, the use of bibliometric indicators become largely diffused in Ukraine for ranking institutions [70], for distributing financial awards, and for evaluating projects [71]. According to [72], “until 2015 publication requirements for becoming associate professor and professor included only articles in Ukrainian journals. In 2015 they were substituted for articles in journals indexed in Scopus and WoS”. According to scientists interviewed by [73], the Ukrainian academy suffers from a significant and inherited problem of misconduct and plagiarism. Possibly grafted on this tradition, new forms of adaptation to the bibliometric game are arising, such as publishing articles in selected national journals [74] and in Scopus de-listed journals [71]. Evidence about over-citations and self-citations has been provided as well [71].

Lastly, Romania started major reforms of tertiary education following the provisions of the Law of Education n. 1/2011 by modifying recruitment, university funding, and quality assurance [75]. It introduced also a research-driven classification and ranking system for universities managed by the Romanian Ministry of Education University, which constituted the informative basis of performance-based funding [76]. Academic and research staff recruitment and promotion changed radically from a model based on in-breeding to one taking into account individual performances measured by the number of publications and citations [75]. The Romanian Program for Rewarding Research Results, which had already started in 2007, was strengthened with direct payment to authors for publication in indexed journals [77]. [78] described in detail the functioning of the program, and highlighted that articles are rewarded according to the metrics of the journal where they are published. According to their analysis, monetary incentives supported productivity, but not the impact of Romanian research. The presence of incentives to publication might push toward misconduct, as suggested by the Romanian high level of retractions [79]. Evidence of self-citation and citation stacking for Romanian journals may be correlated to the necessity to boost journal metrics [80].

Southeast Asia: Malaysia, Thailand, Indonesia

According to a UNESCO report [81], university rankings are central in the research policies of Malaysia and Thailand: “A key ingredient in high rankings is a university’s publication rate. Consequently, faculty members—particularly those teaching in graduate programmes—are under pressure to publish in top-tier international journals”. According to [82], Asian higher education institutions witnessed the “proliferation of policies surrounding the fanaticism with metrics [for] incentivizing scholars to publish through selected publication”.

In Malaysia, the discussion about university performances and rankings started early [83]. Policy interventions happened in a highly centralized structure, where salaries and promotion criteria were defined directly by the Ministry of Education. In 2007, the Malaysian government adopted a National Higher Education Strategy Plan introducing performance-based funding of universities [84]. The strategic objective to “empower research teams with new teamwork concepts to produce international level research output” was defined in an action plan for “improving the quality of faculty publications”. Remarkably, the only indicator adopted was the “Increased percentage of staff achieving at least 100 citations” [85]. The upward trend of Malaysian country self-citations coincides with the years following this government intervention. In recent years, decentralized policies adopted by universities provide individual payment schemes for publications [86] (https://tinyurl.com/4t5kpd2h; https://tinyurl.com/mr33rzcn) and received citations (https://tinyurl.com/n3779dz3).

As to Thailand, the government initiated the establishment of the National Research University (NRU) project in 2009. “This plan aimed at developing academic excellences to enhance the country’s research activities and to promote the better university-industry linkages for national competitiveness. The Office of the Higher Education Commission’s selection criteria were mainly based on the ranking system conducted by Times Higher Education-Quacquarelli Symonds (THE-QS) and the impact factor of publications published on Scopus database” [87]. A list of excellent universities was also defined [88].

Indonesia was “the weakest nation in all relative scientometric indicators”, with respect to Malayisia, Philippines, Thailand, and Vietnam [89]. According to [90] “the increase in the number of publications in recent years […] is a reflection of government policy on research and academic careers and attempts to improve the position of Indonesian universities in international ranking”. The main intervention was the Indonesian law on higher education Number 12/2012. [91] documented that “Academics who are successful in publishing their articles in Scopus- indexed journals would be rewarded with a certain amount of money and it goes directly to their pocket. Such a standard is also used as a measure for promotions […], for payment of certified academics, and for honorary allowance. […] The obligation to publish articles in reputable international journals has become integrated into doctoral programs and serves as a requirement to be met prior to the completion of the study”. According to a survey [91], Indonesian “academics have attributed their interpretation of the rewards as a mere completion of publishing in any kinds of journals indexed in Scopus apart from the consideration of the quality. Consequently, academics have performed their own way or strategy of publishing through the easiest, fastest, cheapest open access journals and proceedings”. The performance in terms of citations per paper appeared not so high with respect to other neighborhood countries until 2017 [92]. [93] documented that Indonesia has made a recent and remarkable shift towards conference proceedings publishing. Rochmyaningsih [94] criticized the adoption of this aggressive policy by arguing that “the developing world needs more than numbers”. In 2017 Indonesia’s Ministry of Research, Technology and Higher Education introduced Indonesia’s Science and Technology Index (SINTA; https://sinta.kemdikbud.go.id/), based on Scopus data, that “gives recognition to Indonesian scientists, triggers competition among them, and motivates them to be better” [95]. The number of papers, citations, and H-index are mixed in an index used for evaluating research grants applications, promotions, and salary negotiations. According to [95] several top scorers “had inflated their SINTA score by publishing large numbers of papers in low-quality journals, citing their own work excessively, or forming networks of scientists who cited each other”. The present paper documents that the adoption of policies both in 2012 and 2017 was flanked by a modification of self-citation strategies of Indonesian scientists.

Muslim-majority countries: Egypt, Iran, Saudi Arabia, Pakistan

According to [96], Egypt, Iran, Pakistan, and Saudi Arabia differ from other Muslim-majority countries in terms of research performance. In particular, Egypt and Saudi Arabia have been the most active research producers from the Arab world in the last 20 years. Like other Arab countries, they adopted a reform of tertiary education and witnessed a remarkable growth of publications and citations [9698], matched however by a rising number of retractions due to misconduct [99, 100]. More in general, according to [101], the problem of corruption is widespread in Arab universities.

Since the mid-2000s, Saudi Arabia adopted National Development plans aiming to support research productivity [102], by mixing centralized strategies, such as the National Science Technology and Innovation Plan inspired by the US National Science Foundation, and decentralized ones adopted by universities [103]. According to [104] this catching-up strategy of Saudi Arabia universities started in 2007, and it was mainly based on attempts to raise research outputs, prestige, and rankings (e.g. https://tinyurl.com/4s8b6hef), by allocating “significant research funding to support industry-based Research Chairs as well as the employment of high-profile international researchers to lead projects that will be staffed by university faculty and postdoctoral students” [102]. Individual incentives for researchers are widely adopted, with salary increases and promotions tied to publications and citations (e.g. https://tinyurl.com/2s3zchwk; https://tinyurl.com/5n96nmrt). The strategy of affiliating to Saudi Arabia universities external highly cited researchers for improving rankings received early criticism [105107]. Nonetheless, [108] claims that self-citations are not anomalous in Saudi Arabia, at least in the medical specialties.

In comparison to Saudi Arabia, the governance of Egyptian universities is traditionally much more centralized [109]. According to [98], Egypt “demonstrate[s] the importance of incentives within hiring organizations, specifically whether researchers are rewarded for publications or obtaining funding”. Indeed, the Ministry of Scientific Research introduced competitive funding to research in 2007, by favoring basic research, the formation of research groups, and international collaborations. At a single university level, “internal practices recognize and reward certain forms of performance more than others—such as teaching, administration, graduate supervision, advising and outreach—as well as the expected quantity and prestige of scientific publications” [98]. According to some researchers interviewed in [110], individual financial incentives and national awards [111] represent the main push leading to the improvement of Egypt’s higher education sector. Others have a less positive attitude and highlight the diffusion of malpractices in research such as plagiarism, data fabrication, and manipulation [112]. According to a comprehensive survey of Egyptian researchers, financial rewards for publications together with low salaries are among the most important risk factors leading to research misconduct [112]. Finally, it should be noted that Egyptian universities provide individual awards for citations (e.g. https://bu.edu.eg/BUNews/25947).

Regarding Pakistan, in 2002, the country’s government established a Higher Education Commission in alignment with the World Bank’s task force recommendations. This commission was aimed at expanding the higher education sector and improving research within the country. Various measures were adopted, including incentives aimed to promote research and scholarship [113115]. Since 2002, a financial incentive based on the number of publications, number of citations received, and Impact Factor of journals was also introduced [116, 117]. The growth of international publications, collaborations, and citations is considered a result of these policies [114, 117, 118]. Still, Pakistan is currently one of the leading countries in terms of retractions [79, 99], and, according to [119], under the rising menace of scholarly black-market pushed-up by monetary incentives.

Iran had between 1980 and 2010, “one of the fastest rates of growth in scientific production that the world has witnessed”, probably due to the nuclear technology development program [120]. In 2009, Iran announced a 20-year “comprehensive plan for science” focused on higher education and stronger links between industry and academia [121, 122]. The quantitative growth of Iranian science has continued until now [121, 123]. It is a controversial issue, however, whether this development has been matched by the increase of scientific quality too [123, 124]. According to a critic, “The state has imposed deeply short-sighted research policies on universities with the sole objective of increasing the number of publications, which is in turn used in its propaganda to demonstrate progress in technological self-sufficiency and mask significant shortcomings caused by decades of isolation due to the regime’s international policy. […] A top-down incentive for publication along with lack of real demand from the economy, which is not based on new technology development, have pushed Iranian researchers to focus on the publishability of their works rather than their relevance and practical impact” [124]. As in the other countries, “the government’s policy in higher education makes academic promotions and student graduation contingent upon publication of papers in scientific journals. These policies have created an environment that lends itself to the most extreme form of the publish-or-perish paradigm” [125]. Retractions of articles authored by Iranian scientists have attracted attention worldwide [126, 127]. An anomalous rate of self-citations has been documented as well [128].

Colombia

Since 2002, Colombia has introduced a model of wage incentive based on research productivity, the Faculty Promotion Policy for Colombian Public Universities (Decree 1279 of 2002). The performance rating is based on a national index of scientific journals called Publindex. In 2009, the Ministry of Education started to measure production in terms of citations in the WoS database. In 2018, the system was finally strengthened [129]. By and large, this legislation established a pay-for-performance system through salary points calculated in accordance with the higher education degrees, academic rank (fixed components), and academic productivity (variable component) [130132]. [129] documented by anecdotal evidence that self-citations and citation clubs are perceived as current problems by Colombian scholars.

Italy

Italy is among the top 10 producers of science in terms of global production and total citations (Table 2) and it has been considered the main European competitor of the United Kingdom for citation impact [133, 134]. However, data shows that Italy is the only G10 country exhibiting an anomalous trend of self-citations. The Italian anomaly is especially visible when compared with the trends of the other G10 countries: from 2010, Italy starts to diverge from the benchmark countries in both indicators (Fig 8). At the end of the observation period, it results to be the country with the highest SRII, above Japan.

thumbnail
Fig 8. G10 countries’ trends in SRI and SRII.

Italian trends are in bold green.

https://doi.org/10.1371/journal.pone.0294669.g008

The beginning of the change in the trend coincides with the wide process of reformation of the Italian university system started by the government in 2010 (Law 240/2010). A governmental agency (ANVUR) was established in charge of monitoring and evaluating the Italian research system and, in 2011, the first national research assessment exercise started, followed by a second round in 2015. In both, the evaluation was largely based on the automatic or semi-automatic use of algorithms fed by citation indicators [135, 136]. Universities started to be funded according to their performance in the research assessments. Moreover, the reform modified also the recruitment and advancement system for university professors by introducing a national scientific qualification (ASN). This qualification is mandatory both for hiring and promotion and, in order to obtain it, candidates in natural sciences, life sciences, and engineering must exceed publication and citation thresholds centrally defined by ANVUR [26]. As a result of these reforms, bibliometric performance has acquired a central role in the career of Italian scientists [137]. At the same time, anomalous rises in Italian self-citations have been documented by several studies [26, 3438].

Conclusions

Since the early times of citation indexes, preoccupations with the opportunistic use of citations have been voiced by bibliometricians and the scientific community [138, 139]. The centrality acquired by metrics in the various ganglia of the research system, from the career of individual scientists to the ranking of institutions until the evaluation of the scientific performance of entire countries, has further sharpened these concerns [6, 140143]. Self-citation, in particular, has persistently been indicated as among the easiest strategies available to scientists for artificially boosting their citation-related performance indicators [6, 144], raising the question of whether scientists, under pressure, do indeed attempt to game citation scores by increasing self-citation.

In the present study, we investigated how the propensity to self-citation changed in 50 countries all over the world from 1996 to 2019, using two different indicators based on country self-citations. The results show that, for most countries, self-citation rates have decreased over time following similar patterns. Tendency to self-citation, both at the country level and for individual scientists, seems in fact to be declining. Against this background, however, there are some countries that exhibit anomalous behavior, showing self-citation trends that are significantly different from those of “standard” countries.

The analysis of the research policies adopted by these anomalous countries in the last years has revealed that they all share a common trait, namely the introduction of direct or indirect rewards for the bibliometric performance of scientists. The temporal association we found in all anomalous countries between changes in policies on the one hand and changes in the self-citation behavior of the national scientific community on the other hand suggests that scientists do indeed respond to the new climate of incentives by altering, among other things, their citation habits. Policy pressure seems therefore capable of affecting rapidly and visibly the citation behavior of entire countries, possibly distorting global rankings of countries based on citations [31].

Clearly, we cannot offer a full-fledged causal explanation of our data, displaying the causal chains that start from the policy and end with the citation choices of individual authors. Neither we can demonstrate that the whole self-citation gain of anomalous countries is due to opportunistic adaptation to research policies. Still, the most likely high-level explanation of the macro-trends we observe is that the policies centered on or emphasizing citation performance do significantly affect the behavior of scientists.

Remarkably, our results show that merely having a performance-based funding system involving citation-based indicators is insufficient for a country to exhibit anomalous self-citation behavior. For example, countries like Belgium, Bulgaria, Croatia, and Sweden employ citation-based indicators in evaluating their university systems, yet they do not display anomalous self-citation tendencies [145]. It appears that the crucial factor influencing this anomalous self-citation behavior is the proximity of incentives based on citations to an individual researcher’s career and wage. It seems that the closer these incentives are to affecting researchers’ career progression and wages, the more likely they are to influence citation behavior. In Italy, for instance, the national scientific qualification, which is awarded based on citation indicators, has a direct impact on the career of researchers, as without it researchers cannot access tenure-track positions or advance to full professorships. By contrast, in countries where citation-based indicators contribute to complex funding formulas at the university level, the influence on citation behavior is more distant from individual researchers and, consequently, less significant.

In future research, it will be important to further investigate how varying architectures of performance-based funding systems have more or less impact on citation behavior, exploring in particular the extent to which the proximity of citation-based incentives to individual careers and wages induces strategic behavior. Other promising topics for future research include examining whether research fields characterized by varying epistemological, social, and institutional features respond differently to science policies and investigating whether citation incentives may affect countries’ mix of scientific output, inducing specialization in areas rewarded by citation metrics (e.g., the medical sciences) at the expense of less “performing” areas (e.g., the humanities).

On its part, this study contributes to the ongoing discussion on research evaluation systems by showing that when bibliometric indicators, and citation-based indicators in particular, are integrated into systems of incentives, they cease to be neutral measures to become active components in the research system. As such, they are able to modify the behavior of entire scientific communities. Hence, they should be handled by science policy makers with the most extreme caution.

Supporting information

S1 File. SRI and SRII trends using 1-year and 5-year citation windows; MDS solutions for SRI and SRII using 1-year and 5-year citation windows; publication output of the 50 countries between 1996 and 2019.

https://doi.org/10.1371/journal.pone.0294669.s001

(PDF)

S1 Data. Raw data used for computing the indicator.

Further indicators computed: over-citation ratio, odds-ratio, and relative self-citation rate.

https://doi.org/10.1371/journal.pone.0294669.s002

(CSV)

Acknowledgments

This work uses Scopus data provided by Elsevier through the ICSR Lab. We are very grateful to Lucio Barabesi, Jeroen Bass for his help in retrieving the data, and Giuseppe De Nicolao for insightful discussions in the early stage of the research.

References

  1. 1. Garfield E. Is citation analysis a legitimate evaluation tool? Scientometrics. 1979;1(4):359–375.
  2. 2. Garfield E, Sher IH. New factors in the evaluation of scientific literature through citation indexing. American Documentation. 1963;14(3):195–201.
  3. 3. Glänzel W, Debackere K, Thijs B, Schubert A. A concise review on the role of author self-citations in information science, bibliometrics and science policy. Scientometrics. 2006;67(2):263–277.
  4. 4. MacRoberts MH, MacRoberts BR. The mismeasure of science: Citation analysis. Journal of the Association for Information Science and Technology. 2018;69(3):474–482.
  5. 5. Schubert A, Glänzel W, Thijs B. The weight of author self-citations. A fractional approach to self-citation counting. Scientometrics. 2006;67(3):503–514.
  6. 6. Szomszor M, Pendlebury DA, Adams J. How much is too much? The difference between research influence and self-citation excess. Scientometrics. 2020;
  7. 7. Aksnes DW. A macro study of self-citation. Scientometrics. 2003;56(2):235–246.
  8. 8. Costas R, van Leeuwen TN, Van Raan AFJ. Is scientific literature subject to a ‘Sell-By-Date’? A general methodology to analyze the ‘durability’ of scientific documents. Journal of the American Society for Information Science and Technology. 2010;61(2):329–339.
  9. 9. Fowler JH, Aksnes DW. Does self-citation pay? Scientometrics. 2007;72(3):427–437.
  10. 10. Thijs B, Glänzel W. The influence of author self-citations on bibliometric meso-indicators. The case of european universities. Scientometrics. 2006;66(1):71–80.
  11. 11. Bonzi S, Snyder HW. Motivations for citation: A comparison of self citation and citation to others. Scientometrics. 1991;21(2):245–254.
  12. 12. Pichappan P, Sarasvady S. The other side of the coin: The intricacies of author self-citations. Scientometrics. 2002;54(2):285–290.
  13. 13. Hyland K. Self-citation and self-reference: Credibility and promotion in academic publication. Journal of the American Society for Information Science and Technology. 2003;54(3):251–259.
  14. 14. White HD. Authors as citers over time. Journal of the American Society for Information Science and Technology. 2001;52(2):87–108.
  15. 15. Sugimoto CR, Larivière V. Measuring research: what everyone needs to know. What everyone needs to know. New York: Oxford University Press; 2018.
  16. 16. Khelfaoui M, Larrègue J, Larivière V, Gingras Y. Measuring national self-referencing patterns of major science producers. Scientometrics. 2020;123(2):979–996.
  17. 17. Ioannidis JPA. A generalized view of self-citation: Direct, co-author, collaborative, and coercive induced self-citation. Journal of Psychosomatic Research. 2015;78(1):7–11. pmid:25466321
  18. 18. Leydesdorff L. Caveats for the use of citation indicators in research and journal evaluations. Journal of the American Society for Information Science and Technology. 2008;59(2):278–287.
  19. 19. Andrade A, González-Jonte R, Campanario JM. Journals that increase their impact factor at least fourfold in a few years: The role of journal self-citations. Scientometrics. 2009;80(2):515–528.
  20. 20. Hartley J. To cite or not to cite: author self-citations and the impact factor. Scientometrics. 2012;92(2):313–317.
  21. 21. Yu G, Wang L. The self-cited rate of scientific journals and the manipulation of their impact factors. Scientometrics. 2007;73(3):321–330.
  22. 22. Gul S, Ahmad Shah T, Shafiq H. The prevalence of synchronous self-citation practices at the institutional level. Malaysian Journal of Library and Information Science. 2017;22(1):1–14.
  23. 23. Hellsten I, Lambiotte R, Scharnhorst A, Ausloos M. Self-citations, co-authorships and keywords: A new approach to scientists’ field mobility? Scientometrics. 2007;72(3):469–486.
  24. 24. Egghe L, Rousseau R, Yitzhaki M. The “own-language preference”: Measures of relative language self-citation. Scientometrics. 1999;45(2):217–232.
  25. 25. Tagliacozzo R. Self-Citations in Scientific Literature. Journal of Documentation. 1977;33(4):251–265.
  26. 26. Baccini A, De Nicolao G, Petrovich E. Citation gaming induced by bibliometric evaluation: A country-level comparative analysis. PLOS ONE. 2019;14(9):e0221212. pmid:31509555
  27. 27. Bakare V, Lewison G. Country over-citation ratios. Scientometrics. 2017;113(2):1199–1207. pmid:29081556
  28. 28. Bardeesi AM, Jamjoom AAB, Sharab MA, Jamjoom AB. Impact of country self citation on the ranking of the top 50 countries in clinical neurology. eNeurologicalSci. 2021;23:100333. pmid:33732913
  29. 29. Jaffe K. Do countries with lower self-citation rates produce higher impact papers? Or, does humility pay? Interciencia. 2011;36(9):694–698.
  30. 30. Minasny B, Hartemink AE, McBratney A. Individual, country, and journal self-citation in soil science. Geoderma. 2010;155(3-4):434–438.
  31. 31. Shehatta I, Al-Rubaish AM. Impact of country self-citations on bibliometric indicators and ranking of most productive countries. Scientometrics. 2019;120(2):775–791.
  32. 32. Board NS. Science and Engineering Indicators 2018. Alexandria, VA: National Science Foundation; 2018. Available from: https://www.nsf.gov/statistics/indicators/.
  33. 33. Lancho Barrantes BS, Guerrero Bote VP, Rodríguez ZC, de Moya Anegón F. Citation flows in the zones of influence of scientific collaborations. Journal of the American Society for Information Science and Technology. 2012;63(3):481–489.
  34. 34. Abramo G, D’Angelo CA, Grilli L. The effects of citation-based research evaluation schemes on self-citation behavior. Journal of Informetrics. 2021;15(4):101204.
  35. 35. Peroni S, Ciancarini P, Gangemi A, Nuzzolese AG, Poggi F, Presutti V. The practice of self-citations: a longitudinal study. Scientometrics. 2020;123(1):253–282.
  36. 36. Scarpa F, Bianco V, Tagliafico LA. The impact of the national assessment exercises on self-citation rate and publication venue: an empirical investigation on the engineering academic sector in Italy. Scientometrics. 2018;117(2):997–1022.
  37. 37. Seeber M, Cattaneo M, Meoli M, Malighetti P. Self-citations as strategic response to the use of metrics for career decisions. Research Policy. 2019;48(2):478–491.
  38. 38. Vercelli S, Pellicciari L, Croci A, Cornaggia CM, Cecchi F, Piscitelli D. Self-citation behavior within the health allied professions’ scientific sector in Italy: a bibliometric analysis. Scientometrics. 2022;
  39. 39. Lawani SM. On the Heterogeneity and Classification of Author Self-Citations. Journal of the American Society for Information Science. 1982;33(5):281–284.
  40. 40. Frame JD, Narin F. The national self-preoccupation of American scientists: An empirical view. Research Policy. 1988;17(4):203–212.
  41. 41. Moed HF. Citation analysis in research evaluation. No. v. 9 in Information science and knowledge management. Dordrecht: Springer; 2005.
  42. 42. Ladle RJ, Todd PA, Malhado ACM. Assessing insularity in global science. Scientometrics. 2012;93(3):745–750.
  43. 43. Glänzel W, Schubert A. Domesticity and internationality in co-authorship, references and citations. Scientometrics. 2005;65(2):323–342.
  44. 44. Elsevier. Research Metrics Guidebook; 2018. Available from: https://www.elsevier.com/__data/assets/pdf_file/0020/53327/ELSV-13013-Elsevier-Research-Metrics-Book-r5-Web.pdf.
  45. 45. Hassan SU, Haddawy P. Measuring international knowledge flows and scholarly impact of scientific research. Scientometrics. 2013;94(1):163–179.
  46. 46. Bookstein A, Yitzahki M. Own language preference: A new measure of “relative language self-citation”. Scientometrics. 1999;46:337–348.
  47. 47. Aghabozorgi S, Seyed Shirkhorshidi A, Ying Wah T. Time-series clustering—A decade review. Information Systems. 2015;53:16–38.
  48. 48. Warren Liao T. Clustering of time series data—a survey. Pattern Recognition. 2005;38(11):1857–1874.
  49. 49. Borg I, Groenen PJF. Modern multidimensional scaling: theory and applications. New York; London: Springer; 2010.
  50. 50. Van Raan AFJ. Measuring Science: Basic Principles and Application of Advanced Bibliometrics. In: Glänzel W, Moed HF, Schmoch U, Thelwall M, editors. Springer Handbook of Science and Technology Indicators. Cham: Springer International Publishing; 2019. p. 237–280. Available from: http://link.springer.com/10.1007/978-3-030-02511-3_10.
  51. 51. Todeschini R, Baccini A. Handbook of bibliometric indicators: quantitative tools for studying and evaluating research. 1st ed. Weinheim: Wiley-VCH Verlag GmbH & Co.KGaA; 2016.
  52. 52. Baas J, Schotten M, Plume A, Côté G, Karimi R. Scopus as a curated, high-quality bibliometric data source for academic research in quantitative science studies. Quantitative Science Studies. 2020;1(1):377–386.
  53. 53. Montero P, Vilar JA. TSclust: An R Package for Time Series Clustering. Journal of Statistical Software. 2014;62(1).
  54. 54. Golay X, Kollias S, Stoll G, Meier D, Valavanis A, Boesiger P. A new correlation-based fuzzy logic clustering algorithm for FMRI. Magnetic Resonance in Medicine. 1998;40(2):249–260. pmid:9702707
  55. 55. R Core Team. R: A Language and Environment for Statistical Computing; 2021. Available from: https://www.R-project.org/.
  56. 56. Venables WN, Ripley BD. Modern Applied Statistics with S. 4th ed. New York: Springer; 2002.
  57. 57. Pan RK, Petersen AM, Pammolli F, Fortunato S. The memory of science: Inflation, myopia, and the knowledge network. Journal of Informetrics. 2018;12(3):656–678.
  58. 58. Nicolaisen J, Frandsen TF. Number of references: a large-scale study of interval ratios. Scientometrics. 2021;126(1):259–285.
  59. 59. Lin WYC, Huang MH. The relationship between co-authorship, currency of references and author self-citations. Scientometrics. 2012;90(2):343–360.
  60. 60. Tollefson J. China declared world’s largest producer of scientific articles. Nature. 2018;553(7689):390–390.
  61. 61. Schiermeier Q. Russian science: The battle for Russia’s brains. Nature. 2007;449(7162):524–527. pmid:17968377
  62. 62. Kassian A, Melikhova L. Russian Science Citation Index on the WoS platform: a critical assessment. Journal of Documentation. 2019;75(5):1162–1168.
  63. 63. Moed HF, Markusova V, Akoev M. Trends in Russian research output indexed in Scopus and Web of Science. Scientometrics. 2018;116(2):1153–1180.
  64. 64. Guskov AE, Kosyakov DV, Selivanova IV. Boosting research productivity in top Russian universities: the circumstances of breakthrough. Scientometrics. 2018;117(2):1053–1080.
  65. 65. Kosyakov D, Guskov A. Reasons and consequences of changes in Russian research assessment policies. Scientometrics. 2022;127(8):4609–4630.
  66. 66. Matveeva N, Sterligov I, Yudkevich M. The effect of Russian University Excellence Initiative on publications and collaboration patterns. Journal of Informetrics. 2021;15(1):101110.
  67. 67. Matveeva N, Ferligoj A. Scientific collaboration in Russian universities before and after the excellence initiative Project 5-100. Scientometrics. 2020;124(3):2383–2407.
  68. 68. Abalkina A. Publication and collaboration anomalies in academic papers originating from a paper mill: Evidence from a Russia-based paper mill. Learned Publishing. 2023; 36(4): 689–782.
  69. 69. Abalkina A. Unethical Practices in Research and Publishing: Evidence from Russia. Scholarly Kitchen. 2021. available from: https://scholarlykitchen.sspnet.org/2021/02/04/guest-post-unethical-practices-in-research-and-publishing-evidence-from-russia/.
  70. 70. Akbash KS, Pasichnyk NO, Rizhniak RY. Analysis of key factors of influence on scientometric indicators of higher educational institutions of Ukraine. International Journal of Educational Development. 2021;81:102330.
  71. 71. Nazarovets S. Analysis of publications by authors of Ukrainian institutes in Scopus-delisted titles. Learned Publishing. 2022;35(4):499–515.
  72. 72. Hladchenko M, Moed HF. National orientation of Ukrainian journals: means-ends decoupling in a semi-peripheral state. Scientometrics. 2021;126(3):2365–2389.
  73. 73. Filtenborg E, Weichert S. Academics in Ukraine Fighting Against Rampant Misconduct. The Scientist. 2019;.
  74. 74. Nazarovets S. Controversial practice of rewarding for publications in national journals. Scientometrics. 2020;124(1):813–818.
  75. 75. Vlăsceanu L, Hâncean MG. Policy Incentives and Research Productivity in the Romanian Higher Education. An Institutional Approach. In: Curaj A, Matei L, Pricopie R, Salmi J, Scott P, editors. The European Higher Education Area: Between Critical Reflections and Future Policies. Cham: Springer International Publishing; 2015. p. 185–203.
  76. 76. Vîiu GA, Păunescu M, Miroiu A. Research-driven classification and ranking in higher education: an empirical appraisal of a Romanian policy experience. Scientometrics. 2016;107(2):785–805.
  77. 77. Pisoschi AM, Pop A, Predoi G, Purdoiu L. An Analysis of the Influence of Some Incentives on Bibliometric Performances. Journal of Scientometric Research. 2020;9(3): 344–351.
  78. 78. Vîiu GA, Păunescu M. The citation impact of articles from which authors gained monetary rewards based on journal metrics. Scientometrics. 2021;126(6):4941–4974.
  79. 79. Rivera H, Teixeira da Silva JA. Retractions, Fake Peer Reviews, and Paper Mills. JKMS. 2021;36(24):e165–0. pmid:34155837
  80. 80. Heneberg P. From Excessive Journal Self-Cites to Citation Stacking: Analysis of Journal Self-Citation Kinetics in Search for Journals, Which Boost Their Scientometric Indicators. PLOS ONE. 2016;11(4). pmid:27088862
  81. 81. UNESCO. Higher Education in Asia: Expanding Out, Expanding Up. The rise of graduate education and university research. Unesco Institute for statistics; 2014.
  82. 82. Tham A. A policed state—higher education quality policies and influences on Asian tourism scholarship. Journal of Policy Research in Tourism, Leisure and Events. 2019;11(2):236–256.
  83. 83. Gill J. Malaysian rankings flop’shames’ the nation Times Higher Education. 2008. Available from: https://www.timeshighereducation.com/news/malaysian-rankings-flop-shames-the-nation/404570.article.
  84. 84. Ahmad AR, Farley A. Funding Reforms in Malaysian Public Universities from the Perspective of Strategic Planning. Procedia—Social and Behavioral Sciences. 2014;129:105–110.
  85. 85. Ministry of Education of Malaysia National higher education action plan 2007-2010, 2007. Available from: https://www.ilo.org/dyn/youthpol/en/equest.fileutils.dochandle?p_uploaded_file_id=477.
  86. 86. Wan CD, Chapman D, Hutcheson S, Lee M, Austin A, Md Zain AN. Changing higher education practice in Malaysia: the conundrum of incentives. Studies in Higher Education. 2017;42(11):2134–2152.
  87. 87. Siripitakchai N, Miyazaki K. Thailand’s national research universities (NRUs) and their University-Industry Linkages (UILs) based on university-industry co-publications (UICs). In: 2015 Portland International Conference on Management of Engineering and Technology (PICMET); 2015. p. 287–294.
  88. 88. Sombatsompop N, Markpin T, Ratchatahirun P, Yochai W, Wongkaew C, Premkamolnetr N. Research performance evaluations of Thailand national research universities during 2007-2009. Information Development. 2010;26(4):303–313.
  89. 89. Fiala D. Indonesia’s place in the research landscape of Southeast Asia. Unisia. 2022.
  90. 90. Achwan R, Ganie-Rochman M, Alamsyah AR, Triana L. University reform and the development of social sciences in Indonesia. International Journal of Educational Development. 2020;78.
  91. 91. Gaus N,Jasruddin , Saleh A, Resnawaty R, Paramma MA, Tanjung Y. Trading-off monetary rewards as reinforcers to enhance task motivation and performance of publication in academia. Higher Education Quarterly. 2022;76(4):800–814.
  92. 92. Achsan H, Wibowo WC, Achsan MMB, Purnama DG, Lubis KB. The Quality of Indonesian Scientific Articles and Its Neighboring Countries. In: 2019 IEEE Conference of Russian Young Researchers in Electrical and Electronic Engineering (EIConRus); 2019. p. 174–178.
  93. 93. Purnell PJ. Conference proceedings publications in bibliographic databases: a case study of countries in Southeast Asia. Scientometrics. 2021;126(1):355–387.
  94. 94. Rochmyaningsih D. The developing world needs more than numbers. Nature. 2017;542(7639):7–7. pmid:28150793
  95. 95. Rochmyaningsih D. How to shine in Indonesian science? Game the system. Science. 2019;363(6423):111–112. pmid:30630909
  96. 96. Oldac YI. Global science and the muslim world: overview of muslim-majority country contributions to global science. Scientometrics. 2022;127(11):6231–6255.
  97. 97. Ahmad S, Ur Rehman S, Ashiq M. A Bibliometric Review of Arab World Research from 1980-2020. Science & Technology Libraries. 2021;40(2):133–153.
  98. 98. Currie-Alder B, Arvanitis R, Hanafi S. Research in Arabic-speaking countries: Funding competitions, international collaboration, and career incentives. Science and Public Policy. 2017;45(1):74–82.
  99. 99. Ataie-Ashtiani B. World Map of Scientific Misconduct. Science and Engineering Ethics. 2018;24(5):1653–1656. pmid:28653166
  100. 100. Liu W, Lei L. Retractions in the Middle East from 1999 to 2018: a bibliometric analysis. Scientometrics. 2021;126(6):4687–4700.
  101. 101. Hillman JR, Baydoun E. The Future of Universities in the Arab Region: A Review. In: Baydoun E, Hillman JR, editors. Universities in Arab Countries: An Urgent Need for Change: Underpinning the Transition to a Peaceful and Prosperous Future. Cham: Springer International Publishing; 2018. p. 1–53.
  102. 102. Al-Ohali M, Burdon S. International Collaboration. In: Smith L, Abouammoh A, editors. Higher Education in Saudi Arabia: Achievements, Challenges and Opportunities. Dordrecht: Springer Netherlands; 2013. p. 159–166.
  103. 103. Smith L, Abouammoh A. Higher Education in Saudi Arabia: Reforms, Challenges and Priorities. In: Smith L, Abouammoh A, editors. Higher Education in Saudi Arabia: Achievements, Challenges and Opportunities. Dordrecht: Springer Netherlands; 2013. p. 1–12.
  104. 104. S Al-Khalifa H. Scientometric assessment of Saudi publication productivity in computer science in the period of 1978-2012. International Journal of Web Information Systems. 2014;10(2):194–208.
  105. 105. Bhattacharjee Y. Saudi Universities Offer Cash in Exchange for Academic Prestige. Science. 2011;334(6061):1344–1345. pmid:22158799
  106. 106. Gingras Y. How to boost your university up the rankings. University World News. 2014;329.
  107. 107. Miley G. Saudi University Policy: Overvalued Rankings. Science. 2012;335(6072):1041–1042. pmid:22383829
  108. 108. Bardeesi AM, Jamjoom AA, Algahtani A, Jamjoom A. The Impact of Country Self-Citation Rate Among Medical Specialties in Saudi Arabia. Cureus. 2021. pmid:33552795
  109. 109. Waterbury J. Governance of Arab Universities: Why Does It Matter? In: Baydoun E, Hillman JR, editors. Universities in Arab Countries: An Urgent Need for Change: Underpinning the Transition to a Peaceful and Prosperous Future. Cham: Springer International Publishing; 2018. p. 55–70.
  110. 110. Sawahel W. Why Egypt’s higher education sector continues to improve. University World News Africa Edition. 2021.
  111. 111. Ibrahim B. The role of Egyptian State Awards in changing researchers’ performance in the science and technology sector. Research Evaluation. 2020;29(2):171–190.
  112. 112. Moustafa PE. Curbing the practices of research misconduct: a qualitative study on the perceptions of researchers at Egyptian public institutions [Thesis]. American University in Cairo; 2019. Available from: https://fount.aucegypt.edu/etds/791.
  113. 113. ur Rahman A. Building a Knowledge Economy. In: Baydoun E, Hillman JR, editors. Universities in Arab Countries: An Urgent Need for Change: Underpinning the Transition to a Peaceful and Prosperous Future. Cham: Springer International Publishing; 2018. p. 105–121.
  114. 114. Iqbal HM, Mahmood K, Iqbal SA. Factors Contributing Towards Research Productivity and Visibility: a Case Study of Pakistan. Libri. 2018;68(2):85–98.
  115. 115. Osama A, Najam A, Kassim-Lakha S, Zulfiqar Gilani S, King C. Pakistan’s reform experiment. Nature. 2009;461(7260):38–39. pmid:19727184
  116. 116. Fuyuno I, Cyranoski D. Cash for papers: putting a premium on publication. Nature. 2006;441(7095):792–792. pmid:16778850
  117. 117. Nasir S, Ahmed J. Incentives matter: The role of research productivity award in increasing scientific output of Pakistani scientists. Science, Technology and Development. 2013;32(3):251–6.
  118. 118. Haq IU, Faridi RA. Evaluating the Research Productivity of Pakistan in the 21st Century. In: Chisita CT, Enakrire RT, Durodolu OO, Tsabedze VW, Ngoaketsi JM, editors. Handbook of Research on Records and Information Management Strategies for Enhanced Knowledge Coordination. Hershey, PA, USA: IGI Global; 2021. p. 407–423.
  119. 119. Memon AR, Rathore FA. The rising menace of scholarly black-market: Challenges and solutions for improving research in low-and middle-income countries. Journal of Pakistan Medical Association. 2021;71(6):1523–1526. pmid:34111064
  120. 120. Archambault E. Science Metrix, 30 Years in Science Secular Movements in Knowledge Creation. 2010. Available from: www.science-metrix.com/30years-paperd.pdf.
  121. 121. Moed HF. Iran’s scientific dominance and the emergence of South-East Asian countries as scientific collaborators in the Persian Gulf Region. Scientometrics. 2016;108(1):305–314.
  122. 122. Sawahel W. Iran: 20-year plan for knowledge-based economy. University World News. 2009.
  123. 123. Azadi Ahmadabadi G. Study and Evaluation of Quantitative and Qualitative Growth of Scientific Outputs of the Islamic Republic of Iran. Scientometrics Research Journal. 2022;8(16):265–286.
  124. 124. Azadi P, Mirramezani M, Mesgaran MB. Migration and brain drain from Iran. Stanford Iran. 2020;2040:1–30.
  125. 125. Sadeh S, Mirramezani M, Mesgaran M, Feizpour A, Azadi P. The Scientific Output of Iran: Quantity, Quality, and Corruption. Stanford Iran 2040 Project. 2019.
  126. 126. Dakhesh S, Hamidi A. Scientific misconduct and Iranian scientists. Gaceta Sanitaria. 2019;33(6):598. pmid:30926281
  127. 127. Kamali N, Talebi Bezmin Abadi A, Rahimi F. Plagiarism, Fake Peer-Review, and Duplication: Predominant Reasons Underlying Retractions of Iran-Affiliated Scientific Papers. Science and Engineering Ethics. 2020;26(6):3455–3463. pmid:33146787
  128. 128. Yaminfirooz M, Tirgar A. Self-citation in Iran in Comparison with Other Countries. Acta Inform Med. 2019;27(4):259–262. pmid:32055093
  129. 129. Tejada M. University Research Governance And The Colombian Scientific Journal Index Publindex. Understanding The Tensions [Thesis]. University of Twente; 2022.
  130. 130. Armijos Valdivieso P, Avolio Alecchi B, Arévalo-Avecillas D. Factors that Influence the Individual Research Output of University Professors: The Case of Ecuador, Peru, and Colombia. Journal of Hispanic Higher Education. 2022;21(4):450–468.
  131. 131. Montes IC, Garcia-Callejas D, Ocampo-Salazar C. Faculty Promotion Policy and the Academic Capitalist Regime: Professors’ Actions in Two Colombian Academic Departments. Higher Education Policy. 2022;
  132. 132. Pacheco I. Academic Salaries in Colombia: The Data Tell Only a Small Part of the Story: Iván F. Pacheco. In: Altbach P, Reisberg L, Yudkevich M, Androushchak G, Pacheco I, editors. Paying the Professoriate. Routledge; 2012. p. 114–123.
  133. 133. Elsevier. International Comparative Performance of the UK Research Base 2016. UK Department for Business, Energy and Industrial Strategy; 2017.
  134. 134. UK Department for Science Innovation and Technology. International Comparison of the United Kingdom Research Base, 2022.
  135. 135. Baccini A. Napoléon et l’évaluation bibliométrique de la recherche. Considérations sur la réforme de l’université et sur l’action de l’agence national d’évaluation en Italie. Canadian Journal of Information and Library Science-Revue Canadienne des Sciences de l’Information et de Bibliotheconomie. 2016;40(1):37–57.
  136. 136. Baccini A, De Nicolao G. Do they agree? Bibliometric evaluation versus informed peer review in the Italian research assessment exercise. Scientometrics. 2016;108(3):1651–1671.
  137. 137. Petrovich E. Bibliometrics in Press. Representations and uses of bibliometric indicators in the Italian daily newspapers. Scientometrics. 2022;
  138. 138. Garfield E. Citation indexing—its theory and application in science, technology, and humanities. Information sciences series. New York: Wiley; 1979.
  139. 139. Narin F. Evaluative bibliometrics: The use of publication and citation analysis in the evaluation of scientifc activity. Cherry Hill, NJ: Computer Horizons Inc; 1976.
  140. 140. Biagioli M, Kenney M, Martin BR, Walsh JP. Academic misconduct, misrepresentation and gaming: A reassessment. Research Policy. 2019;48(2):401–413.
  141. 141. Rijcke Sd, Wouters PF, Rushforth AD, Franssen TP, Hammarfelt B. Evaluation practices and effects of indicator use—a literature review. Research Evaluation. 2016;25(2):161–169.
  142. 142. Hicks D, Wouters P, Waltman L, de Rijcke S, Rafols I. Bibliometrics: The Leiden Manifesto for research metrics. Nature. 2015;520(7548):429–431. pmid:25903611
  143. 143. Burrows R. Living with the H-Index? Metric Assemblages in the Contemporary Academy. The Sociological Review. 2012;60(2):355–372.
  144. 144. Ioannidis JPA, Baas J, Klavans R, Boyack KW. A standardized citation metrics author database annotated for scientific field. PLOS Biology. 2019;17(8):e3000384. pmid:31404057
  145. 145. Koen Jonkers and Thomas Zacharewicz Research performance based funding systems: a comparative assessment. Publications Office of the European Union. 2017.