Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

A Large-Scale Analysis of Impact Factor Biased Journal Self-Citations

Abstract

Based on three decades of citation data from across scientific fields of science, we study trends in impact factor biased self-citations of scholarly journals, using a purpose-built and easy to use citation based measure. Our measure is given by the ratio between i) the relative share of journal self-citations to papers published in the last two years, and ii) the relative share of journal self-citations to papers published in preceding years. A ratio higher than one suggests that a journal’s impact factor is disproportionally affected (inflated) by self-citations. Using recently reported survey data, we show that there is a relation between high values of our proposed measure and coercive journal self-citation malpractices. We use our measure to perform a large-scale analysis of impact factor biased journal self-citations. Our main empirical result is, that the share of journals for which our measure has a (very) high value has remained stable between the 1980s and the early 2000s, but has since risen strongly in all fields of science. This time span corresponds well with the growing obsession with the impact factor as a journal evaluation measure over the last decade. Taken together, this suggests a trend of increasingly pervasive journal self-citation malpractices, with all due unwanted consequences such as inflated perceived importance of journals and biased journal rankings.

Introduction

It is well known that the impact factor–arguably the single most important measure for assessing the quality or impact of scholarly journals–is vulnerable to all sorts of manipulation ([1], [2], [3]). The impact factor can for instance be manipulated by publishing large numbers of so-called non-citable articles ([4], [5], [6]). Manipulation is also possible by publishing editorials with many journal self-citations to recently published papers ([7], [8]) or by participating in citation cartels ([9], [10], [11]). Another source of manipulation which received considerable attention lately, is that of coercive journal self-citation ([12], [13], [14], [15]). The phenomenon of coercive journal self-citation has been discussed in various fields of science, ranging from Medicine ([13]) to Information Systems ([16]), Social Sciences ([14]), Sociology ([17]), and Transportation ([18]). Several forms of coercive journal self-citation can be distinguished: for example, a journal may highlight on its website that it requires authors to position their paper with respect to recent papers published in that same journal. Or a journal’s editor, who may perhaps feel pressured by his or her publisher to increase the impact factor of his or her journal, may request authors of conditionally accepted papers to add references to papers recently published in the journal. In response to (anticipated) coercive citation practices, authors may behave strategically by adding references to papers recently published in the journal to which they plan to submit their work, to increase the chance of surviving the (editorial) review process.

This paper studies the phenomenon of ‘impact factor biased journal self-citations’. We use this term to refer to the phenomenon of journals having a disproportionally large proportion of journal self-citations to the past two years, which are the years that determine the impact factor of a journal, relative to their proportion of journal self-citations to earlier years. Impact factor biased journal self-citations may result from coercive citation practices and authors’ strategic response to such (anticipated) practices. They may also result from journal self-citations in manipulated editorials or simply from a tendency of editors to be more inclined to accept submissions that include many references to papers published recently in their journal. However, as we will discuss, there are also legitimate mechanisms that may result in impact factor biased journal self-citations.

Our study contributes to the growing literature on impact factor manipulation in the following two ways. First, we present an easy to use measure of impact factor biased journal self-citations. As we will show, by means of an empirical comparison of our measure with recently reported results of an author survey into coercive journal self-citation malpractices, the measure provides a useful tool to identify (from a large set of candidate journals) a small subset of journals that are relatively likely to engage in self-citation malpractices. As such, it provides a practical tool for a first diagnosis of, for example, the coercive journal self-citation phenomenon. Second, we apply our measure in a large scale study of long term trends in impact factor biased journal self-citations, based on three decades of citation data for thousands of journals. Our empirical analyses show that impact factor biased journal self-citation practices have become much more prevalent over the past decade. This corresponds well with the time frame during which the impact factor has gained widespread attention among researchers and journal editors [19]. It therefore seems likely that the increase in impact factor biased journal self-citation practices relates to the increasing importance of impact factors and reflects increasingly pervasive journal self-citation malpractices.

Results

Consider a measure of Impact Factor Biased Self-citation Practices (from here on: IFBSCP). This measure (see Materials and Methods section for a formal derivation and interpretation) compares the share of journal self-citations in year y to papers published in impact factor years (i.e., year y − 1 and y − 2), with the share of journal self-citations to papers published in the five preceding years. The IFBSCP of a journal equals 1 if the share of journal self-citations to papers published in impact factor years equals the share of journal self-citations to papers published in preceding years. An IFBSCP that is (much) above 1 signals a disproportional share of journal self-citations to papers published in impact factor years. Although there may be perfectly legitimate reasons for a high IFBSCP (see Discussion section), it is a possible indication of malpractices such as manipulated editorials, coercive citation practices and authors’ strategic response to such (anticipated) practices. Table 1 (see S1 File and S2 File for the original data sources) reports the share of journals indexed in the Web of Science database which, in a given year, have an IFBSCP higher than a given threshold. It also presents the mean IFBSCP across journals for a given year.

thumbnail
Table 1. Trends in IFBSCP since 1987 (all scientific fields combined).

https://doi.org/10.1371/journal.pone.0161021.t001

A first observation is that a large majority of journals have an IFBSCP which is higher than 1; see third column from the left. As we discuss in the Discussion section, this can be explained by various legitimate mechanisms that may trigger an overrepresentation of journal self-citations to recently published papers. Nonetheless, it is directly seen that while the share of journals with an IFBSCP higher than given thresholds was relatively stable from 1987 to around 2004, it has markedly increased since then. For example, while between 1987 and 2004, around 33% of journals had an IFBSCP higher than 1.5, this share has rapidly increased since and is now 45%, implying a rise of more than 35%. This recent rise in the share of journals with an IFBSCP that surpasses a given threshold becomes even more pronounced for higher thresholds: for the threshold of 2, the share is stable at somewhat more than 10% between 1987 and 2004, while growing to almost 19% in 2015. For a threshold of 3, the share was stable at around 2% between 1987 and the early 2000s, and has since almost tripled to 5.6% in 2015. The mean IFBSCP, while being quite stable (and if anything, moderately declining) between 1987 and 2004, has since rapidly increased. Fig 1 provides a visualization of the trend concerning the share of journals with an IFBSCP that surpasses 3; Fig 2 does the same for the trend concerning mean IFBSCP.

thumbnail
Fig 1. Percentage of journals with IFBSCP > 3 (all fields of science combined).

https://doi.org/10.1371/journal.pone.0161021.g001

Our data allow us to explore if there are any eye-catching differences in (trends in) IFBSCP between the Physical, Life, and Social Sciences. More specifically, Fig 3 presents the time trend in the percentage of journals in a given domain, whose IFBSCP surpasses 3 in a given year. Fig 4 presents the time trend in the mean IFBSCP in a given domain.

thumbnail
Fig 3. Percentage of journals with IFBSCP > 3 (different fields of science).

https://doi.org/10.1371/journal.pone.0161021.g003

Inspection of Figs 3 and 4 warrants the following observations: first, all three fields exhibit the above described trends–more specifically, the pronounced increase in mean IFBSCP and in the percentage of journals with a high IFBSCP over the past decade. Second, compared to the Social and Physical Sciences domains, the Life Sciences domain exhibits a higher mean IFBSCP and a higher share of journals with a high IFBSCP. In particular, in the most recent years, the percentage of journals with an IFBSCP above 3 on average is almost two times higher in the Life Sciences than in the Physical Sciences.

Relation between IFBSCP and coercive journal self-citations

A survey was held recently, with the aim of exploring how pervasive the malpractice of coercive citation is perceived to be by scholars in selected fields in the Social Sciences, specifically Economics, Sociology, Psychology, and Business ([14]). It was found that about one in every five scholars responding to the survey has experienced coercion him- or herself, and that more than half of the scholars add journal-specific citations before submission. The authors conclude that coercion is “uncomfortably common”. Survey results reported in [14] can be used to study the relation between the IFBSCP measure and a journal’s reputation among scholars in terms of coercive citation practices. More specifically, Table S12 reported in the Supplementary material section in [14], provides a list of “Journals identified as coercers by survey respondents. Number of coercive observations represents the number of times a journal was identified by independent survey respondents as requesting self-citations that (i) give no indication that the manuscript was lacking in attribution, (ii) make no suggestion as to specific articles, authors, or a body of work requiring review, and (iii) only guide authors to add citations from the editor’s journal.” The Table contains 175 journals, from which we selected those 64 journals which met the following conditions: i) the number of coercive observations (as defined directly above) was larger than one, ii) the journal is indexed in the Web of Science database, iii) an IFBSCP value could be computed for the year 2011, going back to 2004 (in other words Y = 7 in the IFBSCP equation presented in the Materials and Methods section). We focus on the year 2011 because journals may have altered their practices in response to the publication of [14] in 2012. For each of these 64 journals, the 2011-IFBSCP was computed. Before we can compare the computed IFBSCP with the listed number of coercive observations per journal, the latter measure needs to be normalized to account for differences in journal sizes. More specifically, we divided the number of coercive observations by the number of papers published in a journal in 2011. This gives us a normalized measure of coercive observations per published paper (from here on denoted as COPPP).

The results of our analyses can be summarized as follows (see also Table 2): the mean IFBSCP for the 64 identified journals equals 1.93. To put this into perspective, it should be noted that the mean 2011-IFBSCP for the entire Social Sciences domain (which includes 581 journals with an IFBSCP in 2011) equals 1.54, implying that on average the IFBSCP of the 64 culprits identified in [14] is 25% higher than that of the average Social Science journal. Remember that the survey in [14] focused on Social Science journals. What’s more, 22 out of 64 identified culprits (i.e., more than a third) have an IFBSCP of at least 2.23, which is the 90th percentile threshold for the Social Sciences domain as a whole. Furthermore, when from the list of 64 journals only those 10 with the highest COPPP are considered, their mean IFBSCP equals 2.28; of these 10 journals, 5 have an IFBSCP of at least 2.23, so half of the 10 journals belong to the top 10% journals with the highest IFBSCP in the Social Sciences domain. On the other hand, 2 journals from this ‘top 10’ have an IFBSCP which is lower than the median Social Sciences IFBSCP; these two journals would not have been identified as potential culprits by an IFBSCP-based diagnosis. Finally, unsurprisingly, we find a positive correlation (of 0.19) between the IFBSCP and the COPPP in the population of 64 journals considered.

thumbnail
Table 2. Relation between the IFBSCP measure and reported coercive citation practices in [14].

https://doi.org/10.1371/journal.pone.0161021.t002

In sum, it turns out that the IFBSCP values of journals which were identified in an independent survey as being guilty of coercive citation practices, are considerably higher than those of the average journal in their field. This holds even more so for many, but not all, journals whose coercive citation practices were perceived to be particularly prevalent. These findings suggest that high values of the IFBSCP measure partly relate to coercive citation malpractices. This in turn warrants the conclusion that the IFBSCP measure may be used as a tool of first diagnosis of potential coercive citation malpractices.

Note that the data underlying our analyses are available as supplementary material to this paper (See S1 File and S2 File).

Discussion

This paper presents a measure, denoted Impact Factor Biased Self-citation Practices (IFBSCP), of the extent to which a journal has a disproportional share of journal self-citations to papers published in years used for computation of the impact factor, compared to papers published in preceding years. Increases in IFBSCP may be due to unethical attempts of journal editors to artificially inflate the impact factor of their journal (potentially in response to pressure from the publisher). For instance, a high IFBSCP may be caused by coercive citation practices ([12], [13], [14], [15]), by editorials with excessive numbers of journal self-citations ([7], [8]), or simply by editors’ acceptance decisions being biased in favor of submissions with many references to papers published recently in their journal. A high IFBSCP may also relate to questionable behavior of authors. Anticipating on the importance of journal self-citations to editors, authors may behave strategically and may include in their paper additional references to papers published recently in the journal to which they plan to submit their work.

However, it cannot be claimed that a high IFBSCP provides proof of unethical behavior of editors or authors. In fact, there are other, perfectly legitimate reasons why a journal may have a relatively high IFBSCP. A recent paper by one of us [18] presents two potential legitimate mechanisms that may trigger an ‘overrepresentation’ of journal self-citations in recent years: First, a reader of a journal may find a paper in its newest issue so interesting, that she decides to do follow up research, e.g. to explore the validity or properties of a recently proposed method, or to test the applicability of a recently reported empirical finding in a different geographical context. She is then of course more likely to submit the resulting paper to this journal, than to another one. This scenario would lead to an increase in the journal’s IFBSCP. Second, an author, after having written a paper, may take a look at her finalized reference list to find out where the most recent papers cited in her reference list have been published. She may consider that the fact that other papers about her paper’s topic were recently published in a particular journal is a signal of that journal’s current interest in the topic; in line with this consideration, she goes on to submit her work to that journal as opposed to some other journal. This scenario too would lead to an increase in the journal’s IFBSCP. A third potential mechanism which would cause high but legitimate IFBSCP values is as follows: in first instance, the spreading of new ideas is likely to take place in the community where the ideas were originally presented. Over time, the most influential ideas then gradually become noticed and gain traction in other scientific communities. This scenario would imply a relatively large number of journal self-citations in years directly after publication of an idea, compared to more distant (future) years.

Taking into account the above-mentioned legitimate reasons for high IFBSCP values, we stress that a high IFBSCP does not in itself prove wrongdoing by the associated journal. Nonetheless, we do believe that the IFBSCP measure has strong potential as a tool for first diagnosis: it can be used to quickly sift through a large set of candidate journals, and to identify the subset which is relatively likely to include potential culprits. This subset then should be subject to further scrutiny before journal-specific conclusions can be drawn. Our empirical comparison between the IFBSCP measure and the number of reported observations of coercive citation practices in a recent survey [14] provides a justification for this interpretation of the IFBSCP measure: journals identified in [14] as being prone to engage in coercive citation practices generally have a high value on the IFBSCP measure.

We apply the IFBSCP measure to analyze citation patterns for all scientific fields combined since 1987, as well as for the particular domains of Life, Physical, and Social Sciences. We believe that, notwithstanding the caution that should be exercised when interpreting the IFBSCP measure, two over-all conclusions are warranted based on our results: first, between 1987 and the early 2000s, the mean IFBSCP and the share of journals whose IFBSCP surpasses a given threshold has been quite stable; this suggests that there is a stable underlying set of legitimate mechanisms that together result in a mean IFBSCP of around 1.4 at the aggregate level. Second, since about 2004, there has been a pronounced rise in mean IFBSCP and particularly in the share of journals whose IFBSCP surpasses relatively high thresholds (of 1.5, 2, and 3); this suggests that during the last decade, and on top of the legitimate mechanisms discussed earlier, there have increasingly been other influences driving up the IFBSCP, the practices of coercive journal self-citation and strategic responses by authors being obvious candidates.

Although it cannot be ruled out that this steep recent increase in IFBSCP is due to profound changes in the legitimate mechanisms discussed earlier, it should be noted that the importance of the (two year) impact factor has increasingly gained attention among researchers, policy makers, and managers of scientific institutions over the past decade. For example, [19] show that the share of scholarly papers indexed in the Web of Science database mentioning the term “impact factor” in their title has increased steadily since the second half of the 1990s, suggesting a strongly increasing interest in the impact factor since the mid-1990s. Our results show an increasing trend in IFBSCP values starting in the first years of the new millennium. Hence, allowing for a delay of a few years, the increasing trend in IFBSCP values matches remarkably well with the increasing interest in the impact factor. That paper [19] also shows that journals in the Life Sciences publish many more editorials about the impact factor than journals in the Physical and Social Sciences, suggesting that interest in the impact factor is strongest in the Life Sciences. Again, this matches very well with our results, since we find that journals in the Life Sciences are more likely to have high IFBSCP values than journals in the Physical and Social Sciences.

Together, it seems fair to consider the strong rise in mean IFBSCP and in the share of journals with a (very) high IFBSCP, as circumstantial evidence for a trend of increasingly pervasive journal self-citation malpractices, with all due unwanted consequences such as inflated perceived importance of journals and biased journal rankings. We recommend that future studies into such malpractices use the IFBSCP measure to zoom in on a small subset of journals whose IFBSCP is suspiciously high, before embarking on the collection of much needed case study evidence for or against the actual malpractice at the level of particular journals.

Materials and Methods

The IFBSCP measure

Assuming that journal self-citation malpractices are aimed at inflating a journal’s impact factor, one would expect that a journal which actively engages in such a malpractice, has a share of journal self-citations to papers published in that journal in recent years that is relatively large compared with the share of journal self-citations to papers published in that journal in earlier years. This is the underlying rationale for the IFBSCP measure, which is formally denoted for a given journal as follows: where y represents the year for which the IFBSCP is computed; denotes the number of self-citations from papers published in the journal in year y to papers published in the same journal in year y – 1; denotes the number of self-citations from papers published in the journal in year y to papers published in the same journal in year y − 2; note that years y − 1 and y − 2 constitute the two years that are used for computing the journal’s impact factor. denotes the total number of citations from papers published in any journal in year y to papers published in the focal journal in year y – 1; , , , etc. are defined likewise, and Y stands for reach of the analysis (i.e., yY represents the earliest year included in the analysis).

Three comments need to be made on the above definition of the IFBSCP measure. First, the idea of analyzing the time trend in the ratio of self-citations to total citations is also used in recent studies on journal self-citations and journal citation cartels ([10], [15], [18]). Second, our measure differs from the procedure outlined in [18] in terms of how the denominator is defined; [18] uses all years preceding y in the analyses, which generates overlap between the years represented in the numerator and denominator. Third, the IFBSCP measure as defined above assumes that journal self-citation malpractices typically refer to papers from the past two years, which are the years that determine the traditional two year impact factor. Of course, following the same ideas as above, a variant of the IFBSCP measure for the five year impact factor can be easily derived.

Data

We extract citation data from the Web of Science database for all fields in the Sciences and the Social Sciences. We do not consider fields in the Arts and Humanities. Journals in the Arts and Humanities do not have an impact factor, and presumably editors of these journals therefore have less incentive to engage in coercive self-citation practices. We compute the IFBSCP for every journal and for every year in the period 1987–2015. We set Y = 7, so in effect we compare the share of journal self-citations to papers published in ‘impact factor-years’ with the share of journal self-citations to papers published in the five years preceding the ‘impact factor-years’ (see below for a stability analysis, using different values for Y). To focus our analyses on journals for which sufficiently reliable statistics can be computed, we exclude those journals that–in a given year–cite their own papers in the past seven years fewer than 50 times. This results in the exclusion of 88,842 out of 164,757 cases, where each case represents a combination of a journal and a year). The total number of journals considered increases from 1520 in 1987 to 4767 in 2015; the total number of cases in the period 1987–2015 being 75,915. For the analyses reported in Fig 3 and Fig 4, we make comparisons across different fields of science: this is based on 366 journals on average for the Social Sciences, 938 for the Physical Sciences, and 1532 for the Life Sciences. Note that we perform no tests of statistical significance of differences across domains and between years, since we are effectively observing the entire population of interest as opposed to a sample thereof. Note also that IFBSCP statistics for all journals included in our analyses are available from the authors upon request. Interested readers can use these statistics for follow-up analyses at the level of individual journals.

Stability analysis

The decision to set Y at 7 (implying a comparison between self-citations for the two impact factor years and the five preceeding years) is somewhat arbitrary. We therefore report the results of a series of stability analyses which we performed to investigate the impact on IFBSCP of setting Y at different values. Specifically, we consider the year 2015 (i.e., y in the IFBSCP formula equals 2015), and select all journals that continuously published during the previous decade, i.e., in each of the years 2005–2015. This results in 7223 journals. For each of these journals, we compute IFBSCPs for Y ∈ {4,5,6,7,8,9,10}. We report three different stability metrics: first, we report the Pearson correlation between the IFBSCP with Y set at 7 and each of the alternative specifications. These correlations range from 0.71 (Y = 4) to .95 (Y = 8); note that it was expected that the highest correlations would be found for values of Y close to 7, since the underlying citation data have large overlap. To study stability in rank ordering (i.e., whether the position of a journal in a rank ordering in terms of IFBSCP remains stable), we compute Kendall’s tau correlations as well. These correlations range from 0.69 (Y = 4) to 0.90 (Y = 8). Finally, we compute–for each journal–the ratio of the IFBSCP for Y = 7 and the IFBSCP for Y = 4 and Y = 10; i.e., we compare Y = 7 with the two most extreme values of Y in our data. The ratio (IFBSCP for Y = 7) / (IFBSCP for Y = 4) has a mean (across 7223 journals) of 1.16; the 10-th percentile IFBSCP equals 0.84, and the 90-th percentile IFBSCP equals 1.51. The ratio (IFBSCP for Y = 7) / (IFBSCP for Y = 10) has a mean (across 7223 journals) of 0.94; the 10-th percentile IFBSCP equals 0.78, and the 90-th percentile IFBSCP equals 1.07. In combination, these analyses indicate that the IFBSCP is stable across different values of Y.

Supporting Information

S1 File. Citation data.

Contains citation data from which Table 1 is computed.

https://doi.org/10.1371/journal.pone.0161021.s001

(XLSX)

S2 File. IFBSCP Results.

Contains the IFBSCPs reported in Table 1 and Figs 14.

https://doi.org/10.1371/journal.pone.0161021.s002

(XLSX)

Acknowledgments

We would like to thank the members of the Advanced Bibliometric Methods and the Evaluation Practices in Context working groups at the Centre for Science and Technology Studies (CWTS), Leiden University for their valuable feedback on this research.

Author Contributions

  1. Conceptualization: CC LW.
  2. Data curation: LW.
  3. Formal analysis: CC LW.
  4. Investigation: CC LW.
  5. Methodology: CC LW.
  6. Resources: LW.
  7. Software: LW.
  8. Visualization: CC LW.
  9. Writing – original draft: CC.
  10. Writing – review & editing: LW.

References

  1. 1. Falagas ME, Alexiou VG. (2008). The top-ten in journal impact factor manipulation. Archivum Immunologiae et Therapiae Experimentalis, 56(4), 223–226. pmid:18661263
  2. 2. Martin BR. (2016). Editors’ JIF-boosting stratagems–Which are appropriate and which not? Research Policy, 45(1), 1–7.
  3. 3. Vanclay JK. (2012). Impact factor: Outdated artefact or stepping-stone to journal certification? Scientometrics, 92(2), 211–238.
  4. 4. Heneberg P. (2014). Parallel worlds of citable documents and others: Inflated commissioned opinion articles enhance scientometric indicators. Journal of the Association for Information Science and Technology, 65(3), 635–643.
  5. 5. Moed HF, Van Leeuwen TN. (1996). Impact factors can mislead. Nature, 381, 186. pmid:8622752
  6. 6. PLoS Medicine Editors (2006). The impact factor game. PLOS Medicine, 3(6), e291. pmid:16749869
  7. 7. Campanario JM, González L. (2006). Journal self-citations that contribute to the impact factor: Documents labeled “editorial material” in journals covered by the Science Citation Index. Scientometrics, 69(2), 365–386.
  8. 8. González L, Campanario JM. (2007). Structure of the impact factor of journals included in the Social Sciences Citation Index: Citations from documents labeled “editorial material”. Journal of the American Society for Information Science and Technology, 58(2), 252–262.
  9. 9. Davis P. (2012). The emergence of a citation cartel. Available: http://scholarlykitchen.sspnet.org/2012/04/10/emergence-of-a-citation-cartel/
  10. 10. Heneberg P. (2016). From excessive journal self-cites to citation stacking: Analysis of journal self-citation kinetics in search for journals, which boost their scientometric indicators. PLOS ONE, 11(4), e0153730. pmid:27088862
  11. 11. Van Noorden R. (2013). Brazilian citation scheme outed. Nature, 500, 510–511. pmid:23985850
  12. 12. Mahian O, Wongwises S. (2015). Is it ethical for journals to request self-citation? Science and Engineering Ethics, 21(2), 531–533. pmid:24728611
  13. 13. Smith R. (1997). Journal accused of manipulating impact factor. BMJ, 314, 461. pmid:9056791
  14. 14. Wilhite AW, Fong EA. (2012). Coercive citation in academic publishing. Science, 335, 542–543. pmid:22301307
  15. 15. Yu T, Yu G, Wang MY. (2014). Classification method for detecting coercive self-citation in journals. Journal of Informetrics, 8(1), 123–135.
  16. 16. Gray P. (2009). Journal self-citation I: Overview of the journal self-citation papers–The wisdom of the IS crowd. Communications of the Association for Information Systems, 25, article 1.
  17. 17. Cohen PN. (2015). Journal self-citation practices revealed. Available: https://familyinequality.wordpress.com/2015/11/30/journal-self-citation-practices-revealed/
  18. 18. Chorus CG. (2015). The practice of strategic journal self-citation: It exists, and should stop (A note from the editor-in-chief). European Journal of Transport and Infrastructure Research, 15(3), 1–7.
  19. 19. Hicks D, Wouters P, Waltman L, De Rijcke S, Rafols I. (2015). The Leiden Manifesto for research metrics. Nature, 520, 429–431. pmid:25903611