Skip to main content
Advertisement
Browse Subject Areas
?

Click through the PLOS taxonomy to find articles in your field.

For more information about PLOS Subject Areas, click here.

  • Loading metrics

Enabling transparent research evaluation: A method for historical RCR retrieval using public NIH metadata

  • Fabian Haupt ,

    Roles Conceptualization, Data curation, Investigation, Project administration, Visualization, Writing – original draft

    Fabian.Haupt.MD@gmail.com

    Affiliation Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland

  • Jan F. Senge,

    Roles Data curation, Validation, Writing – review & editing

    Affiliations Department of Mathematics and Computer Science, University of Bremen, Bremen, Germany, Dioscuri Centre for Topological Data Analysis, Warsaw, Poland

  • Hendrik von Tengg-Kobligk,

    Roles Project administration, Resources, Writing – review & editing

    Affiliation Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland

  • Wolfram A. Bosbach

    Roles Conceptualization, Formal analysis, Methodology, Software, Writing – original draft

    Affiliations Department of Diagnostic, Interventional and Pediatric Radiology, Inselspital, Bern University Hospital, University of Bern, Bern, Switzerland, Department of Nuclear Medicine, Inselspital, Bern University Hospital, University of Bern, Switzerland

Abstract

The demand for improvement of research output evaluation by the international science community has been recently formulated as the Declaration on Research Assessment (DORA) by the American Society for Cell Biology. The Relative Citation Ratio (RCR), introduced by the National Institutes of Health (NIH) is a novel metric indicating the influence of a publication on its peer group — publications from the same research field, as determined by co-citation analysis. The RCR can be viewed for the actual month. While historical RCR data exists within the NIH database, it is not exposed in a way that allows direct or simple access for researchers. We present a method to reconstruct otherwise inaccessible RCR data. A Python-based approach was deployed to extract RCR data from the NIH database and to plot RCR-values for every time point since introduction of the database. This method demonstrates the feasibility of recovering historical bibliometric information and may contribute to more transparent and accountable use of metrics in academic evaluation — in line with the goals of DORA and open science initiatives.

Introduction

The Relative Citation Ratio (RCR), introduced by the National Institutes of Health (NIH), was developed to address key limitations of traditional citation-based metrics by providing a field-normalized, article-level measure of influence [1,2]. The RCR is defined as the total number of citations divided by the yearly number of citations of publications in the same field — defined by the NIH. It is article specific and only accounts for PubMed-listed sources [1]. As such, the RCR serves as a field-normalized measure of individual article influence — making it increasingly relevant in academic evaluation and funding decisions. An RCR above 1.0 reflects citation rates above the median of PubMed-listed publications of the same scientific area.

In addition to the RCR, the h-index remains one of the most commonly applied bibliometric indicators in academic evaluation. It is defined as the maximum value h such that a researcher has published h articles that have each been cited at least h times. This composite measure is intended to capture both productivity (number of publications) and citation impact (number of citations), and therefore continues to be frequently used in promotion and funding contexts [3].

The Declaration on Research Assessment (DORA), developed in 2012 by the American Society for Cell Biology, recognized the need for improvement of research output evaluation (sfdora.org) [4].

In Switzerland, DORA has already been signed by 60 organizations (status of December 2024). These include various foundations, such as the Swiss National Science Foundation (SNF), as well as numerous universities.

Quantitative metrics play a central role in academic decision-making, yet concerns have been raised about their transparency, fairness, and reproducibility — as highlighted by initiatives such as DORA. The University of Bern has adopted DORA in the Faculty of Medicine and included the RCR 2018 in the regulations for habilitation [5]: To successfully achieve habilitation, a candidate must — among other items — provide at least two publications with an RCR of 1 or above [6].

The common approach to obtain the RCR from the National Institutes of Health (NIH) is the utilization of iCite, https://icite.od.nih.gov/ an online tool provided directly by the NIH [7]. However, iCite provides only the current RCR, with no access to historical values — creating a gap for those seeking to verify past performance thresholds or assess citation dynamics over time.

Thus, a habilitation candidate could meet the RCR criterion but not provide the necessary proof. To address this transparency gap, we present an open and reproducible method to extract historical RCR data from NIH archives — supporting traceable, data-driven research assessment. The underlying code has been written under assistance of the large language model (LLM) ChatGPT (Chat Generative Pre-trained Transformer) by OpenAI (San Francisco, CA, USA), a language model-based bot [8].

The method presented is non-commercial, simple to apply and intended to provide researchers access to historical RCR data.

Materials and methods

Study design

Retrospective analysis of a database provided by the NIH.

Ethics

No human health related data or animal experiments were involved. The data repository/ structure/ storage and handling are explained below.

Local hardware

This study used an Intel Core i5-6200U central processing unit (CPU) at 2.30 GHz with 8.00 GB random access memory (Intel Corporation, Santa Clara, CA, USA).

Accessing data storage

The database for the actual month was accessed through the online interface of iCite (https://icite.od.nih.gov/analysis, date of access October 13th 2023) (6).

For historic RCR-values, databases were obtained from the NIH figshare deposit (https://doi.org/10.35092/yhjc.c.4586573) (13), which is updated on a regular basis by the NIH. Each month three files are deposited:

  • Metadata twice (as.zip and as.tar.gz),
  • as well as the open citation collection.

The relevant file containing the data for accessing historic RCR-values is the metadata.

Data structure

Each metadata file is for its respective date a complete register of all PubMed-listed manuscripts, sorted by the PubMed ID (PMID) including benchmark information for each individual manuscript, such as RCR, total citation count, or year of publication.

Data download and storage

The metadata file name (icite_metadata.zip) does not include any identifier for the respective date, therefore each file was stored in a folder system with one directory per month. The folders were named after their respective time point in the format YYYYMM.

Code generation

Due to the size of the databases contained in the metadata files, a computer code was written for data extraction. The code applied for RCR-value extraction was written in Python language (Python Software Foundation, DE, USA) by W.A.B. Running of the code was performed in Spyder (v5.4.3, Anaconda, TX, USA). Metadata from 10/2022–07/2023 was analyzed. All code and data handling steps were designed to support reproducibility and align with open science principles. Code debugging was supported by the LLM [8]. It has demonstrated in the past a cetrain level of competence in various medicine related tasks [9].

RCR-value extraction/ code-1.py

Running code-1.py starts a search in the monthly databases from top to bottom for the specified PMID. A local copy is saved for each month of the database entry. Searching one metadata file for one PMID required in this study approximately 20 minutes.

S1 File supplementary example code of code-1 (Publication A).

S3 File: extracted data entries (Publication A, and B)

Data summary/ code-2.py

Running code-2.py allows data aggregation and the making of Table 1 and Fig 2.

thumbnail
Table 1. RCR and citation count, retrieved from the iCite databases.

https://doi.org/10.1371/journal.pone.0340697.t001

S2 File supplementary example code of code-2 (Publication A, and for Publication B).

ChatGPT coding help and README file

As part of the open access supplement, a file named read.me is included with detailed instructions for replicating the published code. Based on the provided example of PMID 34067405, code variations can be created and run for other (e.g., any) PMID. For debugging of code-1.py and code-2.py, ChatGPT 3.5 was employed.

Historic RCR-value extraction

The previously described steps were applied on two publications of our group: Publication A (PMID 31278348) Haupt et. al. from 2019 [10] with a maximum RCR of 0.9 in 07/2021, and Publication B Bosbach et. al. (PMID 34067405) [11] with a maximum RCR of 1.2 in 10/2022. Historic RCR-values were obtained in table format (*.csv).

Results

Feasibility of historic RCR-value extraction

Historic RCR-values of two sample publications, Publication A (PMID 31278348) from the field of Cellular Biology and Publication B (PMID 34067405) from the field of Public Health Sciences, were successfully retrieved using the approach outlined in the Methods section. These values, which are not directly accessible through NIH’s iCite tool, were reconstructed by systematically parsing archived NIH metadata files. The extracted data spans from the publication dates to July 2023 and is presented in Table 1.

RCR data over time

For publication A, the RCR started from 0.86 in September 2020 at two citations with an all-time maximum at 0.90 in July 2021 with a total count of 7 PubMed-listed citations at that time point. The RCR declined after this time point, receiving 4 more citations till 2023−07. The course of the number of citations and the RCR-value over time are displayed in Fig 1.

thumbnail
Fig 1. RCR-values and citation count of Table 1 for publication A plotted by Python over time, reconstructed from NIH metadata using Python.

https://doi.org/10.1371/journal.pone.0340697.g001

From June 2020 on, a “saw-tooth profile” was observed. The RCR climbed multiple times whenever an additional citation was found and, dropped again in the following months where no new citations were found but peer group citations increased.

For publication B, the RCR started in Oct 2021 above the 50th percentile with a value of 1.2 from two PubMed-listed citations. The publication was not cited after this time point and the RCR continuously dropped. The course of the RCR-value over time of publication B is displayed in Fig 2.

thumbnail
Fig 2. RCR-values of Table 1 for publication B plotted by Python over time.

https://doi.org/10.1371/journal.pone.0340697.g002

Discussion

Relevance of the RCR

The h-index remains the most established bibliometric indicator for academic evaluation, combining productivity and citation impact into a single number. However, it has well-recognized limitations: it favors senior researchers over early-career scientists, varies substantially across disciplines, and is insensitive to highly cited individual publications. In contrast, the RCR provides an article-level, field-normalized measure of influence, thereby enabling more equitable comparisons across researchers at different career stages and in different domains. This conceptual distinction has led to increasing interest in RCR as a complementary or alternative metric to the h-index in promotion and funding contexts [1].

Historical RCR extraction from NIH database

This project demonstrates the feasibility of extracting historical Relative Citation Ratio (RCR) values from publicly available NIH metadata. By providing a transparent, script-based approach to reconstruct monthly RCR values, this method addresses a significant gap in the bibliometric infrastructure: the absence of accessible historical citation influence data. While the NIH’s iCite tool provides a valuable overview of RCR at a single time point, it does not support longitudinal analysis or retrospective benchmarking [1,7]. Our contribution offers a reproducible solution that aligns with the goals of responsible metrics frameworks such as DORA and the Leiden Manifesto [4,12].

Practical importance

The practical need for such retrospective data is evident in academic evaluation and research policy. RCR has increasingly been used to assess individual academic performance in a field-normalized manner. For example, in ophthalmology, Patel et al. showed that mean and weighted RCR better reflect individual impact than the h-index, particularly for early-career researchers [2]. Similar applications exist in plastic surgery, neurosurgery, and orthopedic hand surgery [1315]. Our method can complement these efforts by enabling time-resolved evaluations — such as whether a publication surpassed an RCR threshold during a candidate’s qualifying period for promotion or grant eligibility.

The RCR has also been applied by funding agencies such as the NIH to evaluate research impact at the portfolio level. In program evaluations, RCR aggregates have been used to assess whether certain grant mechanisms or funding strategies correlate with greater scientific influence [16]. The availability of historical RCR data may enable novel retrospective analyses, including tracking time-based citation influence or comparing the performance of funding programs over distinct periods.

Limitations to the presented process

From a technical perspective, our method leverages publicly available resources provided by the NIH, including monthly iCite snapshots published via Figshare [17]. While real-time RCR values can be accessed via the iCite API, historical data reconstruction is not natively supported. Our code provides a missing link by automating access to and aggregation of these data over time. Existing tools such as the R package iCiteR enable programmatic access to RCR, but do not reconstruct RCR trajectories — a gap our system addresses.

Further, integration of RCR into widely used platforms like Dimensions and ReadCube has expanded its visibility and utility [18]. Our method enhances this ecosystem by allowing institutions to analyze not just current RCR values, but also their historical progression. This enables applications in academic dashboards, evaluation audits, and institutional meta-research.

The main limitation of our method lies in its manual steps — downloading, storing, and processing monthly metadata files. While functional, these steps could be further streamlined via automation or integration into platforms like iCite. We encourage the NIH and bibliometric tool developers to explore this as a valuable feature extension. In terms of scalability, the process is primarily limited by local hardware performance and internet bandwidth. Since data retrieval relies on stable NIH figshare repositories, error margins are negligible, and adapted Python code can be executed on standard computing environments. What has been demonstrated for individual publications can, in principle, be expanded to the research output of individual scholars, research groups, or institutions, supporting large-scale bibliometric analyses.

Limitations to the explanatory power of RCR

It is important to note that RCR, while valuable, has its limitations. It only includes PubMed-indexed publications and therefore may underrepresent influence in disciplines not well covered by PubMed indexing. At the same time, in biomedical fields where PubMed coverage is comprehensive, this focus can be seen as an advantage, since it benchmarks against the corpus most relevant for academic success in these domains. Additionally, RCR depends on co-citation networks, which may not form reliably in newly emerging or interdisciplinary fields [1]. Beyond this, the metric is bounded by contemporary citation practices — it reflects what is deemed “worth citing” at the time, which may shift over years and across disciplines.

Comparison groups underlying RCR calculations are not retrievable by users form the NIH data, as the NIH does not disclose the precise peer set used for normalization. This lack of transparency contradicts reproducibility as a major scientific requirement and limits independent verification of the metric’s validity.

Another important limitation concerns the time lag with which RCR values are provided by the NIH. Since the metric depends on accumulating co-citation networks and benchmark citation rates, newly published articles only receive an RCR after a delay of several months. This latency reduces the suitability of RCR for assessment, particularly in fast-moving research areas. Research bodies should therefore consider for which time windows alternative indicators (e.g., raw citation counts, field-normalized citation scores from other sources) or carefully defined ad hoc values may provide a more appropriate proxy until RCR estimates become available.

Furthermore, while historical RCR values can illuminate citation trajectories, they should not be misapplied for retrospective evaluation in ways that contradict DORA’s principles, since such use risks reinforcing metric-driven assessments rather than contextualized evaluation. Therefore, RCR should be used as one of several indicators and interpreted in context — as recommended by both DORA and the Leiden Manifesto [4,12].

Conclusions

In conclusion, our method provides a practical and reproducible solution to retrieve historical RCR values. This supports more transparent, time-aware, and traceable use of citation metrics in both academic evaluation and funding contexts. By bridging this data gap, our work contributes to the broader movement toward responsible, open, and reproducible research assessment. At the same time, the inherent limitations of RCR — including its restriction to PubMed sources and the non-disclosure of comparison groups by the NIH — highlight challenges to reproducibility that must be considered when applying the metric.

Supporting information

S2 File. Python code-2 (Publication A, and B).

https://doi.org/10.1371/journal.pone.0340697.s002

(PY)

S3 File. Extracted data entries (Publication A, and B) The study’s raw data is deposited for open access on figshare under the DOI:

https://doi.org/10.6084/m9.figshare.30067852.

https://doi.org/10.1371/journal.pone.0340697.s003

(PY)

Acknowledgments

The authors wish to thank colleagues for all the useful discussions leading to this manuscript.

References

  1. 1. Hutchins BI, Yuan X, Anderson JM, Santangelo GM. Relative Citation Ratio (RCR): A new metric that uses citation rates to measure influence at the article level. PLoS Biol. 2016;14(9):e1002541. pmid:27599104
  2. 2. Patel PA, Ali MJ. The Relative Citation Ratio: A brief primer on the national institutes of health-supported bibliometric. Semin Ophthalmol. 2022;37(5):539–40. pmid:35695556
  3. 3. Hirsch JE. An index to quantify an individual’s scientific research output. Proc Natl Acad Sci U S A. 2005;102(46):16569–72. pmid:16275915
  4. 4. About DORA. [Cited 2025 May 23]. https://sfdora.org/
  5. 5. Steck N, Stalder L, Egger M. Journal- or article-based citation measure? A study of academic promotion at a Swiss university. F1000Res. 2020;9:1188. pmid:33149900
  6. 6. University of Berne. Habilitation regulations. 2024. https://www.medizin.unibe.ch/unibe/portal/fak_medizin/content/e17203/e529687/e907140/200604_Habilitationsreglement_Neuerlass_def_ger.pdf
  7. 7. NIH. iCite: Impact metrics for publications. 2024. [Cited 2025 May 23]. https://icite.od.nih.gov/
  8. 8. AI O chat GPT. [Cited 2025 May 23]. https://chat.openai.com/
  9. 9. Bosbach WA, Schoeni L, Beisbart C, Senge JF, Mitrakovic M, Anderson SE, et al. Evaluating the diagnostic accuracy of ChatGPT-4.0 for classifying multimodal musculoskeletal masses: a comparative study with human raters. Rofo. 2025:10.1055/a-2594–7085. pmid:40461006
  10. 10. Haupt F, Krishnasamy K, Napp LC, Augustynik M, Limbourg A, Gamrekelashvili J, et al. Retinal myeloid cells regulate tip cell selection and vascular branching morphogenesis via Notch ligand Delta-like 1. Sci Rep. 2019;9(1):9798. pmid:31278348
  11. 11. Bosbach WA, Heinrich M, Kolisch R, Heiss C. Maximization of open hospital capacity under shortage of SARS-CoV-2 Vaccines-an open access, stochastic simulation tool. Vaccines (Basel). 2021;9(6):546. pmid:34067405
  12. 12. Hicks D, Wouters P, Waltman L, de Rijcke S, Rafols I. Bibliometrics: The leiden manifesto for research metrics. Nature. 2015;520(7548):429–31. pmid:25903611
  13. 13. Herzog I, Park JB, Pei M, Didzbalis CJ, Reed LT, Weisberger J, et al. Utilizing relative citation ratio to compare academic productivity by gender in plastic surgery. J Surg Res. 2024;303:141–7. pmid:39340939
  14. 14. McNamara JK 3rd, Henderson MN, Sudah SY, Faccone RD, Michel CR, Dijanic C, et al. Evaluation of the relative citation ratio among academic orthopedic hand surgeons: a novel measure of research impact. Cureus. 2022;14(5):e25362. pmid:35774708
  15. 15. Reddy V, Gupta A, White MD, Gupta R, Agarwal P, Prabhu AV, et al. Assessment of the NIH-supported relative citation ratio as a measure of research productivity among 1687 academic neurological surgeons. J Neurosurg. 2020;134(2):638–45. pmid:32005024
  16. 16. Heyard R, Hottenrott H. The value of research funding for knowledge creation and dissemination: A study of SNSF research grants. Humanit Soc Sci Commun. 2021;8(1).
  17. 17. Hutchins BIGS. iCite database snapshots (NIH open citation collection). 2019. https://doi.org/10.35092/yhjc.c.4586573
  18. 18. Digital Science. 2025. www.digital-science.com