Peer Review History

Original SubmissionMarch 7, 2022
Decision Letter - Hugh Cowley, Editor

PONE-D-22-06759Building an annotated corpus for automatic metadata extraction from multilingual journal article referencesPLOS ONE

Dear Dr. Kong,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

Please see the reviewers' comments below. We ask that you ensure you address each of the reviewers' comments when revising your manuscript.

Please submit your revised manuscript by Oct 14 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.

If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Hugh Cowley

Staff Editor

PLOS ONE

Journal Requirements:

When submitting your revision, we need you to address these additional requirements.

1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at

https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and

https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf

2. Thank you for stating in your Funding Statement:

“This research was supported by the Korea Institue of Science and Technology Information (KISTI) of the Ministry of Science and ICT, South Korea (MSIT) (No. K-22-L01-C01-S01: Construction on Intelligent SciTech Information Curation). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.”

Please provide an amended statement that declares *all* the funding or sources of support (whether external or internal to your organization) received during this study, as detailed online in our guide for authors at http://journals.plos.org/plosone/s/submit-now.  Please also include the statement “There was no additional external funding received for this study.” in your updated Funding Statement.

Please include your amended Funding Statement within your cover letter. We will change the online submission form on your behalf.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: Yes

Reviewer #2: Yes

**********

2. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: Yes

Reviewer #2: N/A

**********

3. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: Yes

Reviewer #2: Yes

**********

4. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: Yes

Reviewer #2: Yes

**********

5. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: This work describes the use of data from a large database of manually annotated article references to create a corpus of reference strings and tagged token sequences from those reference strings which represent correct parses of the reference strings. The tags used are of the IOB type and are made specific to the thirteen field types studied. At this point the references are limited to references found in journal articles. Currently the corpus contains 3,815,987 multilingual references. While most references are in English there are a significant number in Korean, Chinese and Japanese. A few are in German or French. The creation of this corpus began with 4,078,624 references which were tokenized and automatically tagged based on the metadata contained in the original database. A set of 144,934 of these were selected for manual review by eight annotators working in pairs. The pairs of annotators had a high level of agreement in deciding if references were correctly parsed (Cohen’s Kappa average of 0.903). From this work 135,367 references were either correctly parsed or could be corrected to be correctly parsed and thus usable and correct to a human standard. The remaining references were subjected to an automatic screening process and 3,680,620 were judged correct based on this screening. These were used to train a BERT model to automatically tag tokenized reference strings. The 135,367 references were divided into validation and tests and the trained BERT model predicted the test data with an F1 of 99.83. The BERT model was compared with the GROBID algorithm for the same task and BERT proved to be much better.

The paper is well written in good English and details are clearly explained.

Suggestions: 1) Is locating the references strings in journal articles a problem? If so, is your data helpful in solving it?

2) You state that previous methods of parsing reference strings do not perform sufficiently well to be used without human intervention. Does your BERT model produce results good enough to be used without a human examining results? Are you using it in your curation process for your KISTI database?

3) p. 12, line 183 “each group delicately examined” would be better “each group carefully examined.”

Reviewer #2: The paper presents a procedure to build an annotated corpus of bibliographic references using

metadata. The authors proposed reference labeling using pre-trained BERT models.

The topic is relevant, and the article explores promising ideas in the context of the annotated corpus.

All assessment scenarios have better results (Tab 10, 11, and 12). In this sense, doubts about

the work results can be pointed out. One suggestion is to propose an extra analysis to validate

the results through textual data visualization techniques. For example, agglomerative clustering

methods can be proposed to visualize the results.

**********

6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #2: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Revision 1

Reviewer#1, comment #1: Is locating the references strings in journal articles a problem? If so, is your data helpful in solving it?

[Author response] Thank you for the comment. We have collected, stored and managed all journal articles in PDF format. Therefore, it is necessary to find and extract the list of references included in the PDF before the metadata extraction process starts. We have been collecting journal articles from over 500 domestic journals annually. Because each journal has a different design of the paper and reference numbering style, it is difficult to automatically locate and extract the list of references in journal articles. Therefore, we have manually extracted them thus far. The corpus mentioned in this paper is built to extract metadata automatically under the assumption that a single reference is given. Recently, we started a study aimed extracting the bibliography list from pdf papers automatically. In more detail, we plan to construct an image-based corpus for detecting the locations of references and for separating into individual references. Using this corpus, we will train the object detection models such as YOLO. For extracting reference texts from the detected objects, we will use the OCR technologies. In the near future, we plan to fully automate the bibliographic metadata extraction process.

Reviewer#1, comment #2: You state that previous methods of parsing reference strings do not perform sufficiently well to be used without human intervention. Does your BERT model produce results good enough to be used without a human examining results? Are you using it in your curation process for your KISTI database?

[Author response] Thank you for the pertinent comment. The journal articles collected by KISTI include many non-English references and English. For these non-English references, existing methods did not correctly extract their metadata. On the other hand, as mentioned in the paper, the metadata extracted by BERT-REF1 were almost identical to the strings of answer metadata verified by humans. In addition, after we conducted this research, we continuously accumulated the results of both the manual extraction of metadata and the automatic extraction using BERT-REF1 for several months. In conclusion, it was found that BERT-REF1 produced sufficiently good results to be used without human intervention. Currently, the metadata for given references are automatically extracted without any human intervention and accumulated in the KISTI DB. Therefore, BERT-REF1 is now applied to fully automate the metadata extraction of references during the curation process for the KISTI Database.

Reviewer#1, comment #3: p.12, line 183 “each group delicately examined” would be better “each group carefully examined.”

[Author response] Thank you for the comment. We have corrected the term accordingly. We changed the term "delicately" to "carefully."

Reviewer#2, comment #1: All assessment scenarios have better results (Tab 10, 11, and 12). In this sense, doubts about the work results can be pointed out. One suggestion is to propose an extra analysis to validate the results through textual data visualization techniques. For example, agglomerative clustering methods can be proposed to visualize the results.

[Author response] Thank you for the suggestion. Tables 10, 11, and 12 show the accuracies calculated by the annotators manually comparing the answer metadata with the metadata predicted by GROBID and BERT-REF1. Therefore, questions concerning the the results can arise because they result from manual evaluation by annotators. Therefore, we attempted to perform an additional analysis to validate the results through textual data visualization techniques such as agglomerative clustering. However, we found it challenging to visualize the validation results using clustering methods because our experiment, as shown in Tables 10, 11, and 12, simply aimed to verify equality between the answer metadata and the predicted metadata by each model. Therefore, we performed another experiment that automatically compares the string similarities between the answer metadata and the predicted metadata by each model.

In the experiment, we automatically calculated the string similarities between the answer metadata and the metadata predicted by BERT-REF1 and GROBID using the Levenshtein distance. The answer metadata for the 12,887 references mentioned in the paper can be considered accurate and reliable since the KISTI database has been regularly evaluated by the Korea Data Agency, a government-affiliated organization, to check the DB quality, and the administrator of this research additionally inspected them. Thus, it can be considered that the higher the similarity with the answer metadata is, the better the model performance. Figure 8 shows the similarities according to the metadata field types. Similar to the manual evaluation results shown in Tables 10, 11, and 12, BERT-REF1 slightly outperformed the similarity performance of GROBID for the English references. However, for non-English references, BERT-REF1 showed significantly better performance than GROBID. In particular, it was found that the similarity performance between the two models for author names, titles, and journal names in non-English references was significantly different. We also visualized the similarity difference between BERT-REF1 and GROBID for the above three metadata field types in 3,348 non-English references, as shown in Figure 9. In Figure 9, the red similarity points of metadata predicted by BERT-REF1 were distributed closer to 1.0 than GROBID. This is because GROBID could not extract accurate metadata from non-English references. Based on this experiment, we concluded that BERT-REF1 outperformed GROBID not only in English references but also in non-English references. We also added Excel sheets containing the answer metadata and the metadata predicted by BERT-REF1 and GROBID for all 12,887 references as supplementary information.

Attachments
Attachment
Submitted filename: Response to Reviewers.doc
Decision Letter - Sanaa Kaddoura, Editor

PONE-D-22-06759R1Building an annotated corpus for automatic metadata extraction from multilingual journal article referencesPLOS ONE

Dear Dr. Kong,

Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process.

The paper needs to be proofread for English enhancements.

Please submit your revised manuscript by Jan 20 2023 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file.

Please include the following items when submitting your revised manuscript:

  • A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'.
  • A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'.
  • An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'.
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter.

If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols.

We look forward to receiving your revised manuscript.

Kind regards,

Sanaa Kaddoura

Academic Editor

PLOS ONE

Journal Requirements:

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

[Note: HTML markup is below. Please do not edit.]

Reviewers' comments:

Reviewer's Responses to Questions

Comments to the Author

1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation.

Reviewer #1: All comments have been addressed

Reviewer #3: All comments have been addressed

**********

2. Is the manuscript technically sound, and do the data support the conclusions?

The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented.

Reviewer #1: (No Response)

Reviewer #3: Yes

**********

3. Has the statistical analysis been performed appropriately and rigorously?

Reviewer #1: (No Response)

Reviewer #3: Yes

**********

4. Have the authors made all data underlying the findings in their manuscript fully available?

The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified.

Reviewer #1: (No Response)

Reviewer #3: No

**********

5. Is the manuscript presented in an intelligible fashion and written in standard English?

PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here.

Reviewer #1: (No Response)

Reviewer #3: Yes

**********

6. Review Comments to the Author

Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters)

Reviewer #1: (No Response)

Reviewer #3: the author (s) improved the paper based on the received comments. the paper can be published after improving language.

**********

7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files.

If you choose “no”, your identity will remain anonymous but your review may still be made public.

Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy.

Reviewer #1: No

Reviewer #3: No

**********

[NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.]

While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.

Revision 2

[Comments from Editor]

Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice.

Author response: In response to the above comments, we reviewed our reference list to ensure that it is complete and correct.

Author action: We updated some of the references as follows:

(reference #3, page 19) Pranckut˙e, R Web of Science (WoS) and Scopus: The titans of bibliographic information in today’s academic world. Publications. 2021; 9(1):12. https://doi.org/10.3390/publications9010012

(reference #5, page 19) Khabsa M, Giles CL. The Number of Scholarly Documents on the Public Web. PLoS ONE. 2014; 9(5):e93949. https://doi.org/10.1371/journal.pone.0093949

(reference #18, page 20) Sennrich R, Haddow B, Birch A. Neural machine translation of rare words with subword units. arXiv:1508.07909 [Preprint] 2016. Available from: https://arxiv.org/abs/1508.07909

[Comments from Reviewer 1]

(No Response)

Author response: The reviewer 1 did not give any comment.

[Comments from Reviewer 3]

The author (s) improved the paper based on the received comments. the paper can be published after improving language.

Author response: In response to the above comments, all spelling and grammatical errors pointed out by the reviewers have been corrected. We highlighted all the changes within the revised manuscript.

Attachments
Attachment
Submitted filename: Response to Reviewers.doc
Decision Letter - Sanaa Kaddoura, Editor

Building an annotated corpus for automatic metadata extraction from multilingual journal article references

PONE-D-22-06759R2

Dear Dr. Kong,

We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements.

Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication.

An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org.

If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org.

Kind regards,

Sanaa Kaddoura

Academic Editor

PLOS ONE

Additional Editor Comments (optional):

Reviewers' comments:

<quillbot-extension-portal></quillbot-extension-portal>

Formally Accepted
Acceptance Letter - Sanaa Kaddoura, Editor

PONE-D-22-06759R2

Building an annotated corpus for automatic metadata extraction from multilingual journal article references

Dear Dr. Kong:

I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department.

If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org.

If we can help with anything else, please email us at plosone@plos.org.

Thank you for submitting your work to PLOS ONE and supporting open access.

Kind regards,

PLOS ONE Editorial Office Staff

on behalf of

Dr. Sanaa Kaddoura

Academic Editor

PLOS ONE

Open letter on the publication of peer review reports

PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.

We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.

Learn more at ASAPbio .