Peer Review History
| Original SubmissionApril 9, 2020 |
|---|
|
PONE-D-20-10079 Using de-identified electronic health records to research mental health supported housing services: a feasibility study PLOS ONE Dear Mr Dalton-Locke, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. We would appreciate receiving your revised manuscript by Jul 03 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols Please include the following items when submitting your revised manuscript:
Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. We look forward to receiving your revised manuscript. Kind regards, Sreeram V. Ramagopalan Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. We note that you have indicated that data from this study are available upon request. PLOS only allows data to be available upon request if there are legal or ethical restrictions on sharing data publicly. For information on unacceptable data access restrictions, please see http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. In your revised cover letter, please address the following prompts: a) If there are ethical or legal restrictions on sharing a de-identified data set, please explain them in detail (e.g., data contain potentially identifying or sensitive patient information) and who has imposed them (e.g., an ethics committee). Please also provide contact information for a data access committee, ethics committee, or other institutional body to which data requests may be sent. b) If there are no restrictions, please upload the minimal anonymized data set necessary to replicate your study findings as either Supporting Information files or to a stable, public repository and provide us with the relevant URLs, DOIs, or accession numbers. Please see http://www.bmj.com/content/340/bmj.c181.long for guidelines on how to de-identify and prepare clinical data for publication. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. We will update your Data Availability statement on your behalf to reflect the information you provide. 3. Please include a separate caption for each figure in your manuscript. Additional Editor Comments (if provided): [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Dear Editor, Thank you very much for giving me the opportunity to review the manuscript: PONE-D-20-10079. In this feasibility study the authors investigated whether de-identified electronic health record (EHR) can be used effectively as a tool to identify large samples of users of mental health supported housing using structured fields and free text searches. The authors concluded that it is feasible and resource efficient to use the Clinical Record Interactive Search (CRIS) tool to identify individuals who have used mental health supported accommodation services. The manuscript is well structured and is relevant in the digital world where we have data from patients with huge potential for research. The study is a good first step in utilizing EHR data in the field of mental health support services. The study sample is very large, and the authors discussed multiple ways to identify individuals with mental health supported accommodation services. However, there is potential to improve the description of methodology of the study and to add more analysis to increase the value of the study. I see main problem with the methodology of the study that needs further clarification and explanation to convey the results of the study clearly. First, this feasibility study explains different approaches to identify users of mental health supported accommodation services. But estimated true positive value of only one approach (i.e. free text search approach) is presented. It would add value to the manuscript if true positive values of the structured field search approaches are also estimated and presented. In line 265 the authors mentioned that it is unlikely that a clinician would add false information on supported accommodation service in structured field; however, there has been multiple validation studies showing less than perfect positive predictive value (PPV) of clinical diagnosis in EHR. Therefore, I think it is likely that the PPV of the CPA structured approach is less than 100%. Second, the study is missing the ‘validity’ of the identification methods used by the authors. Ideally, manual chart review of ‘random’ samples from the identified individuals should be performed to get the positive predictive value (PPV) e.g. PPV of combining the CPA structured field approach and free-text approach; PPV of combining CPA structured approach and structured demographic field approach; and PPV of combining all three approaches together. The closest estimate provided is the true positive rate of free text search, which was performed for the first 10% of individuals after sorting the results by note date (line 166) (i.e. not random). Having the information about the validity of the individual and combined search approach will certainly add value to this study. Finally, the technical details in the Method section should be expanded to ensure that readers understand exactly how the authors identified individuals; and missingness and selection bias need to be further discussed. Please see my detailed comments for each section below: Abstract: • The abstract is well-written. • Line 28: the study is not based on data over the last 10 years. • It would be informative if the authors can add something about the ‘setting’ of the study or add name of the mental health trust in the Method section. • Result section line 34: “A manual review of these notes…” Please add, “…manual review of 10% of the notes…”. • Result section, line 35-36: is there any reason of using the term ‘true positive rate’? I think more widely used term is ‘Positive Predictive Value’. • Result section, line 39: The statement that these 337 individuals are likely to be false positive assumes that individuals identified by structured field search are all true positive. This is a strong assumption. Please see my comment on the Discussion section below. • Conclusions: The term ‘efficient’ is very subjective and I suggest using it carefully. In this study authors have fixed the resources before the study. Hence, I cannot see the conclusion of efficiency is based on evidence generated by this study. Please see my comment on the Discussion section below. • Conclusions are based only on results of structured fields, why free text is not mentioned? Background: • This section provides background and good overview of the key literature. However, the section is missing the background on the need of the problem addressed. It appears that the problem is ‘identification’ of people in EHR who have used mental health supported accommodation services. What methodologies have been used in the past to identify such people in the EHR in the same field or other closely related field using CRIS platform? And what were the challenges? Methods: Setting: • Line 127-128: Are there any studies that have investigated the completion of EHR data over the years since it started in 2008? I expect the completion of data to improve over time. If there are differences, then it is a good idea to stratify the results according to years. • I assume there must be changes to the EHR system or healthcare system in the 10 years of the study. Was there any reason of including all available EHR data since 2008 and not restricting the study to recent few years only? I assume the reason was to increase the sample size, but since it is a feasibility study a smaller sample would be acceptable. Search approach: • Line 135-136: “….sample in terms of their sociodemographics using structured fields, and…” I think Table 2 also has information from free text search. • Line 136-137: What was the reason for deciding to compare the sociodemographic data to the national survey from 2014? The study is based on only 2 of the 326 local authority areas, spread from 2008-20017, and is known to be different (as the authors mentioned in the first paragraph of the Setting). • Line 140-142: I could not find any details on how the authors will assess resource effectiveness? How it was measured and what was measured? We cannot assess any effectiveness by fixing the time (=resources). E.g. if we provide only 8 hours to work on something then we will get results, but quality will be compromised. Therefore, it depends on what quality was desired, which is not explained. Hence, we cannot make the resource assessment. However, I do agree that database studies are in general less resource demanding than a prospective real-world study or a clinical trial, but this study does not provide evidence supporting that. Free text search of de-identified clinical notes: • Line 166: “Results were ordered by identification number and note date.” What was the reason for sorting the results by date and the identification number? Ideally manual chart review should be on a random sample. • Line 169: I assume true positive rate is same as positive predictive value (PPV). If this assumption is correct then the definition of true positive rate is incorrect, it should be the ratio of true positive to total positive. Although the calculation is correct in the result section, but the definition is incorrect. • Line 186-187: Do the authors have any reference to support this methodology of balancing sensitivity and specificity? Any previous study that has used similar approach and calculated sensitivity and specificity providing evidence that this approach truly balances sensitivity & specificity? Results Structured fields search: • What was the total base population? I assume it was 126,769 (line 128). • Line 197-199: It is not clear how many individuals have no records of mental health accommodation services. I assume out of total individuals (i.e. 126,769), 1635 had records of mental health supported accommodation services, 9545 had missing/unknown values, and the rest did not get any mental health supported accommodation services. Is that correct? Please clarify. Also, it is not clear how many individuals in total had CPA field records. • Missingness could reflect true absence of the of use of mental health supported accommodation. What was the assumption made for missing/unknown subjects? were they assumed to have no mental health support accommodation, or the data was ‘missing’? May be a flowchart would help. • Table 1: The authors should add footnote explaining that the true positive value was derived from manual review of 10% of the identified individuals. • Line 218-220: Please mention this exclusion criteria in the Method section, it is not currently explained there. Please also add how many individuals were excluded with this exclusion criteria in the final search. • Line 218-220: “Therefore, a condition was added to the search whereby individuals were removed from the results if they only had a single note matching the search term.” why these individuals were not added again after iterating the search term? Comparing the structured fields and free text search approach: • Table 2: It does not illustrate only the sociodemographic but also clinical characteristics. Discussion: • Line 255-261: These can be moved to result section or deleted. • Line 260-261: “….it is likely that many of these 337 are false positives.” This is a big assumption considering we have ‘estimated’ true positive rates. With similar assumptions 45.2% (Figure 2) individuals identified in CPA structured field would ‘likely’ be false positive as they appeared only in CPA structured search. Because the authors did not identify exactly who are false positive in text search field and the authors did not estimate true positive value for CPA structured field search, it will be difficult to conclude anything. • Line 285, “However, the free text search did not appear to significantly enhance sensitivity.” The authors did not estimate ‘sensitivity’ in this study, so there is no data to support this statement. • Line 299: “However, an unforeseen issue arose that inevitably reduced the number of free text results”. I think it would be worthwhile to mention what was the issue and how it had impacted the results. If it was fixed, then this can be deleted. • May be the authors can put light on impact of ever-changing technology during the study. I assume there has been changes to EHR and I assume quality of data (e.g. completion) vary as familiarity to the system increase. • Some of the single service searches have reached PPV of 100% and others remained as low as 35%. Is it possible that the data quality varies between different housing services? Is it possible that data quality varies between providers as well? e.g. some clinicians would not ask or report on accommodation status? • I assume the structured variable that are used in this study to identify the cases are not systematically reported. Therefore, missingness is not at random leading to selection bias. Maybe it is worthwhile to discuss selection bias in this study. Conclusions: • I think the authors’ claim of ‘resource efficiency’ is not substantially supported with evidence. Other comments: • References are not as per the journal’s acceptable style. Many are missing the volume and page number or DOI. • The authors have used some terms inconsistently e.g. effectiveness and efficiency are interchangeable. I suggest using terms consistently to make the manuscript easy to read. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
Using de-identified electronic health records to research mental health supported housing services: a feasibility study PONE-D-20-10079R1 Dear Dr. Dalton-Locke, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Sreeram V. Ramagopalan Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: |
| Formally Accepted |
|
PONE-D-20-10079R1 Using de-identified electronic health records to research mental health supported housing services: a feasibility study Dear Dr. Dalton-Locke: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Sreeram V. Ramagopalan Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .