Peer Review History
| Original SubmissionMay 24, 2021 |
|---|
|
PONE-D-21-17113 Representativeness of individual-level data in COVID-19 phone surveys: Findings from Sub-Saharan Africa PLOS ONE Dear Dr. Wollburg, Let me apologize first for taking much more time than expected. All reviewers kept requesting for extensions of deadlines and in the current situation it is hard for editors to not grant these. Even at this point, one review is still outstanding, but I decided we should be able to proceed with the two I received at this point. From the assessments of the reviewers, you will see that both reviewers feel this is important research, but also have some important questions on the methodology used (why deciles?) and suggestions on how to make the paper stronger overall. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Sep 04 2021 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Bjorn Van Campenhout, Ph.D. Academic Editor PLOS ONE Journal requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and Reviewers' comments: Reviewer #1: During the COVID-19 pandemic some of the LSMS-ISA household surveys were used as sampling bases for new phone surveys. This paper assesses the reliability of these phone surveys to provide individual-level statistics. One of the main constraints to doing this is that 74% to 83% of the phone calls were made to household heads, so the question arises how you can use this information to say something about the population as a whole. The public use data already include household-level weights that correct for non-response in the phone survey. This paper shows that those weights are not adequate to get accurate individual level representation and shows that a reweighting procedure using individual weights provides some improvement. My main comments are about (i) framing of the paper and (ii) calculation of the weights and (iii) the tests. (i) Clearly the main problem the authors have to deal with is that phone numbers were mainly collected from household heads and even where they were not, the interviewers ended up speaking to the heads. It is obviously not going to be easy to say something about the general population if you have such a specific sample of household heads. So the problem is really: with four fifths of the sample household heads, how can we say something about individual responses (on, say, knowledge of COVID-19) of everyone else. Presumably this is going to come about by giving lots of weight to those who are not heads and don’t have the typical head characteristics. There is no methodological innovation in this paper as such reweighing is quite standard, but because this is an important and widely used data source there is, in my opinion, merit in knowing whether we can use LSMS-ISA data for individual level analysis. The problem is also quite specific to this survey constellation in LSMS-ISA. If the survey designers had known that a phone survey would follow, they very likely would not have set it up this way. So the question is one of how much reweighing can help in this specific survey. These important points to highlight when framing the paper. (ii) There is something odd about the two-step procedure you use. You estimate the probability of each individual to end up being interviewed in equation 1 and use the inverse (of the decile-average) to reweigh. But instead of applying that weight directly you multiply it with the household weight (in equation 3). I don’t understand the logic of doing this. Why not just use the inverse of the probability that each individual ends up in the sample as sampling weight? That is the tried and tested technique used widely to correct for such problems. If you do want to argue in favor of the two-step procedure then you should take account of your nested procedure when estimating equation 1. (iii) There is something circular about your tests: you use a set X of observed characteristics of individuals to reweigh the observations. Then you see whether after reweighing this same set of characteristics X better reflects the population average. That is setting yourself mechanically up for success. And I think that if you drop the two-step procedure and use the single step procedure you will get a pretty good match. This is something to acknowledge and address in the paper. Four specific questions: • Could you explain why you use deciles of the predicted probabilities rather than the probabilities themselves? • Section 2.6 is not very clearly written; I only understood what you were trying to do after reading the results section, which defeats the purpose of having a section explaining the empirical set-up. I suggest a careful rewrite. • What do you mean that the winsorized weight are post-stratified (p19). Could you explain this part better? • Why is winsorization necessary? You have already taken deciles for you weights so presumably there are no outliers. Could you provide some specific numbers to show what is going on in the extremes? Reviewer #2: This article analyzes whether unweighted phone survey data are subject to selection bias, and whether the use of household- and individual-level weights correct for such bias. The paper uses household rosters obtained from pre-COVID face-to-face household surveys as nationally representative data, and as a benchmark for what the survey population of phone surveys administered during the pandemic should look like. Obviously, as the phone survey protocols targeted the household head, the unweighted phone survey sample does not resemble the nationally representative adult population. Interestingly, the use of household-level weights does not help address this bias, and in many cases, makes the phone survey sample even less representative of the adult population. Although individual-level weights perform better, this method is not sufficient as a bias correction technique, as differences between the nationally representative adult sample and the phone survey sample remain significant for several variables. The analyses performed in this paper are interesting and worthwhile publishing, but I believe that the paper needs to be rewritten. First and foremost, I find the question of whether the phone survey data are representative of the adult population not very novel or interesting and would suggest taking that result for granted or as an obvious outcome of the survey protocols, which were not designed to get a representative sample of individuals. The protocols targeted households, and were mostly blind towards selection of respondents within the household. The real added value of this paper is therefore the analyses of the bias correction methods: when we do our phone surveys with just one member of the household, usually the household head, can we use weighting methods to correct for a potential selection bias? The paper finds that this is not the case, and this has important implications for respondent selection protocols, if one is interested in individual- instead of household-level data. This could be focused on much more throughout the abstract, introduction, and conclusion. For instance, it is not salient from the abstract that the focus is on selection bias from an individual rather than household point of view, and as a result, the last sentence of the abstract comes as a surprise. The introduction could be shorter and more focused on the issue at hand - that most phone surveys are done with just one person in the household, typically the household head, and thereby not representative of the adult population. The main question, then, is whether bias correction methods can make the data more representative, or whether we should adjust the methods through which we select respondents within a household. To make this salient, it would help if the paper introduced in the last paragraph on page 8 (using the same data) is introduced much earlier in the introduction. Second, and related to the first comment, the data from Malawi appear underutilized in better understanding the drivers of the selection bias. In all other countries, the protocol was to first call the household head; but in Malawi, phone numbers were called in random order. Despite that, Malawi still has significant selection bias, and it would be good to know whether this is because household members are passing the phone to the household head, or whether certain types of household members are less likely to pick up their phone, and replaced by the next number on the list. Third, the weighting is an important aspect of the paper, but the discussion of how the weights are constructed is difficult to follow and although I understand that the authors do not want to provide the technical details here, as they are presented in a different paper, the presentation could benefit from more intuition. For instance, why are deciles being created? What is their use (and why deciles as opposed to for instance quintiles or quartiles)? And why is post-stratification applied to the weights? Fourth, I am not convinced by the added value of the employment outcomes. The supposedly representative data are reported by proxies, except for the data that the respondents provide for themselves. If we find differences between the individual-weighted and benchmark data, is that because the weighting does not work, or because the benchmark data are biased? The authors discuss this in the conclusion, but it raises questions around the added value of this comparison; unless the authors could show that findings are the same regardless of whether we study variables on which there should be less asymmetric information and more accurate reporting by proxies, versus variables where we would expect more error. Finally, with the large set of variables that the authors look into, times the 4 countries for which the analyses are replicated, there are a lot of numbers to review in order to draw conclusions. Although I see why the weights are estimated at the country level, in presenting the results on bias correction, the regressions could be presented at the aggregate level, combining all countries in one regression, and the country-specific analyses could be presented in an appendix. Alternatively, would there be scope for an aggregation over the different variables, in order to reduce the number of coefficients that need to be interpreted and aggregated by eyeballing before drawing conclusions? Other comments: - The first part of the introduction appears negligent of a large body of literature on phone surveys and survey design; but in fact this literature is mentioned towards the end of the introduction. In this case, I would consider integrating the existing literature in the first part of the introduction, and presenting the findings along with their implication as the key contribution, to shorten the introduction and appear less negligent of the work that has already been done in this field. - The figures were not visible in the manuscript itself, only as the appendices of the submission. - A bit more explanation on why certain things vary across countries would be useful. For instance, why are the fifth round data for two countries excluded? Did these not include the employment outcomes? But how does that rhyme with all surveys being standardized across the four countries? And why were the selection protocols different in the four countries? - At times, the paper reads a bit like a promotional brochure for the World Bank (for instance the introduction claims that the World Bank is the prime institute doing and learning around phone surveys, followed by 3-4 other institutions, among others). The fact that the anonymized surveys are published online a few weeks after completion could be a footnote in the methods section and does not require an entire paragraph. The paragraph in the conclusion stating how widely these data are used is irrelevant unless it is used to illustrate that people are using the data for individual-level analyses and therefore using the wrong household-level weights. - Bottom of page 19 (last sentence), note that there is a small typo: households instead of household. - IRB: Clarify that the data to which the authors had access were anonymized and that the merging of phone survey and face-to-face data was based on an anonymized household ID. Even if for data collection no ethics approval was needed, for a researcher to work with data that contain personal identifiers, and merge data from different sources, IRB should be obtained. Overall, though, the paper presents an extremely useful analysis that needs to be published, since we should be more aware that our household-level weight adjustments often do more harm than good if using individual-level data, and that survey respondent selection protocols need to be adjusted when the objective is individual-level analysis. While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
Representativeness of individual-level data in COVID-19 phone surveys: Findings from Sub-Saharan Africa PONE-D-21-17113R1 Dear Philip, I heard back from the two reviewers and they both indicated that all their comments and suggestions were satisfactory addressed. Therefore, I have decided to accept the article as is. Congratulations! Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Bjorn Van Campenhout, Ph.D. Academic Editor PLOS ONE |
| Formally Accepted |
|
PONE-D-21-17113R1 Representativeness of individual-level data in COVID-19 phone surveys: Findings from Sub-Saharan Africa Dear Dr. Wollburg: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Bjorn Van Campenhout Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .