Peer Review History
| Original SubmissionApril 16, 2020 |
|---|
|
PONE-D-20-10986 Framing COVID-19 How we conceptualize and discuss the pandemic on Twitter PLOS ONE Dear Mr. Wicke, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. I have benefitted from reviews by five colleagues who are all leading experts in this area. As such, you will be delighted to see that the reviewers have provided incredibly constructive and clear feedback, not only on technical issues relating to methodology, but also on the interpretation of the findings in the grand scheme of things. I invite you to engage thoroughly with every one of the reviewers’ points, as I am sure this will result in a stronger contribution. We would appreciate receiving your revised manuscript by Jul 02 2020 11:59PM. When you are ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. To enhance the reproducibility of your results, we recommend that if applicable you deposit your laboratory protocols in protocols.io, where a protocol can be assigned its own identifier (DOI) such that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols Please include the following items when submitting your revised manuscript:
Please note while forming your response, if your article is accepted, you may have the opportunity to make the peer review history publicly available. The record will include editor decision letters (with reviews) and your responses to reviewer comments. If eligible, we will contact you to opt in or out. We look forward to receiving your revised manuscript. Kind regards, Panos Athanasopoulos, Ph.D Academic Editor PLOS ONE Journal requirements: When submitting your revision, we need you to address these additional requirements: 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at http://www.plosone.org/attachments/PLOSOne_formatting_sample_main_body.pdf and http://www.plosone.org/attachments/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. In your Methods section, please include additional information about your dataset and ensure that you have included a statement specifying whether the collection method complied with the terms and conditions for the websites from which you have collected data. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes Reviewer #3: Partly Reviewer #4: Partly ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: I Don't Know Reviewer #4: No ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes Reviewer #4: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: No Reviewer #3: Yes Reviewer #4: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The paper analyzes Tweets that are used to talk about Covid-19 in three studies. The first identifies topics of these tweets. The second quantifies the prevalence of WAR metaphors in general, and across the topics identified in the first study. The third explores alternative figurative frames for the virus, along with one non-figurative frame (FAMILY). It is a timely study with interesting methods and results. I enjoyed reading it. I have suggestions for a revision, along with some methodological questions detailed below. Introduction 1. The introduction could be better connected with the methods used in and research questions explored in the studies. Why is it interesting and important to determine what topics are discussed on Twitter in relation to Covid-19 or how frequently WAR and other metaphors are used? For example, could the topic modeling help to improve surveillance and forecasting about the disease? What frequency of WAR metaphors should we expect? How and why were the specific alternative frames chosen? Are there relevant theories in cognitive linguistics that the current work informs? 2. The emphasis and major strength of the paper seems to be the focus on WAR metaphors. The prevalence of WAR metaphors is quantified for an important issue, by topic, and in the context of other metaphoric frames (and a non-metaphoric frame). To highlight these strengths, I would encourage the authors to emphasize the novelty of quantifying the WAR frame using automated methods and real world data. Maybe switching the order of studies 1 and 2 would be helpful? (The results of Study 1 are difficult to interpret on their own). 3. The “Theoretical Background” section describes past work that is certainly interesting and relevant, but it doesn’t really identify theoretically motivated questions that the current work is well-suited to address. In most cases, it emphasizes practical applications and real world issues and/or methods (e.g., the relationship between measuring public sentiment on Twitter and addressing public health issues). Maybe rename this section for clarity. Smaller points related to the introduction: 4. On p. 3 the authors note that around “16K Tweets are posted by Twitter users every hour, containing a hashtag such as #coronavirus, #Covid-19 or #COVID.” Is this based on the current data collection or a metric that is computed by Twitter? What is the temporal window for this statistic? 5. It seems like there is a typo on p. 4: “Unlike the articles on magazines and journals typically used for corpus analyses of this kind, Twitter does contain messages written by journalists and other experts in mass communication, as most tweets are provided by non-expert communicators.” Maybe “as” should be “although”? Methods 1. Please clarify the following questions: a. How were the 25k tweets per day selected? It seems like the algorithm picked up the first 25k tweets that included one of the specified hashtags per day, but that’s not explicitly stated. How long did it typically take to get to 25k tweets? b. Were retweets included? c. For a given person who tweeted multiple times in a day about the virus, was it only their first tweet that was included? What was the filtering criteria/method? d. Were Tweets in languages other than English included? 2. It should be possible to clarify some of the hedging in the sentence, “our corpus arguably encompasses mainly tweets produced by users residing in the USA, where the time of data collection corresponds to awake hours, and the targeted language corresponds to the first language of many (if not most) US residents” (p. 12). I realize that the location data was stripped from the text before it was stored, but the text is initially tagged with the location data, so maybe it is possible to estimate the percentage of tweets from the US. Language use questions are asked on the US census every 10 years and roughly 80% of US residents report speaking English at home. A citation would make the case stronger. 3. How were the topic numbers (n = 4 vs. 16) decided in the first study? I find the results of this study hard to interpret. I think they are more informative when presented in concert with the results of study 2. 4. Categorizing language as metaphorical or not is tricky. There is a fair amount of work on this and the most common approach uses expert coders (see, e.g., Steen et al, 2010, A Method for Linguistic Metaphor Identification). I think the approach taken in the paper is reasonable but I think some limitations should be acknowledged. For example, there would probably be some disagreement by coders over which conceptual metaphors individual instances of “fight” appeal to (war vs. boxing vs. games, etc) and even whether or not particular instances are metaphorical or not. 5. Include some discussion about the relationship between the STORM and TSUNAMI categories. At first glance it seems like all instances of TSUNAMI should also be instances of STORM, but the current approach establishes these categories as mostly (completely?) non-overlapping. 6. I like the comparison of the metaphorical frames to the FAMILY non-metaphorical frame, but the FAMILY frame also seems fairly different in that it is more of a topic than a frame (i.e. in some ways more akin to the topics identified in Study 1). I don’t think this needs to be changed, but it seems worthy of a little discussion. 7. Include a note on how comparisons will be made. No inferential statistics are presented, which is common in cognitive linguistics, but the question about whether the number of cases is meaningfully different by frame or topic type will likely arise for readers. Results 1. It’s hard to interpret the results of Study 1. Are these topics the ones we would expect? Do they inform theory? Do they inform practice? Could they have come out any other way? 2. What inferences can we draw about the relationship between WAR metaphors and topics from the results of Study 2? 3. If space is an issue, I think the comparison of the 30 vs 50 terms approaches on pp 26-28 could be cut. 4. Small point: There is no General Discussion section, although it is alluded to on p. 24. 5. Small point: Include a citation (and, ideally, a more precise statistic) in the sentence “…as previous literature shows that metaphor-related words cover only a percentage of the discourse, and that literal language is still prevalent” (p. 28). What percentage? Are there other metaphors that might be prevalent? 6. It would be nice to ground some of the qualitative observations noted in the discussion. For example: “Words in the STORM and TSUNAMI frames seem to relate to events and actions associated with the arrival and spreading of the pandemic…” (p. 28). 7. The paper ends by introducing the idea of a Metaphor Menu, which is interesting but it doesn’t logically fall out of the current study in my opinion. Maybe this idea could be discussed a little more. Reviewer #2: The authors examined the (metaphorical) content of tens of thousands of English tweets surrounding the Covid-19 pandemic, scraped from two recent weeks from largely American twitter users. Topic modeling revealed several common themes (4 and also 16; more on this below), and that war metaphors were somewhat common (~4% prevalence), for some topics more than others, and appeared more frequently than other metaphorical domains. They argue that this is consistent with other empirical and theoretical research on the use of war metaphors in public discourse, but they now provide evidence this extends to everyday lay discourse online. In general, I thought this was a timely article dealing with an important topic of interest to a variety of scholars, and a nice extension of previous work and theoretical musings on the use and prevalence of war metaphors. I think the methods and analyses were thoughtful and for the most part sound (though, as I detail below, I was confused about some of the details), and the results were solid. That said, I have a variety of comments and concerns that I think the authors should address in a revision before the manuscript is considered for publication. One overarching concern is that the paper feels like it was rushed to submission and therefore the writing and overall organization are not quite up to the standards of a publishable manuscript. I understand the authors’ sense of urgency in getting this paper out there while the global pandemic crisis is still at its peak, and that they literally wrote it over the past few weeks, but I think extra care needs to be taken during any revision process to make sure the writing is improved. There were many grammatical and punctuation errors throughout the paper, along with confusing sentences and shifts in tense (if they want to refer to “now,” they should stick with phrasing like “at the time of writing,” which they were not consistent with throughout the paper). At times it was difficult to follow the logic of their thinking or make sense of some of the details of the methods and analyses. One of the issues is mostly organizational: the authors chose to frame their work as three “studies,” presenting the “methods” for each first, followed by the results for each, etc. I found this structure to be confusing and hard to follow, as I had to jump back and forth between methods and results and discussion sections to remember what was done (and why) as I proceeded. At one point they discuss the topics modeling results, for example, but it had been so long since they had discussed the methods, and then they waited until the subsequent section to actually give the topics meaningful labels (which they never do in the main text for the 16 topics). It was very hard to keep track of everything because of this structure. As the research itself really strikes me as one single study, not three, but with many analytic components, I think the authors should restructure the paper in a more logical, linear fashion. For example, they could still preview the whole set of big questions and their approach in the introduction, and then the main sections could be each question in turn, with meaningful headings/subheadings rather than traditional “Study 1” and “methods” headings. So, they could still start by describing the procedures for gathering the data from twitter and the organization of the dataset. A sub-heading in that section could be something like “Themes in the data: Topic modeling” where they go through all of the methods, results, AND discussion and labels (each with their own subheadings…) for the topic modeling. Then they can move on to a section about defining their WAR (AND alternative!) dictionaries and analyses and discussion, and then conclude with their general discussion. I think something like this would help make the flow of the paper clearer and more effective. Some additional comments: While the authors reviewed a good amount of research on war metaphors, they neglected to discuss any of the dozens of articles have been written very recently about the war metaphor framing for Covid-19 (and its plusses and minuses), in both mainstream and independent outlets online. I think citing and discussing at least some of these would help situate the article in the present moment, provide additional context, and highlight the importance of the present research. Here are some examples: https://www.vox.com/culture/2020/4/15/21193679/coronavirus-pandemic-war-metaphor-ecology-microbiome https://time.com/5821430/history-war-language/ https://www.theguardian.com/commentisfree/2020/mar/21/donald-trump-boris-johnson-coronavirus https://medium.com/@steve.howe_63053/were-at-war-the-language-of-covid-19-e3d4f4a1ae2e https://www.counterpunch.org/2020/04/24/trump-is-not-a-wartime-president-and-covid-19-is-not-a-war/ https://www.afsc.org/blogs/news-and-commentary/how-to-talk-about-covid-19-pandemic https://theconversation.com/war-metaphors-used-for-covid-19-are-compelling-but-also-dangerous-135406 On Page 8, 167, the authors say “As explained in [1], war metaphors are pervasive in public discourse and span a wide range of topics because they provide a very effective structural framework for communicating and thinking about abstract and complex topics, notably BECAUSE of the emotional valence that these metaphors can convey” (emphasis added). This makes it sound like the emotional valence of WAR is part of its structural framework, but I think this is a bit confused. War provides both a structural schema as a source domain AND it conveys an emotional tone; these points are actually separated in the paper referenced in the sentence. The authors break this down on the following page, but this sentence was unclear. Again, this may be part of the broader need to edit and revise some of the language in the article. P12, Line 243-4: “…and the targeted language [English] corresponds to the first language of many (if not most) US residents.” Look this up and cite a source instead of speculating. P11-12. I was terribly confused by the whole data gathering and filtering procedure. It was unclear how many tweets there were vs. individual tweeters vs. used tweets. The table tracks cumulative tweets but didn’t say that, which was confusing. It was not explained how the filtering was done (i.e., how did you choose which tweet to keep from each user that posted multiple tweets? Did the same tweeters post on multiple days and how was that dealt with?). I think this whole section could be streamlined and made much clearer. Lines 270-72: The authors note they expected to find broader and more generic topics when they included 4 as compared to 16 topics. Well, of course, how else could that have turned out? In general, I found the use of two sets of topics to be unnecessarily confusing and did not feel it added much to the overall message in the paper. I suggest the authors stick with one set of topics that have easily identifiable and meaningful labels/clusters of attributes. Perhaps they could split the difference and choose 8 or 10 topics. Whatever makes the most sense for interpreting the metaphor data later is fine. I should also note this was all very exploratory/arbitrary, which is OK, but perhaps should be noted in the text (they could add a footnote explaining that using different numbers of topics doesn’t fundamentally change the pattern of findings). Lines 309-310: The authors write, “The term list includes the following 79 terms“… but no list was forthcoming yet until the authors discussed their other method for generating terms. Either separate out into two lists (79 + 12) or, better, use one list but BOLD the ones coming from tool two (metaNet), and do not say “the following terms…” until you are actually planning to list the terms. The authors use FAMILY as their “literal” comparison, but it should be noted that family terms COULD be figurative (and indeed, Lakoff, for example, has written much about the figurative uses of FAMILY in describing governments…). For example, “all Americans are one family.” “the president is the father of the American household,” etc. Is there any way to check to make sure all of the instances of family terms in the dataset are indeed literal and not figurative (and to remove the latter)? On lines 503-4, the authors note that war words have a “very negative valence, OF COURSE” [emphasis added]. But I am not so sure I agree with that. Some people might get excited and motivated by ‘FIGHTING” the virus (which feels much less negatively valenced than THREAT, for example). Especially in the United States, which comprises many subcultures that glorify guns and wars and the military, I think some of these terms may be quite positively valenced. Maybe draw on some empirical work and use actual ratings of emotional valence of these words (e.g., using Pennebaker’s LIWC or some other database) Reviewer #3: This paper adopts a topic modelling approach to study a dataset consisting of just over 200,000 tweets about Covid-19 posted in English (and primarily from the USA) in March and April 2020. The approach is employed to: identify the main topics in the data (set at 4 and 16); study the prevalence of a WAR metaphorical framing; compare that framing with three alternative metaphorical framings and a literal topic; and investigate any correlations between the WAR framing and the topics that were automatically identified. The findings are relevant, if somewhat predictable: the WAR framing is more prevalent than the alternative metaphorical framings, and it tends to correlate with discussions of diagnosis and treatment. Concerning the creation of the dataset, the authors provide some justification for limiting tweets from the same account to one. However, this makes it impossible to capture the actual prevalence of the various framings on Twitter. The consequences of this decision should therefore be explicitly acknowledged. The labelling of the groups of terms associated with each automatically generated topic imposes more coherence on each set of words than is actually the case, especially in the version of the analysis that only involves four topics. This is typical of this kind of computational approach to discourse analysis, but it should minimally be pointed out as a methodological issue. As for the alternative metaphorical frames, the terms under TSUNAMI are generally to do with natural disasters, rather than tsunamis specifically. Finally, it should be acknowledged more explicitly that this kind of analysis cannot shed light on how the WAR framing, or any other framing, are actually used. For example, it cannot distinguish between cases where the WAR framing is adopted and where it is critiqued (as has also been the case on Twitter). Ideally, the subset of tweets that employ WAR-related vocabulary could have been subjected to a more fine-grained analysis, but this usually goes beyond the scope of studies such as this. Reviewer #4: Thanks for the opportunity to review. Interesting look at how discussion on Twitter may be framed using frames from the disease literature, and a brief discussion of results of topics models on a limited Twitter dataset. This is certainly a timely thing, so I recommend major revisions. With work I think this could bring value to the public health community as we endeavor to perform contact-tracing and subject to mis- and dis-information around this pandemic. --------------------------- Main critiques --------------------------- What I am missing is the theoretical and practical contribution. Specifically, how would the authors answer the "so what" if the tweets are framed like WAR, STORM, etc., and "so what" if they're not? (which, they're not - 90-95% of the posts are not according to the results.) - Are relative frequencies of frames statistically different from each other, and do they happen often enough to be significant in general? Put another way, does this frame analysis work or matter on Twitter? - How would the authors characterize the other 90% of the discussion, and why / how is it important? Are there any themes related to mis- or dis-information, or to political polarization? Second, I have concerns about sampling bias. This amounts to a study of 12 days' worth of tweets, only a few thousand. Line 60 states 16k tweets are posted every hour (do the authors have a citation?), and yet the authors collected 25k tweets per day. This equation does not balance, even when accounting for a 1% sampling rate from the Twitter API. This uncertainty undermines the efficacy of this paper - either the collection has a problem or the statement is false. Regardless, at the time of data collection multiple datasets of Twitter related to COVID-19 existed. I strongly recommend repeating this analysis in two ways to see if the results change or hold: - one, now that the authors have been collecting more data for a while, - and two, perhaps more pressingly, using one of the public open datasets for Twitter with millions of tweets. See e.g. this collection of resources: http://www.socialmediaforpublichealth.org/covid-19/resources/ "Twitter Data" (This also suggests an opportunity to do temporal analysis, to see if the frames and discussion have changed and if so, how they are changing. This may help with a practical contribution - to answer if discussions are moving in a healthy or helpful direction, or the opposite, and why? - For example, how often do these topics found happen over time, how often do these frames happen over time, and why is that important? How would we interpret these topics, and why might they be important? How do these frames correspond with hashtags or the literal discussion of the disease?) Thirdly, please see critiques of the methods, related to LDA and Twitter pre-processing. Fourthly, I also include more minor points and notes about statistical significance. --------------------------- I also have concerning methods critiques that may undermine results: --------------------------- On tweet processing decisions: - I'm struggling to understand why the authors eliminated all but one tweet per user. This is a limitation. It looks like the methods and results are at the level of a tweet, not at the level of a user. In addition to the sampling bias, the authors could be discarding data that is important to their analysis. If the authors insist on retaining only one tweet per user, how this was performed? Was this random? If not, this could bias one's data again. - I'm struggling to understand why the authors excluded retweets and mentions. How many retweets and mentions are there? Together, these choices severely limit the amount of analysis possible, to show how often the frame of the discourse is spreading, occurring, or changing. I understand wanting to exclude them initially, but what about repeating the analysis with them included to see how it changes? On LDA implementation: - Did the authors use Gibbs sampling or variational inference? Gibbs sampling has been shown to yield vastly superior topics. I'd recommend repeating analysis if used variational inference and see if results hold. Related: - how did the authors choose 4 vs 16 topics? why not other numbers? did they check perplexity - what number of topics has the lowest perplexity? (most likely to explain the data) - How did the authors handle hashtags and URLs and usernames? These may contain information, or not, depending on the design of the study. What happened? These may be useful to report if analyzing the discussion. On LDA interpretation and results: - the authors look at significant words in topics, but what about tweets most about those topics? it can make a difference, per the coming citation. I recommend evaluating topics in both ways, as it may affect results of lines 585-596. See https://scholar.google.com/scholar?hl=en&as_sdt=0%2C21&q=reading+tea+leaves+humans+interpret+topic&btnG= - did the authors interpret all of the 16 topics like they did the 4 topics? (lines 450-453) - for topic figures, I suggest putting names of topics in the figure axes where possible. without them it's inconvenient to remember which is which On Twitter pre-processing, I'm worried about the authors' use of general-language tools on tweets which have been shown to use vastly different language structure. - stopwords from 2012... check/justify that these are up-to-date and apply to Twitter? need to come up with domain-specific ones? - along the same lines there's a twitter tokenizer (e.g., stanfordNLP, NLTK) that are custom-built for this... what about emoji, how were these handled? - line 279: better would have been to use tf-idf and leave the common terms in... these would have been reduced by the weighting organically On literal framing control: - What about the literal frame "it's a disease"? The authors chose family as a literal frame- this may strongly coincide with incidence or deaths from the disease, which may not be exactly what the authors want to measure. - In addition... how are the authors controlling by including this? Should this be used for normalization, or testing statistical significance of frequencies or of differences among frame? --------------------------- On Results and discussion --------------------------- table 2 - are these results statistically significant? This would give weight to the authors' statement about the relative amount. lines 512-536, about topics predicting occurrences of frames... are these differences between frequencies statistically significant? are these frequencies high enough to matter? Continuing down the path about frames vs. topics: - How often do the family or alternative frames show up in the predicted topics? like likes 383-384 for the WAR frame. - In addition, how many topics include words in the frames? This may be an indicator if the frames are even worth studying on this domain. (see 90% number and earlier comment about frequencies) Lines 55-67 do the authors have any citations for any/all of these statements? ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: Yes: Paul Thibodeau Reviewer #2: No Reviewer #3: No Reviewer #4: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files to be viewed.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email us at figures@plos.org. Please note that Supporting Information files do not need this step.
|
| Revision 1 |
|
PONE-D-20-10986R1 Framing COVID-19 How we conceptualize and discuss the pandemic on Twitter PLOS ONE Dear Dr. Wicke, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. ============================== I have heard back from 3 of the original reviewers. As you can see, once again the colleagues have made incredibly helpful suggestions for revisions. I invite you to consider each point carefully. Please pay particular attention to the clarity of language issues highlighted by reviewers 1 and 2. I agree with them that you have your manuscript proof-read prior to submission of the next revision. And please especially focus on the technical issues with the data highlighted by reviewer 3. The reviewer has given no less than 4 possible ways forward, any of which I find entirely reasonable. At the very least, you should acknowledge the limitations of your data processing method, and modify your theoretical claims accordingly, as suggested by the 1st option the reviewer puts forward. I look forward to receiving your revision as per the guidelines below. ============================== Please submit your revised manuscript by Sep 14 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, Panos Athanasopoulos, Ph.D Academic Editor PLOS ONE [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: (No Response) Reviewer #2: (No Response) Reviewer #4: (No Response) ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes Reviewer #4: No ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: I Don't Know Reviewer #2: Yes Reviewer #4: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes Reviewer #4: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: No Reviewer #4: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: I appreciate the careful work of the authors to address the concerns raised in the previous round of review. The manuscript is clearer as a result and I think it will make a solid contribution to the literature. I have a few remaining comments and concerns that I describe below. Introduction & Theoretical Background The introductory sections do a better job of describing and situating the contribution of the current work. The paragraph that explicitly details “The innovative aspect of this paper…” (p. 5) is particularly helpful. There is still an issue with the sentence I highlighted in my previous review, which now reads, “Unlike the articles on magazines and journals typically used for corpus analyses of this kind, Twitter contains messages written by journalists and other experts in mass communication: most tweets are provide by non-expert communicators” (p. 4). Maybe: “Although Twitter contains messages written by journalists and other experts in mass communication, most tweets are provided by non-expert communicators…” I still find the heading “Theoretical Background” misleading. While that section certainly does discuss relevant research that helps situate the current work, it doesn’t discuss theoretical background per se (to my eye). But if the authors feel strongly that it does, that’s ok with me. I feel more strongly that the heading “Experimental design” should be changed because the study is not an experiment. Maybe “Study design.” Methods I appreciate the revised structure of the methods and results, which makes the paper more readable. I still find the first section difficult to interpret. I think it would be useful to say explicitly in the set up or discussion of the topic modeling that the categories will be useful for understanding how the war frame is being used. It’s not clear what it means to say that, “The two numbers of clusters were chosen on an empirical basis” (p. 17). What was the empirical method / criteria? How was it evaluated? Adding the LDA coherence measure is useful but seems different from what the authors mean when they say that the “clusters were chosen on an empirical basis.” I appreciate the discussion of the MIP and MIPVU in the response letter, but I think some of these issues should be noted in the manuscript itself. Namely, this method of categorizing and quantifying metaphor is new, different from alternatives, and has its own benefits and limitations (as does any method of coding language). To be clear, I think the method being used is interesting and worthwhile. But it likely misses some instances of metaphor and it likely counts some words as metaphorical that are being used in a literal sense. Maybe, for example, someone tweeted about how COVID was spreading on the aircraft carrier the USS Theodore Roosevelt, which is a war ship. Given that one of the main contributions of the paper is the method, I think some attention should be given to the strengths and limitations of the method. Of note, a major strength of the method is that it can handle a lot of data. This point could also be made more explicitly. [The comment below also relates to a limitation of the coding method.] I agree that the concept of a STORM is different from the concept of a TSUNAMI. However, if the same words (e.g., wave, flood, disaster) are being used to index both concepts, how do we know which concept people are drawing on? For example, the word “disaster” is the second most frequent marker of a TSUNAMI in the corpus and also the second most frequent marker of a STORM in the corpus. The word “wave” is also frequent in both. This means that some of the tweets are double counted as metaphorical in a sense, no? Do the analyses assume that the categories are mutually exclusive and exhaustive? It would be useful to report the percentage of tweets that included multiple categories of metaphor. I understand and appreciate how the cluster of words related to FAMILY are being used, as a point of comparison with the figurative frames (WAR, STORM, TSUNAMI, MONSTER). But I still think there is a qualitative difference between this category and the others. The language of WAR (STORM and TSUNAMI and MONSTER) is being used to frame different aspects of the pandemic and response. In many cases, the issues being framed metaphorically could be discussed using an alternative metaphorical frame or in a literal sense. The tweets about FAMILY, on the other hand, seem to be about FAMILY per se. That strikes me as an important difference that should be discussed in the paper. Results See above for questions about tweets that were categorized as relating to multiple frames. Adding the replication study strengthens the paper. I recommend toning down some of the claims like, “…the percentage of use of the WAR frame reported in our study suggests that this frame is particularly pervasive, because it covers more than one third of all the metaphorical language typically used in texts” (pp. 33-34). The study restricts itself to four metaphorical frames of interest; the full range of metaphor typically used in texts is not quantified here. Reviewer #2: I want to commend the authors on the impressive revisions to this paper. I think they have done just about everything they could to address all of the reviewers concerns, and I feel the paper is much stronger now and will make a nice addition to the literature. Indeed, I could see myself citing the findings in a future talk or article! The only reason I am selecting "minor revision" rather than "accept" is that PLOS ONE does not copyedit accepted manuscripts, and I still feel the paper could use one more round of proofreading and editing. While the overall structure of the paper is greatly improved now, there were still grammatical errors and awkward phrases sprinkled throughout that hindered readability. For example, lines 75-77 contained the sentence "Unlike the articles on magazines and journals typically used for corpus analyses of this kind, Twitter contains messages written by journalists and other experts in mass communications: most tweets are provided by non-expert communicators" This is very confusing to me as it seems to be saying two different things. A more mild example, but still one that should be edited for clarity, comes on lines 179-180: "The military metaphor thanks to which we frame diseases such as cancer is a very common one to be found in public discourse." And in the next sentence, the authors use the article "the" before Time Magazine, which is not necessary. I am not going to go into every example of confusing writing, but I do want to recommend the authors go over the paper again and update the prose as necessary to enhance clarity. At that time, I will be prepared to accept! Reviewer #4: Thank you to the authors for drastically improving many points about the paper. I enjoyed reading this interesting research even more this time. It is much cleaner to read and understand, including contributions, backing statistical analyses, and replication on other datasets. Other reviewers also brought up great points that I did not think of. However, as is written, the study seems to have unacknowledged limitations and concerns that render contributions too broad about understanding general discourse on Twitter that I would recommend addressing before accepting this paper. My concern stems from ambiguity in terms of unit of analysis that biases the representativeness of the data, which causes unacknowledged impact on results and contributions. Claims cannot be made about general discourse on Twitter, but they *can* be made about a biased yet useful subset - those tweets that originate from less-frequent tweeters. Generally, any study that performs social monitoring needs to: - make clear what its unit of analysis is, - then accordingly determine what it means to be representative of these units in its sampling frame, - gather units to analyze and analyze those same units - acknowledge any limitations of representativeness - and limit contributions accordingly At minimum in this particular study, - it should be made clear and consistent through data collection and analyses whether this particular study is analyzing use of frames by users, or use of frames in tweets, (I'm pretty sure it's tweets, right?) - the limitation of a bias against super-tweeters should be made more explicit, and I would accordingly recommend explicitly walking back the study's claims, research questions, and contributions, because claims of general representativeness on Twitter are not defendable. Specifically, if it is chosen to omit the vast majority of tweets by keeping one tweet per user and omitting retweets, claims cannot be made about how the general Twitter discourse discusses and frames COVID-19 (or to put it another way, the study's data is not representative of all tweets). Claims instead can be made about how those tweets *originating from less-frequent tweeters* discuss and frame COVID-19. Either that, or I would recommend at least one of: - switching to analysis at the level of a user - more representatively sample from tweeters (as one tweet per user is usually not seen as representative of a user's content, whether they are among the top tweeters or not) Please find details below - perhaps these can make the issue more clear if it helps. First, the study has a bias against retweeters and sharers, in favor of original tweeters, as is stated. However, many users on Twitter retweet or share someone else's tweets - this is seen as a proxy for behavior and opinion, as usually people share things they agree with. Therefore, more accurate counts of frames might be *including* those frames that are retweeted. At the very least, one may analyze with retweets and without to see any differences. It's fine that the study disregards retweets, but this should be acknowledged as a limitation and claims about discourse should be modified accordingly. And second, to paraphrase, data is being "limited to one tweet per user because of technical limitations and to make the corpus balanced and representative". I would reject the study's claim that the resulting corpus is representative - because the study is not clear about *how* it is representative. As designed, it seems like the study is confused about the unit of analysis that it is trying to represent: the tweet or the user. - If the study wishes to accurately represent tweets, then by Twitter's nature there are super-tweeters, and frequencies of frames will be biased by super-tweeters who use frames more. This is an accurate representation of the Twitter discourse. - If the study wishes to accurately represent users, then it should state this, and tweets should be aggregated or sampled representatively per user, and and all subsequent analyses should be at the level of the user, not the level of the tweet. However, as currently designed, the study starts at the level of a user by "retaining one tweet per user", but then moves back to analyze at the level of a tweet. This results in a bias against super-tweeters with "retaining one tweet per user". Therefore, all subsequent analysis at the level of a tweet omits the vast majority of tweets that compose Twitter discourse - exactly what it seems to want to analyze. - Minor point: The study should address exactly how it "retained only one tweet per user". This was asked in the review. Was the earliest tweet retained? the most recent? the one with the most framing words? Is it simple random sampling? I ask to elucidate, not to be flippant, as ambiguity does not lead to reproducibility - it leads to concern. A brief mention would suffice. I can recommend four courses of action, although the authors would know which would be more appropriate for their desired contributions: 1. Keep current data collection, methods, and results that seem to analyze at the level of a tweet, and acknowledge explicit bias against super-tweeters, to favor those users who tweet less often. This seems to require revisions to contributions and research questions, as the overall Twitter discourse is not being studied for framing, but the Twitter discourse being studied is biased towards discourse by less-frequent users. Is there a theoretical motivation for studying in this way? 2. Keep current data collection, analyze at the level of a user, and acknowledge explicit bias against super-tweeters, to favor those users who tweet less often. This would require more extensive revisions to analyses, but would result in a different contribution: instead of studying the entirety of framing discourse on Twitter, studying how *users*, with emphasis on less-frequent users, frame the discussion. 3. Decreasing bias against super-tweeters by gathering a representative amount of tweets per user, and analyze at the level of a user. This would result in a proper representative analysis of discourse on Twitter at the level of the user, and contributions on how users discuss and frame would follow. 4. Decreasing bias against super-tweeters by gathering a representative amount of tweets per user, and analyze at the level of a tweet. This would result in a proper representative analysis of discourse on Twitter at the level of the tweet, and defendable contributions on how tweets generally discuss and frame would follow. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: Yes: Paul Thibodeau Reviewer #2: No Reviewer #4: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 2 |
|
PONE-D-20-10986R2 Framing COVID-19 How we conceptualize and discuss the pandemic on Twitter PLOS ONE Dear Dr. Wicke, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. ============================== I am happy to accept your paper pending a few minor modifications as outlined by reviewer 1 (see below). Please submit the final revision with a cover letter showing how you have incorporated these final minor revisions as per the instructions below. I will then process the paper for publication without another round of review. Thank you for your diligent effort to address all of the reviewers' helpful comments throughout. I look forward to receiving the final version of your paper in due course. ============================== Please submit your revised manuscript by Oct 23 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, Panos Athanasopoulos, Ph.D Academic Editor PLOS ONE [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The manuscript, which explores metaphors used to frame COVID-19 on Twitter, has been revised to address concerns raised in review. Overall, I think the revision is very successful. The argument is much clearer, and I think the paper is nearly ready for publication. I have a few final comments that should be straightforward to address. I am recommending that the paper be accepted. 1. I still think the use of “empirical basis” is misleading and confusing (p. 17) because, in many research contexts, this term is used to refer to specific criteria or quantitative methods of comparison. Since the current approach is more informal, why not say in the paper what is said in the response letter? Namely, something like, “These two numbers of clusters chosen by looking at the data and are backed up by post hoc testing.” 2. Please include some details about the criteria that Lamsal used to create their database of COVID-19-related Tweets. 3. I was a little confused by the term “individual” in “individual tweeters” on page 13. I think “unique tweeters” would be clearer, assuming that is the intended meaning of “individual tweeters.” 4. Using the Topic labels (e.g., “Communications and Reporting”), rather than the numeric codes (e.g., “Topic III”), would make the figures more readable. Once again, I found this research really interesting and think that it will make a nice contribution to the literature. I appreciate all the careful work the authors have done throughout the review process. Reviewer #2: I am satisfied the authors have addressed all critical reviewer points. The current manuscript is significantly stronger than the initial submission and makes a nice contribution to the literature. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 3 |
|
Framing COVID-19 How we conceptualize and discuss the pandemic on Twitter PONE-D-20-10986R3 Dear Dr. Wicke, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Panos Athanasopoulos, Ph.D Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: |
| Formally Accepted |
|
PONE-D-20-10986R3 Framing COVID-19: How we conceptualize and discuss the pandemic on Twitter Dear Dr. Wicke: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Professor Panos Athanasopoulos Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .