Peer Review History
| Original SubmissionMarch 27, 2022 |
|---|
|
PONE-D-22-09013Two is better than one: Using a single emotion lexicon can lead to unreliable conclusionsPLOS ONE Dear Dr. Stillwell, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. We have received the reports from our advisors on your manuscript and read them carefully. We think the reviewers provided very good assessments and recommendations. Regarding the content of the reviews, we find no inconsistencies and we share their comments. Based on the advice received, we feel that your manuscript could be reconsidered for publication should you be prepared to incorporate revisions. You should appreciate that this is a *major* revision. Minor changes to your manuscript will not be acceptable. Furthermore, the current decision does not guarantee an eventual acceptance of your paper - thus, the importance of your revisions. Please submit your revised manuscript by Aug 14 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Andrea Fronzetti Colladon, Ph.D. Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. In your Methods section, please include additional information about your dataset and ensure that you have included a statement specifying whether the collection and analysis method complied with the terms and conditions for the source of the data. 3. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide. 4. PLOS requires an ORCID iD for the corresponding author in Editorial Manager on papers submitted after December 6th, 2016. Please ensure that you have an ORCID iD and that it is validated in Editorial Manager. To do this, go to ‘Update my Information’ (in the upper left-hand corner of the main menu), and click on the Fetch/Validate link next to the ORCID field. This will take you to the ORCID site and allow you to create a new iD or authenticate a pre-existing iD in Editorial Manager. Please see the following video for instructions on linking an ORCID iD to your Editorial Manager account: https://www.youtube.com/watch?v=_xcclfuvtxQ [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: I Don't Know Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: No Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: Review to: Two is better than one: Using a single emotion lexicon can lead to unreliable conclusions. This is an interesting paper challenging the habit of just using a proven lexicon. The paper shows that when using different lexicons that where developed for a similar goal, there might be differences in outcome and they show in a small example on investigating where such differences in outcome come from: in their case: political words. However, I think that the paper is a bit unclear on it’s message and what it’s main goal is. In general the paper is quite vague which makes it difficult to understand. It is hard to point exactly to why, but I will point at two points where I think it needs improvement. The two main points of vagueness: The influence of Political words, Conclusions The influence of Political Words. In itself this is an interesting section, and it shows the use of a large selection of methods. However, it’s function in the paper is not so clear. It is also not mentioned in the goals or introduction or abstract. The best description of what this chapter tries to achieve lies in the last sentence of the article: “but if not, then researchers need to examine the lexicons to find out what difference might be causing the inconclusive result. This section is very dense in information compared to the other parts of the paper. A lot of methods pass by (LDA, LASSO, autoregressive multilevel model), but it is not clear why a certain method was employed and what the results of the method were and that the results are only . This makes this section hard to comprehend. I would suggest to rewrite this section with more focus to what the method tries to achieve and less attention to details. For example, take more time to explain what the LDA does and why you are using it and less on details such as the amount of iterations (readers can find that in the code if this is published.) or the fact that you used Gibbs sampling (as far as I know this is the most often used). Only mention these details if you diverted from the default for a very specific reason (e.g. if Variational Bayes did not work in your occasion) I would also suggest integrating this part better by more explicitly referring to it in the introduction/goals. Conclusions: The authors spend a significant part of the conclusions explaining what the paper is not. It is not a tutorial and it is not a critique on the lexicons. I would suggest to rephrase the conclusion such that it tells more about what the authors think the paper achieves instead of what they think that it doesn’t. Smaller points/suggestions - Although it is not necessary, It might be interesting to include a small discussion on the difference between Emotion analysis and Sentiment analysis. - Figure 1: I would suggest to replace the titles Panel A, Panel B, Panel C with more descriptive labels: e.g. LIWC, NRC and NRC no political words. - Figure 1: Produce the image in a better resolution. Currently it is not possible to read the axis ticks, even if you zoom in. Alternatively increase the font size of the ticks (and labels). - line 219: how many tweets were deleted in each step? - Line 241: Include a histogram to depict the distribution of ages. Conclusion Substantively I see the value of the paper. It compares and evaluates two lexicons that are often used without questioning. But I think the paper can be written more clearly such that it is clear what the authors did and why, as well as to properly evaluate the quality of the work. For this reason I think a revise and resubmit would be the most appropriate way forward. Reviewer #2: This is an interesting study that compares the use of two separate widely used emotion lexicons in a particular research project. It describes both how some conclusions are in alignment as well as how some conclusions are different when using the two lexicons for the project. Further, the paper examines, why some of the conclusions are different. This is a useful paper to highlight the nuances of emotion lexicon use in a research project. There are three main points of improvement that I would suggest: 1. Some of the claims made towards the end of the paper about why the differences exist are not supported by evidence. For example, it is claimed thatwords like democracy, empathy, and equality are positive but that they are in reality used in negative tweets. Firstly, this needs to be supported with more evidence, than a mere citation to a paper. Secondly, if these words are primarily used in negative tweets then all that negativiness from the rest of those tweets will overshadow the positiveness in the overall affect curve. So in the end it just seems like, what is done here is to remove a large number of positive words that older people use, and that has reduced the positivity score when using NRC to bring the curve closer to LIWC. Another claim without evidence: "several terms without clear emotional meaning are related to politics, e.g., for the positive NRC wordlist: “president”,“police”, “cabinet”; for the negative NRC wordlist: “government”, “foreign”, “liberal”"There is no clear fixed boundary between coarse categories like neutral and positive. In the NRC Lexicon, some majority of human annotators have said that "president" is associated with the positive class. For this paper to now claim/imply that the NRC lexicon should have only marked a word as positive if it has "clear emotional meaning" is a strawman argument. State clearly, what the NRC lexicon captures. Then if one wants to say that for an application that should only make use of terms that denotate an emotion, they shoudl not use the NRC lexicon, that is fair. 2. Lack of clear description of how the two lexicons were created. There is no description of how the NRc lexicon captures word--emotion associations (connotations); and how such lexicons are different from denotative lexicons. There is vague hand-waving that the LIWC is "theory-drive" and created by experts, but pretty much no description of the criteron for inclusion of a word in a category. No descriptions of whether these words are meant to denotate an emotion or can include words that are simply strongly associated with the emotion. The nature of a lexicon greatly impacts what conclusions one can draw from experiments that use the lexicon. This is of much greater significance than portrayed in the paper. The difference in results is not just a matter of coverage in the two lexicons. 3. Attributing gender by appearance can lead to misgendering. Even though te work here involves aggregate-level analysis, the paper should not perpetuate hetero-normative ideas of assuming gender by appearance or name. 4. There is no discussion of ethical considerations involved when suing emotion lexicons to draw inferences about people. See:Practical and Ethical Considerations in the Effective use of Emotion and Sentiment Lexicons. Detailed comments: "Thus, in this study, 74 we examine what would be the conclusions of a study testing a single research question using two 75 different and widely used emotion lexicons. Would the two methods give the same answer?" How generic is this result based on one study? The abstract and intro talk about: theory driven lexicons and data-driven/crowdsourced lexicons at some points and, theory driven lexicons and machine learned lexicons at other places. The machine learned lexicons come a bit out of the blue and need to be introduced appropriately. How exactly does one create a "Theory-diven" lexicon. "Chosen by experts based on theory" seems somewhat vague. In the context of small-LIWC-like lexicons and large NRC-Emotion-like lexicons, it is useful to talk about denotative and connotative lexicons. See Ref.Does LIWC mostly include words that denote emotion?NRC Emotion Lexicon includes word--emotion associations (connotations).Connotative expressions subsume denotative expressions and are much more common than denotative expressions. This seems like a key reason behind the success of connotative lexicons to analyze emotions in text. "against a criterion such as emotion ratings (4–7)"emotion ratings of what? "The 2015 version of LIWC has 6,400 terms within 90 pre-defined categories," cite corresponding paper. Point to where one can access the lexicon. Is the lexicon free for use? "There is also a newer version of LIWC which has been released in February 2022."How is the 2022 version different from the 2015 version? "psychological scales such as the PANAS (23)"How were words included in psychological scales such as the PANAS? Giving some idea of which words were selected and how representative they are of emotion words in general will be helpful. How is "“goodness of fit” to a particular category defined?What is the goodness of fit chosen to be? e.g., the word should denote that category? Be strongly associated with that category? something else? "the NRC emotional wordlist is about 4 times longer than LIWC’s" Is this including all LIWC words like past tense words or only words pertaining to emotions? "The NRC Emotion Lexicon provides a list of 6,388 unique words (13,901 words in total,"Cite the paper for the lexicon. And point to the creators' website that distributes the lexicon:http://saifmohammad.com/WebPages/NRC-Emotion-Lexicon.htm If I recall correctly, the lexicon has 14,000+ unique words marked as associated or not associated with various emotion categories.The R package is a third party product and not a proper reference when describing the lexicon. The R package probably is only using the words that are marked as associated with an emotion category (these might be ~6000). "difference between the LIWC and NRC lexicons"One key difference is also that the NRC lexicon tries to capture how people think or use a word by asking a large number of people (and in that sense has a descriptive nature of language), whereas expert annotations tend to be about how a word should be used. Although, it is unclear how the LIWC annotators made their decisions. "users who publicly disclosed their age"Worth noting that this might lead to a selection bias -- people on Twitter that are likely to disclose their age in a tweet. "gender of the users was noted240 by the research assistants manually inspecting the accounts for clues such as their profile picture or 241 name)."Attributing gender by appearance can lead to misgendering. Even though te work here involves aggregate-level analysis, the paper should not perpetuate hetero-normative ideas of assuming gender by appearance or name. Fig 1: Use the same scale on the Y axis (just as the same scale is used for x axis) for all the sub-figures. Otherwise, somebody who does not look at the y axis labels (which are out of focus and hard to read, BTW) can be easily misled. When using large association lexicons, removing words not appropropriate is recommended:cite paper Key Related work:PoKi: A Large Dataset of Poems by Children "However, excluding all of these words would not be justified as they may accurately convey 404 positive or negative affect."It is not clear that removing "words associated with politics" is appropriate either. "Instead, we aimed at identifying and excluding only the words that do not 405 correlate with affect in our Twitter sample."What does this mean? Thus, we correlated the frequency of each word with the 406 respective LIWC scores: If a word is negatively correlated with the respective LIWC scores, it is likely407 that in our sample it does not convey the emotion suggested by the NRC lexicon." This just seems like you are removing entries from the NRC lexicon that are not aligned with LIWC! This paper is trying to compare NRC Lex and LIWC. In such a case, LIWC cannot be taken as the "right" lexicon. "The majority of these words are rather neutral when out of context."?How do you determine the boundary between neutral and positive? Is this simply a matter of what the authors of this paper feel? is the argument that the authors' boundary is the right one compared to the one determined by the annotators? Does the boundary change based on the application at hand? I would argue these boundaries are artificial, and it is better to use relative sentiment lexicons such as the NRC VAD lexicon or the NRC Emotion Intensity lexicon. Ther the only claim is that one word is more positive than another. The claim is not that a word is "positive" or "neutral". "Importantly, some of the political words in the NRC positive affect wordlist we excluded represent488 values, e.g., “democracy”, equality”, and “empathy”. Although those words have a positive 489 connotation in general, in the context of Twitter they are likely to be associated with criticizing the 490 current political situation or used as part of moralized conflicts between people supporting different 491 political ideologies (e.g., Sterling & Jost, 2018)." what exactly do Sterling & Jost, 2018 show to support the claim made in this paper. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
Two is better than one: Using a single emotion lexicon can lead to unreliable conclusions PONE-D-22-09013R1 Dear Dr. Stillwell, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Andrea Fronzetti Colladon, Ph.D. Academic Editor PLOS ONE Additional Editor Comments (optional): I appreciate your revisions and believe the paper represents an excellent contribution to the journal. Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #2: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #2: The authors have worked hard to improve and refine the paper, and it shows. The revision addresses all of my earlier comments. The paper will be useful to a large number of people that use emotion lexicons. Congratulations on the fine work. Final comment: Instead of citing the arXiv versions of papers, where applicable, cite the journal or conference proceedings where the papers were eventually published. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #2: No ********** |
| Formally Accepted |
|
PONE-D-22-09013R1 Two is better than one: Using a single emotion lexicon can lead to unreliable conclusions Dear Dr. Stillwell: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Prof. Andrea Fronzetti Colladon Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .