Peer Review History
| Original SubmissionFebruary 8, 2022 |
|---|
|
PONE-D-22-03986Quantifying Gender Bias Towards Politicians in Cross-Lingual Language ModelsPLOS ONE Dear Dr. Stanczak, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Jun 05 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Yang Xu, PhD Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. Thank you for stating the following financial disclosure: “This work is partly funded by Independent Research Fund Denmark under grant agreement number 9130-00092B.” Please state what role the funders took in the study. If the funders had no role, please state: "The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript." If this statement is not correct you must amend it as needed. Please include this amended Role of Funder statement in your cover letter; we will change the online submission form on your behalf. 3. Thank you for stating the following in the Acknowledgments Section of your manuscript: “This work is mostly funded by Independent Research Fund Denmark under grant agreement number 9130-00092B.” We note that you have provided additional information within the Acknowledgements Section that is not currently declared in your Funding Statement. Please note that funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: “This work is partly funded by Independent Research Fund Denmark under grant agreement number 9130-00092B.” Please include your amended statements within your cover letter; we will change the online submission form on your behalf. 4. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability. Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized. Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access. We will update your Data Availability statement to reflect the information you provide in your cover letter. 5. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. Additional Editor Comments: This paper provides a computational analysis of gender bias toward politicians in language models. Three expert reviewers have offered detailed comments on the manuscript. Although overall the reviewers agreed that the authors have made an attempt at a valuable problem, they pointed out many issues in the current version (most prominently, the lack of thoroughness in the exposition, and the extension and deepening of analyses) and provided extensive feedback on how to improve the manuscript in a major revision. The manuscript as it stands is not acceptable for publication at PLOS, and I encourage the authors to address the reviewer comments carefully in the revision. R1 raises an issue about the input sources of the language models and controlling for the input. R1 also suggests a controlled analysis based on temporal considerations and raises a question about the discrepancy between English and other languages in the top adjectives generated. R2 raises a concern about the choice and relevance of politician names and suggests separate analyses to be performed on subsets of the names, tailored for individual languages. R2 and R1 indicate on the reviewer feedback form that the data underlying the findings have not been made fully available, which is contradictory to the authors’ indication on the submission form — if this is a misunderstanding, please indicate clearly where the data can be accessed in the manuscript (and in the forthcoming response letter). It is also advised that the authors provide the link to the code repository that replicates all the analyses reported in the manuscript to be revised. R3 suggests that the writing needs be substantially improved in terms of its structure, justification for the methodological choices (see R2’s concurrent comment point #3., and R1’s final comment), and correction for an overstatement about novel methodologies. R3 also points out correctly that the results presented seem surface explorations, and more in-depth analyses should be considered to address this issue in the cross-linguistic setting which include cross-language variability, reliability of methodology across languages and politicians. R2 concurs with this comment suggesting that the paper presented similar methodologies to an existing publication on gender word associates, but more analyses should be developed across languages, as well as those that provide better interpretation of the results and some correlational analyses on term frequencies. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Partly ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No Reviewer #2: No Reviewer #3: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: No ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The submitted work investigates the inherent gender biases learned by large contextualized neural language models (LMs) by probing sentiment biases of words that LMs associate with both male and female politicians. Distinct from previous work that attempts to quantify gender biases in language models, the authors consider a multi-lingual setting where both monolingual LMs and cross-lingual LMs are analyzed. Combined with previous methods in quantifying gender bias (PMI and latent variable model from Hoyle et al. (2019)), the authors propose a novel experimental setup that considers lexical sentiment scores of words that are predicted by the LMs to directly supersede or following the named entity of interest. Under the proposed experimental setup, gender biases in LMs are quantified by categorizing the list of words with the largest deviation w.r.t. gender, as well as statistical tests w.r.t. hyperparameters such as model size, model architecture, and language. The proposed methodology and experiment setup has the potential to be extended to analyze gender biases toward other types of named entities. The authors promise to make the data fully available but I was not able to find the data in the materials submitted. Detailed questions and comments regarding the submission: - What datasets are the language models pre-trained on? It is conceivable that the data sources themselves could be inherently biased in different ways (e.g. different news agencies having different biases). When evaluating the effect of specific model architecture on biases captured, it may be worthwhile to consider a more controlled setting where all models are trained on the same data source. - Have you controlled for the time period in which the politicians are active? If so, which time periods do you consider? Not only that gender biases change over time, but also political sentiment in general. These changes can confound your result where some sentiment differences are not necessarily caused by gender differences. - The top 100 generated adjectives are considered for all languages but only the top 20 adjectives are considered for English. Is there a reason for this discrepancy? - In Table 4, what are the number of parameters associated with the model sizes (i.e large, xlarge, xxlarge)? Are these sizes consistent across the different architectures in terms of parameters counts? Reviewer #2: The paper describes the problem of gender bias in language models and conducts a cross-lingual study to identify it. The gender-bias in language models is proxied through politicians' names and sentiment lexicon. The main claim is straightforward. The degree of gender bias is different in each language, and larger language models are not significantly more biased than smaller ones. The problem studied in this paper has fundamental social implications and the paper has an interesting point of view. However, I have some concerns that require revision. 1) My main concern is about the choice of politicians' names. According to the paper, 250k names are retrieved from Wikidata. It is assumed that because these 250k names are for politicians and the fact that politicians have higher toxicity associations and bias in NLP models (Prabhakaran et al.), then the chosen names should represent the embedded gender bias in language models. However, I believe that not all of the names are relevant in all the 7 languages studied in the paper. Although the BERT-based models are trained on Wikipedia, I doubt that a translated article about a politician on Wikipedia is enough to impose gender bias on the model. I strongly suggest that the experiments be run separately on a smaller set of names specific to each language to reduce the noise (e.g., for the Chinese language, only experiment on the Chinese politicians + a few internationally well-known politicians). 2) Most of the experiments in the paper are done based on the frequency with which the top K words associated with a gender belong to a category (e.g., Negative sentiments, Positive sentiments, or a supersense). Although the results are interesting and should be reported, they are very similar to the ones in the previous work of Hoyle et al., 2019, except that now they are done cross-lingually and based on the politicians' data. I believe more experiments and tests should be done to assess gender bias and its differences across languages and language models. Exploring questions like which supersenses are more/less responsible for gender bias in different languages/LMs, and how systematically similar or different are the top gender-associated adjectives across languages can further strengthen the findings. 3) I faced some problems understanding section 5.1.2 and more specifically figure 3. How are you deriving the 100 largest-deviation adjectives for male and female genders? Isn't \\eta(w, s) used for the deviation metric? If so how are you marginalizing the sentiment in this function? I think adding some detail about deriving the frequency of largest-deviation adjectives helps clarify this section. 4) The paper proposes a critical point about the gendered adjectives and neutral words, and how they might mitigate the bias in language models trained on gendered languages (e.g., in Spanish). I am curious if there are ways to examine this proposal against the biases detected in the LMs corresponding to these languages. Some correlation analyses between the frequency of feminine/masculine inanimate words and the frequency of negative/positive adjectives associated with each gender might be one simple approach to this. My general argument is that the methodology used in the paper is statistically sound and evaluated in previous work, and the findings support the claim (although the claims themselves are not impressive), however, the experiments should be extended to deliver a more extensive set of findings. Reviewer #3: This paper explores gender bias toward politicians in transformer language models in a cross-lingual scenario. In particular, it investigates the adjectives/verbs associated with a large number of politicians in 7 languages, and measures deviations in sentiment toward male/female/'other' politicians in the respective language groups. While the motivation and contribution of data set is laudable, as is the treatment of gender as non-binary. However, the paper (1) is not well-structured, at times repetitive and difficult to follow (esp for non-domain expert readers as attracted in PLOSONE); and (2) lacks motivation for several methodological decisions. Finally, the paper abstract/intro suggests a methodological contribution, while in section 4.2 it becomes clear that it simply re-uses a previously proposed method; and this method is not even described in detail. In sum, I think this paper can make a worthwhile contribution by studying an important phenomenon at a broad scale, however, it needs careful rewriting and very careful motivation and validation of all steps of the approach. Please find detailed comments/suggestions below. Comments on paper quality / structure - Section 3 dives right into templated probing methodology without motivating this approach. This compromises understandability of the paper, especially for an audience not focused on NLP (like PLOSONE). I think the paper would be more readable if Related Work was presented earlier on. - Similarly, for section 3.3. I think an illustration early on in the paper, of the overall approach (and goal) would help. - Information is often presented in confusing order, e.g., a long list of models used (lines 133ff.) interrupting the explanation of generation methodology/POS filtering. - The experiment setup is incomprehensible. A paper should be self-contained and "we conduct the same analysis as presented in [12]" (l. 244) is not sufficient. Comments on methodology - A discussion of the type of bias targeted would help, in particular (as far as I understand) it addresses bias based on properties of the *speaker* (i.e., language user) not bias toward properties of the entity being *spoken about* (i.e,. the politician). These dimensions have been disentangled and addressed separately (and jointly) in prior work (e.g., [1]). - The exact method of retrieving politicians names remains unclear. Was the process quality-checked? How well did it work for each language? How do you align politicians' names across languages? - The generation method still relies on syntactic cues, and will work better for languages in which the required bigrams are more plausible. E.g., for EN presumably all adjectives and most verbs will have very low probability. To what extent is the ranking meaningful? Was there any quality analysis across language? - Please motivate all parameters, including choice of languages, cut-off parameters (show the top20 / top100 words, rather than counts only) - An ablation study on comparable names (e.g., same number of subword tokens, comparable popularity) would have been interesting to disentangle LM perplexity from actual bias effects. - Similarly, the approach for creating sentiment lexica seems highly complicated. How did you choose the base models? Is there any evidence for the hypothesis that "combining a larger number of individual lexica with SentiVAE leads to more reliable results" (l. 173), i.e., how well did the most competitive base model perform in the tasks in Table 10? - Table 10 does suggest significant drops for Chinese, Hindi and French; as well as a very low overall performance for Spanish. The conclusion that "the performance of over 72% is satisfactory and validates the lexicon approach" seems premature given that the bias detection method relies on quality sentiment classification, and the unconvincing results in Table 10. - Line 219 treats gender as binary, while the data set seems to be more nuanced (m, f, other). Please explain / unify. Comments on results / analyses - Given that the contribution is a *cross-lingual* analysis, the corresponding results section is very shallow, the main take away being relative bias of small vs larger LMs. - There's a missed opportunity of exploring in-depth research questions/analyses such as: are there systematic differences in bias across languages? How reliable is the method across languages (incl name selection, data generation, bias measurement)? How reliable is the method for high/low coverage politicians (potentially including politicians of the same vs. different native language). [1] https://aclanthology.org/2020.emnlp-main.23.pdf ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
PONE-D-22-03986R1Quantifying Gender Bias Towards Politicians in Cross-Lingual Language ModelsPLOS ONE Dear Dr. Stanczak, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Sep 22 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Yang Xu, PhD Academic Editor PLOS ONE Additional Editor Comments: Two reviewers provided comments on the revised manuscript. Although R1 found the paper acceptable, R2 raised a series of issues that should be addressed and discussed. Specifically, R2 suggested an in-depth analysis on the semantic representations of language models (that could have given rise to the results), a clearer explanation of the methodology in Sec 5.2 (with the possible addition of an illustrative figure) and the working of SentiVAE in Sec 4.3, a more rigorous evaluation in Sec 4.3 e.g., using potentially human annotations and ablation, and a more detailed description of the supersets in Sec 6.1 and 6.1.2. These are reasonable comments, and I look forward to reading the next revision. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #2: (No Response) ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: No ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The authors have adequately addressed comments raised in my previous review and have specifically improved clarity on their experimental setup as well as motivations behind various experimental decisions. An additional experiment has been performed to investigate the temporal effect of political sentiment. To further improve the paper’s clarity, it may be worthwhile to quickly elaborate on the decision to only include 20 adjectives in the main paper (e.g. selecting top 20 adjectives in English already offers a larger pool of adjectives) similar to what has been provided in the response letter. A sample of the dataset used has now been provided and can be accessed. The authors promise the public release of their data upon publication. Reviewer #2: The paper describes the problem of gender bias in language models and conducts a cross-lingual study to identify it. The gender bias in language models is proxied through politicians' names and sentiment lexicon. The main claim is straightforward. The degree of gender bias is different in each language, and larger language models are not significantly more biased than smaller ones. Regarding my previous comments, the authors included new experiments for different sets of politician names. The results were like that of the initial experiment. However, I still believe that the paper lacks a more in-depth analysis. The current experiments show that language modes suffer from gender bias toward politicians, but this bias does not depend on the size of the model but on the language itself. Investigating how the representation of these languages varies in the semantic space generated by the language model could explain the current results. The methodology is almost completely borrowed from Hoyle et al., 2013. The paper, however, does not clearly explain how this method works in section 5.2. The current description is rather incomplete and difficult to follow without reading Hoyle et al., especially for those unfamiliar with NLP methodologies. I recommend rewriting this section to increase its readability. Adding an illustrative figure to summarize the approach would be helpful as well. The same thing applies to section 4.3. How does SentiVAE work? What are the differences in the chosen sentiment lexica? The generation of the multilingual sentiment lexica is a valid and important contribution of the paper and should be emphasized, but only a few paragraphs are dedicated to explain the approach and the results. Also, a more rigorous evaluation is needed for this section, especially for Hindi and Spanish, as the F1 scores are lower than the rest of the languages. Currently, the authors evaluate the generated sentiment lexicon based on the average sentiment of words in a sentence. Having a token-level approach toward sentiment analysis would indeed generate some noise. Using human annotations (even for a small section of the words) or conducting correlation analysis with the initial sentiment Lexia might be more accurate. As an ablation study, the authors could repeat the experiments in section 6 with one of the monolingual sentiment Lexia. This experiment clarifies whether using SentiVAE helps to identify gender bias in the language models or not. I still recommend the authors to provide a more in-depth analysis that would go beyond the experiments and the results of the previous studies. The paper studies an important question and the authors have collected valuable data to investigate the question, but the experiments and the results provide only a surface-level analysis of this problem. The readability of some parts, especially the methodology and the experiments, should be improved as well. Writing comments: In section 6.1., the authors could provide a description of the supersets used and the set of words each contains. I find it difficult to understand the first paragraph of section 6.1.2. Could you provide some examples for each step mentioned in this section? ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No ********** [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 2 |
|
Quantifying Gender Bias Towards Politicians in Cross-Lingual Language Models PONE-D-22-03986R2 Dear Dr. Stanczak, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements and the final comments raised by the reviewer. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Yang Xu, PhD Academic Editor PLOS ONE Additional Editor Comments (optional): I have received final comments from Reviewer 2, who indicated that the authors have managed to address the main issues raised in the previous revision. I will accept the manuscript for publication if the authors can aim to address the lingering (more minor) issues raised by the reviewer in the final version, quoted as follows: 1) In Table 4, why is the word "great" in Tmasc_neg? Similarly, why "first" and "out" are in this column? There seem to be some noises in the proposed methodology, even in the top 10 adjectives. Some explanation or error analysis is needed here. 2) Where are the results on other languages and language models for the clustering experiment in Table 7? Currently, it is only done (or reported) for French and m-BERT. 3) The clustering and PCA experiments could both be discussed in more detail. There are no significant patterns discovered. How does this correspond to the results of the previous sections? (Other than the fact "that more gendered words are generated more frequently in association with their respective gender", which is expected regardless of the gender biases). 4) The quality of the images could be improved (especially Figure 3, and Figure 6). Additionally, I'd encourage the authors to offer some discussion on the language-specific effects found in their study---why gender bias patterns turned out the way they did, e.g. prior to or within the conclusive paragraphs. Wording suggestion in Abstract: "we use to effect a multi-lingual study of gender bias towards politicians ..." ==> "we use to enable a multi-lingual study of gender bias towards politicians ..."?, or perhaps "we use to facilitate a multi-lingual analysis of gender bias towards politicians ..." Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #2: (No Response) ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #2: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #2: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #2: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #2: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #2: The authors have addressed my previous comments by adding new experiments and have improved the clarity of their method section. I have some minor comments and questions for the authors that perhaps need to be addressed in the paper before submission. 1) In Table 4, why is the word "great" in Tmasc_neg? Similarly, why "first" and "out" are in this column? There seem to be some noises in the proposed methodology, even in the top 10 adjectives. Some explanation or error analysis is needed here. 2) Where are the results on other languages and language models for the clustering experiment in Table 7? Currently, it is only done (or reported) for French and m-BERT. 3) The clustering and PCA experiments could both be discussed in more detail. There are no significant patterns discovered. How does this correspond to the results of the previous sections? (Other than the fact "that more gendered words are generated more frequently in association with their respective gender", which is expected regardless of the gender biases). 4) The quality of the images could be improved (especially Figure 3, and Figure 6). ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #2: No ********** |
| Formally Accepted |
|
PONE-D-22-03986R2 Quantifying Gender Bias Towards Politicians in Cross-Lingual Language Models Dear Dr. Stanczak: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Yang Xu %CORR_ED_EDITOR_ROLE% PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .