Peer Review History
Original SubmissionFebruary 17, 2021 |
---|
PONE-D-21-05380 Passive citizen science: the role of social media in wildlife observations PLOS ONE Dear Dr. Edwards, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by July 3rd 2021. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Daniel de Paiva Silva, Ph.D. Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and 2. In your Methods, please provide a statement indicating that the use of this dataset was done in compliance to the National Biodiversity Network and Flickr Terms and Conditions and our requirements for this type of study (https://journals.plos.org/plosone/s/submission-guidelines#loc-personal-data-from-third-party-sources). Additional Editor Comments (if provided): Dear Edwards et al., Greetings! After the evaluation by two independent reviewers, both of them found that your manuscript has great publication chances. Still, improvements are always required. Please take a closer look on what both reviewers indicated about your manuscript, and arrange the necessary improvements they asked. By the time you are required to resubmit a new version of your text, please do not forget to prepare a rebuttal letter indicating all the changes you did and justifying all of those you did not. Considering the pandemic scenario, a three-months period (resubmission due on July 3rd 2021)may be enough for you to prepare all the changes. If not, please let me know how could we arrange a better schedule for you. In case you are able to resubmit earlier then the provided due, do not hesitate to do it. Sincerely, Daniel Silva [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Partly Reviewer #2: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #2: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: In the manuscript “Passive citizen science: the role of social media in wildlife observations”, the authors assess the usefulness of social media data (images on Flickr) for providing species occurrence data. To do this, they extract geolocation and temporal information from Flickr photos and, treating this as a species occurrence record, compares this to records on the National Biodiversity Network (NBN) Atlas. They use Google Cloud Vision API and species’ taxonomic classifications to verify the species tagged in the Flickr images. After evaluating 1500 UK species, as well as numerous invasive species, they found that the Flickr data best represents the NBN data when considered over large scales (40km grids) without temporal constraints. The authors conclude that Flickr can potentially provide useful data on species’ distributions, especially when combined with existing biodiversity data. Overall, the manuscript is succinct and well written, and I thoroughly enjoyed reading it. The topic of the study is interesting and timely. Traditional citizen science has gained traction in recent years as a reliable and efficient means of gathering wildlife data. Therefore, it seems only fitting to explore this further by evaluating the potential contributions of ‘passive citizen science’ and specifically social media. The authors provide a good overview of the current state of the literature and identifies a clear research gap. In general, I found the methods to be appropriate and relatively clear, but some sections would benefit from further clarification (please see specific comments below). Likewise, I found the discussion to be sound based on the presented results (which were generally clear but inconsistent in some areas; see comments below). However, most, if not all, citizen science datasets are subject to questions surrounding data quality and the Flickr data is no exception. In this case, I think that there are some caveats to the analyses that need to be addressed. INTRODUCTION: Lines 78-83: The authors refer to a paper by ElQadi et al. (2017), describing the methods in this paper as “exact matching between species names … and the labels (tags) returned by Google’s Reserve Image Search”. They state that this can lead to “false negatives where the Google API provides only a more general label or another name for the species”. While I concur that an “exact matching” method would result in false negatives, I do not completely agree with the authors’ interpretation of ElQadi et al.’s methods. ElQadi et al. searched for images on Flickr using species’ scientific and common names. These images were then given a text tag using Google Reverse Image Search. They used tag frequency to filter out irrelevant images, such that tags with low frequency (regardless of their relevance) were excluded. Consequently, images with relevant tags, including tags of the species’ scientific or common names, were sometimes nonetheless excluded if the tags were infrequent (see paragraph 2 of their discussion). Yes, ElQadi et al.’s method still results in false negatives, but it is not due to “exact matching” between species names and Google’s tags. “Exact matching” implies that an image was deemed relevant (and thus included after the filtering step) only if the Google tags exactly matched the species names, which was not the case. For example, in ElQadi et al.’s study, Flickr images of the Blue banded bee (Amegilla cingulata) that were tagged by Google as “Amegilla cingulata” and “blue banded bee” were nevertheless excluded after their filtering step because of the relative infrequency of these tags. In other words, false negatives occurred even when Google API tagged the exact species name. Furthermore, photos of blue banded bees that were given frequent general tags, such as “insect” and “bee”, were included after the filtering step. Thus, false negatives do not always occur when general labels are provided. The authors’ description should therefore be clarified. Further mentions of the “exact-match” method in relation to ElQadi et al.’s paper (Lines 278 and 393) should also be clarified. METHODS: Line 135: To conduct their temporal analyses, the authors downloaded date information (the date that the image was uploaded) from each Flickr image. However, I expect some images would have been uploaded some time (potentially a long time) after being taken. Is there any way of knowing when the photo was actually taken, as opposed to uploaded? This information might not be readily available, but if there are frequent discrepancies between when photos were taken and when they were uploaded, it is difficult to have confidence in the results of the temporal analyses (seasonal, half yearly, and yearly patterns). If this information is not available, the authors should at least address how this might have affected the results of the temporal analyses. Lines 171-174: This is a fair statement, but I am wondering what happens if the Flickr image features another species of grass that has been incorrectly tagged on Flickr as Timothy grass? Would the coarse match still be successful in this case? If so, how often might this occur and how might this impact the results? While I think the authors’ image verification technique goes some way in improving previous methods, it appears that the reliability of Flickr-derived species occurrence data still depends significantly on the ability of the Flickr users (i.e., the ‘passive citizen scientists’) to correctly assign species name tags to Flickr images. The authors should discuss this if this is the case. RESULTS AND DISCUSSION: Lines 277-279: What were the 10 species and how were these chosen? I think it would be ideal to choose the species randomly, but either way, a justification should be provided. My concern is that the image verification approach may perform better for certain species or taxonomic groups than others. Given this, I think it would be best to evaluate the performance of the verification technique using the full range of species available in the dataset. General comment: The authors did a good job addressing the key points in this section. I would have liked some more discussion around why the authors thought that constraining the analyses to a 12-month window led to poorer results than a 3 or a 6-month window. On Lines 323-325, the authors attribute the poor results in the 12-month analyses to a “lack of data on Flickr”. Intuitively, should this phenomenon (i.e., the lack of data) not be even more pronounced when constraining the analyses to an even shorter time frame (i.e., 3 months and 6 months)? FIGURES: Overall, the figures were relevant and suitable for the data. I particularly appreciate the inclusion of Figure 1. The flow chart and the matching sub-headings in the methods were helpful. MINOR ISSUES: Lines 33-34: The authors state that “geotags given to photos on social media sites … are assigned automatically by GPS location systems”. Whilst this may be true of photos taken on a location enabled device, there would be cases where automatic geotags are not available (e.g., on a device without a GPS system), or where the geotags are low in accuracy. Does Flickr allow users to manually add or edit location data? If so, I think this needs to be stated, either here or in the methods. Line 126: What is meant by “exact search”? It would also be good to get an idea of how many records included irrelevant species in any given search. Line 136: It would be helpful to add more detail about the Flickr tags. Are they manually entered by the user/uploader, or are they automatically assigned by the computer, or both? How many tags are allowed? The authors did an excellent job of explaining the Google Cloud Vision API tags (lines 151-157), so something like this, explaining the Flickr tags and clearly distinguishing them from the Google API tags, would be great. I initially had some difficulty following the authors’ methods in relation to which tag they were referring to (i.e., the Flickr tags or the Google API tags). Line 169: The authors write, “Using the exact matching on the Flickr label of Adder would not have found any match”. This is unclear to me. Why would the Flickr label “Adder” not have found an exact match when “Adder” is indeed one of the species’ common names in the NBN taxonomic classification (Figure 2)? Is it because Google API does not label the image of the adder as “Adder”? Please clarify. Lines 177-178: “… for a 100 randomly selected instances”. Either a word is missing here, or the “a” should be removed. Lines 229-230: “… and a 100% of species”. Either a word is missing here, or the “a” should be removed. Lines 263-266: This sentence would be clearer if these species were listed in order based on their Flickr species count, starting with the best represented species. Lines 278-279: “… for a 1000 randomly selected instances”. Either a word is missing here, or the “a” should be removed. Line 307: Rather than writing “correct”, I suggest simply stating that 38% of the Flickr data were also reflected in the NBN data. These mean slightly different things as an absence of NBN data from a location (in the presence of Flickr data) does not necessarily mean that the Flickr identifications were incorrect. Rather, the species may have just never been recorded there previously (which the authors note on Lines 415-416). In fact, these “false positives” could actually be interesting datapoints as potential range expansions, especially given invasive species are a substantial focus of this manuscript. Line 331: Did the authors mean to write “true positives” rather than “false positives”? False positives were not used to calculate recall. Line 372: The authors mentioned seven distinct species but presented only six in the table. Please clarify. Line 407: It would be good to provide a number or percentage here alongside the claim of “a large number”. Table 1: I think there is a missing row here. The table legend (and Line 234) refers to the “top 10” most frequently recorded species on Flickr, but only 9 species are shown in the table. Tables 1 & 2: Please check the Flickr count number for Erithacus rubecula (Continental Robin). The count is given as 19,248 in Table 1 but 2,786 in Table 2. Meanwhile, the NBN count is the same in both tables. If this is not an error, please provide an explanation as to why these differ. Table 3: The table legend states, “Species occurrences … with number of species above 100”. Did the authors mean to write, “with number of occurrences above 100”? Figures 4-7: I suggest making the left and right panels/graphs the same width, especially in Figures 4 and 6, so that the gradients of the lines can be compared. Reviewer #2: This paper has the aim to put into value the big amount of biodiversity data hidden in social media. Specifically, they put to the test the photo base on Flickr. Flickr data were evaluated relative to the National Biodiversity Network (NBN) Atlas, the largest collection of species distribution data in the UK. They used an innovative image verification technique that uses the Google Cloud Vision API in combination with species taxonomic data to determine the likelihood that a mention of a species on Flickr represents a given species. Authors found that data are useful for representing some common and diurnal species on a space scale but not on a temporal scale. This work is an interesting and useful contribution to better understanding the value of biodiversity data from social media to contribute to science. However, there are some recommendations to get this article improved for publication. Introduction First, I recommend better explain the relation of this manuscript in the context of citizen science on introduction. The definition of citizen science still an open discussion today and this concept of “passive citizen science” could be controversial. I propose to better definition of passive citizen science immerses in a framework of citizen science definition. Material and Methods I consider the material and methods are ok. However, I recommend better explain how the NBN database is performed. Where data came from? Results and discussion Results are generally well explained and represented. But I consider some parts of the results are weakly discussed and more scientific references would be welcome. You could discuss also the potential of this data, for example comparing it with other more active citizen science methods based on submitted photographs (e.a, iNaturalist). More specific comments: Line 3: Please check the references on the whole text. Is this the correct format for the journal? Lines 54-56: Please, better explain how the NBN database was made. Line 62: Please, abbreviate “National Biodiversity Network Atlas” Lines 371-376: Discuss these results. The occurrence of common species in citizen science data is registered in some articles you could refer to. Lines 413-416: I do not understand this sentence. NBN database is a citizen science database? Line 416: there is a parenthesis ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: Yes: Maria Isabel Hermoso Beltrán [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
Revision 1 |
Passive citizen science: the role of social media in wildlife observations PONE-D-21-05380R1 Dear Dr. Edwards, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Daniel de Paiva Silva, Ph.D. Academic Editor PLOS ONE Additional Editor Comments (optional): Dear Edwards et al., after a final review round by one of the reviewers who previously reviewed your original MS, I believe you text is now suitable (and formally accepted) for publication in PLoS One! Congratualations! Sincerely, Daniel Silva, Ph.D. Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: I thank the authors for their careful and considered responses to my concerns and queries. These have been satisfactorily addressed. The authors’ revisions have improved the clarity of the manuscript and I specifically appreciate their efforts to clarify their methods. The additional information should provide readers with sufficient information to make an informed judgement on the strengths and potential limitations of the current study. Overall, this is a great manuscript that demonstrates the potential of social media to provide valuable biodiversity data. I noticed just a few additional (very minor) issues which can be easily addressed: Line 324: Are you missing a word in “This allows to identify…”? Line 345: Please check the date range (2006-2018). Is this date range meant to be different to the temporal analysis date range (2004-2017; Line 327)? Line 397: Are you missing a “with” in “associated a given species”? Line 513: Please remove the duplicate “that”. Line 589: Should “observation” be “observations”? Figure 2: “Coarse match b-n NBN Google labels” – what is “b-n”? Between? ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No |
Formally Accepted |
PONE-D-21-05380R1 Passive citizen science: the role of social media in wildlife observations Dear Dr. Edwards: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Daniel de Paiva Silva Academic Editor PLOS ONE |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .