Peer Review History
| Original SubmissionMay 19, 2020 |
|---|
|
PONE-D-20-14992 The effect of metacognitive training on confidence and strategic reminder setting PLOS ONE Dear Dr. Engeler, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. I have now received three expert reviews and have also read the paper myself. As you can see, the comments are largely positive and I agree that there is much to like in this paper. To echo Reviewer 3, one laudable aspect is the preregistration of the analytic plan and the fact that the paradigm produces several metacognitive measures is also in itself a contribution to the literature. Each reviewer has provided informative comments that I hope you will find useful. There are two types of key concerns that arose from the reviews. The first type of key concern is a relatively simple issue of clarifying the methods sections, to give provide more details about the online administration of the study and the task itself. I think part of the confusion might be the use of the word “trial” to refer to each block of 25 circle-drags (note that Reviewer 1’s comment 6 uses “trial” to refer to each circle drag; it was also a confusion that I had). Perhaps referring to each trial as a “block” might sidestep the issue. The second type of key concern is about how exactly the data should be interpreted. For example, Reviewer 2 asked about the specific wording of the global metacognitive judgments questions. I suspect that one reason why they ask that question is to get at what participants specifically participants were being asked to make judgments about. For example, was the first global metacognitive judgment a prediction of how participants would perform on future trials or a postdiction of how well they did on the forced trials? Similarly, did participants interpret the second global judgment as a postdiction of how well they did in the choice trials? One reason this matter is that metacognitive bias (which is the DV on which effects are found) is based on the difference between the judgments and their actual performance on the forced trials. However, interpretation of this difference as over- or underconfidence is muddied if participants think they are making metacognitive judgment about anything other than performance on the forced trials. Another example is comment 5 by Reviewer 1 -- if participants are aware of flagging attention, this relatively judgment could appear as a positive reminder bias. I do not expect that these are questions that can be definitively answered without a new study, but should at least be raised and addressed in the general discussion. Finally, I had another more minor suggestion for the analysis of the relationship between reminder bias and metacognitive bias: currently, four correlations are reported and an interaction is implied. I would recommend considering using regression analyses to test the implied interaction (e.g., is the relationship between reminder bias and internal metacognitive bias moderated by group?). I look forward to seeing the next iteration of this manuscript. Please submit your revised manuscript by Aug 23 2020 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: http://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols We look forward to receiving your revised manuscript. Kind regards, Veronica Yan, Ph.D. Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. We note that you have stated that you will provide repository information for your data at acceptance. Should your manuscript be accepted for publication, we will hold it until you provide the relevant accession numbers or DOIs necessary to access your data. If you wish to make changes to your Data Availability statement, please describe these changes in your cover letter and we will update your Data Availability statement to reflect the information you provide. Additional Editor Comments (if provided): <accurate> [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions</accurate> Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Partly ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No Reviewer #2: Yes Reviewer #3: Yes ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #2: Yes Reviewer #3: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: The authors examine whether a training period will shift metacognitive judgments about reminder-setting and adjust how many reminders learners choose to set. Learners move digital circles to assigned different sides of a box. Learners are randomly assigned to a metacognitive training condition in which they predict how many they will correctly move and receive feedback about their performance and predictions during forced trials OR to a baseline condition in which they do not practice metacognitive judgments. Practicing metacognitive judgments improves later metacognitive predictions but does not shift when and how they choose to use reminders. I think cognitive offloading is an incredibly interesting and important area for investigation and understanding influences on it is worthwhile. This manuscript provides an incremental step towards greater understanding of cognitive offloading and reminder setting. 1. I wondering if practice making metacognitive judgments is truly a kind of “training.” The authors consider training to be making metacognitive predictions and getting feedback about their predictions. This does not fit my ideal of what training entails. Training seems like it would entail some kind of instruction. It may be more accurate to think of practice predictions as experience with the task, which then improve later predictions. Those that get experience making predictions become better at making predictions (and this improvement at the task continues throughout the experimental trials, even without feedback about estimates – i.e. the interaction of judgment time with other variables). Practice with this unusual or difficult task improves later performance at the task. 2. The primary conclusion (practice making metacognitive predictions does not change reminder setting) is evidenced by a null effect. The authors find no differences between conditions in reminder selections. This concern is somewhat allayed by differences in metacognitive monitoring accuracy, but null effects are not as powerful as significant differences. The authors suggest reasons why there may be null effects but cannot test their hypotheses. Understanding how to shift reminder setting and showing how to do so would be interesting future directions. 3. To calculate optimal selections, participants must compute a math equation – there is no other way to select optimally. Therefore, this “metacognitive” task may be more of an algebra problem than what is typically considered as metacognition. Succeeding at a math problem seems very different than on-going metacognitive monitoring. Do optimal selections in this task relate to math ability? How can learners choose optimally if they do not know how to solve the math problem? 4. I think that future directions could explore item-by-item reminder selections. Item-by-item selection processes could better mimic how learners choose to offload memory in the real world. Learners do not choose to offload an entire block of information; rather, they choose what items and information to offload on an item-by-item basis. 5. To calculate optimal allocation, the authors use performance on the forced trials at the beginning of the experiment. This could be inaccurate if learning occurs during training and performance improves. Alternatively, this could be inaccurate if fatigue and proactive interference happens and performance declines. In other words, the temporal separation of the forced trials from the choice trials could produce biases in calculating optimal selection. The procedure assumes that performance is static across the entire experiment, which seems unlikely. Further, we know attention waxes and wanes. If learners are sensitive to changes in their attention (as could be suggested by prior research, e.g., Markant, DuBrow, Davachi, and Gureckis, 2014), then they may be exercising better metacognitive control over reminder setting than can be picked up by the rough measurements in this procedure. 6. I think the authors need to be clearer about the 1-9 point assignment phases. It was unclear whether individual circles were assigned point values or whether the entire block of trials was assigned those point values (I figured out that it was the latter possibility, but it took a bit of time). 7. The authors need to be clearer about where they got their degrees of freedom around lines 279 8. The authors sometimes average across the two judgments and sometimes include the time point as a variable in the analysis. Can they be clearer about the differences in accounting for time point? And given the interactions with time point and other variables, does averaging across time point in some analyses obscure important differences? 9. I do not see a link to the data file used for this project. Reviewer #2: I find this manuscript interesting, informative, and clearly written. I appreciate the authors’ work and can see why it would be of interest to those examining cognitive offloading. However, there are some potential areas for improvement, particularly in the motivation of the current work and the specification of the sample. I outline my suggestions below, in the order I noticed them in the paper. Overall, I found that the authors could improve their situating of the current work, and describing why it is compelling – why would improving metacognitive judgment accuracy/reducing bias in offloading behavior be helpful? I think this motivation could be improved in both the Introduction and Discussion sections, perhaps using practical examples of why these factors matter. The larger issues at play here are interesting and could be enhanced in this manuscript. --- The sample was collected via Amazon Mechanical Turk. Given the rather sizeable existing literature on the proper screening of MTurk participants, I think it's important to have additional detail about the sample. How was age confirmed? Did you use Amazon MTurk's pre-screening age qualifications? To confirm that people were in the US as you stated, did you use some kind of IP address locator? Or were these items self-reported? If participants did self-report, how did you control for lying? (MTurk participants commonly respond to demographic questions based on their perceptions of demand characteristics, and are known to share those demand characteristics with their fellow MTurkers once they are discovered). Some studies have reported that American MTurkers are more highly educated on average than other samples. Did you collect educational data, and if so, was that a factor (given the extant literature on the relationship between education and metacognition)? How about across age, as the sample ranges quite widely from 20-71 years old, and some literature suggests that younger and older adults differ in both their memory and metacognitive capabilities? Did you use “catch questions” or “bot checks” to ensure participants were paying attention? If so, what were these items? When did they appear? If not, how did you verify that your sample did not contain bots? Were participants screened to be fluent in English, or to have English as a first language? Among those whose first language was not English, how did you ensure fluency/understanding of the instructions, especially given that the task has multiple components? Do the authors have a measure of how often Mturk participants left the page (e.g., to click to another window) when they were doing this task to assess potential distraction/lack of attention? The authors report that “the experiment took approximately 45 minutes.” What was the standard deviation in time spent? Were data from any especially quick or especially slow outliers discarded? Could participants have clicked through very quickly (e.g., skimming through instructions), or were there minimum requirements for time spent on each page? ---- For conciseness, some repetitiveness between the Optimal Reminders Task section and the Procedure section could be decreased or consolidated. Lines 236-237, please provide exact wording of the final metacognitive questions – was it some version of “how accurately do you think you performed on this task?”. Finally, this may or may not be possible, but I think it would be interesting if readers could be directed to a quick video of the task in action in supplemental materials. Reading about it and seeing photos help, but a screen recording of the task could be even clearer and more compelling. If this exists elsewhere, directing readers toward it would be helpful. Reviewer #3: The authors investigate behavior in a recently-developed metacognitive task in which participants drag a set of numbered circles to the bottom of the screen in order. Certain circles are initially marked in a different color and must be dragged to different locations, for which participants may optionally set an "external reminder." By manipulating the rewards & costs associated with setting a reminder, the paradigm allows the researchers how optimally people use external reminders. In the present study, the researchers additionally created an intervention in which experimental (but not control) participants made predictions and received feedback in the first block of trials. In the second block of trials, the experimental group showed more accurate performance predictions for both internal and external memory, but both groups still showed a behavioral bias towards setting more external reminders than optimal. Overall, I found much to like about this manuscript. One clear strength of the manuscript is the authors' commitment to open science. The authors have preregistered the participant exclusion criteria, experimental procedure and the analytic plan, and they have adhered to this closely. Among other laudable aspects of this plan, power analysis was used to set the target sample size a priori to avoid any risk of p-hacking. Another strength is the data and analytic procedure. An exciting aspect of the paradigm is that it produces several measures, including actual performance, participants' bias in choosing external vs. internal memory, and metacognitive predictions, that each characterizes a separate aspect of behavior. The authors nicely outline each of these measures on p. 11 and clearly delineate the results and conclusions stemming from each one. I did have some initial concerns about how strongly the results from this particular experimental paradigm support the claim of a general bias towards external reminders (with or without metacognitive training), but the authors generally acknowledge these concerns in their Discussion. First, participants might not be motivated to set an optimal indifference point simply because there is no real incentive to earn more points, but the authors acknowledge this in lines 389-394. An additional counter-explanation, which the authors may want to also discuss, is that participants use external reminders because they don't want to "look stupid" by making errors (Hawkins, Brown, Steyvers, & Wagenmakers, 2012), even if that results in a suboptimal payoff. Second, while it's also unclear whether the present intervention is effective because of the predictions, the feedback, or their combination, the authors similarly acknowledge this on lines 378-380 (although I think it would also be worthwhile to consider the possibility that feedback alone produces the effect, which the authors do not discuss). Thus, I find the authors' Discussion section generally effective. But, there is the potential for one other major confound that was unclear to me when reading the manuscript. Lines 148-150 state that "when participants set an external reminder, target circles had to be dragged near the instructed alternative location." Am I to understand that offloading in this case involved moving the special-colored circles near or at their eventual target location? If so, I feel that this referring to this operation as just an a "external reminder" is a bit of a misnomer because putting the circle in the alternative location also helps to complete the eventual task. That could explain participants' bias to set reminders, since in doing so they are not only creating a reminder but also helping to complete the task ("pre-crastinating"; Rosenbaum et al., 2019). (Compare that to say, setting a reminder in one's phone to mow the lawn, in which case the phone alert serves as a reminder but does nothing to accomplish the lawn-mowing itself.) At the very least, the manuscript could make this issue more clear: Where, exactly, where the items moved when setting a reminder? And, if the items were indeed dragged to or near the target locations, I think the authors need to discuss how that may temper or alter their conclusions. One other weakness of the current manuscript is that, while the empirical contributions are clear, the manuscript could do more to speak to the theoretical contributions. The theoretical motivation is a single paragraph in the introduction that briefly covers both the limits of working memory and the notion of offloading with technology. More contact with broader theory on external memory and metacognition could help contextualize the results. For example, can broader theoretical perspectives on metacognitive monitoring and control suggest anything about why people might have a reminder bias? Nevertheless, I think this is a well-executed and well-analyzed study, and with some revision to the introduction and discussion, it will make a valuable contribution. References Hawkins, G. E., Brown, S. D., Steyvers, M., & Wagenmakers, E.-J. (2012). An optimal adjustment procedure to minimize experiment time in decisions with multiple alternatives. Psychonomic Bulletin & Review, 19, 339–348. Rosenbaum, D, A., et al. (2019). Sooner rather than later: Precrastination rather than procrastination. Current Directions in Psychological Science, 28, 229-223. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #2: No Reviewer #3: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. |
| Revision 1 |
|
The effect of metacognitive training on confidence and strategic reminder setting PONE-D-20-14992R1 Dear Dr. Engeler, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Veronica Yan, Ph.D. Academic Editor PLOS ONE Additional Editor Comments (optional): Thank you for the submitted revision. As you can see, the reviewers felt that their comments were sufficiently addressed, and I agree. I do want to draw attention to one note that Reviewer 3 made about the consistency of the use of terms 'sequence' and 'trial' -- this is something that you may want to pay attention to for the final proofs. Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed Reviewer #3: (No Response) ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes Reviewer #3: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes Reviewer #3: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes Reviewer #3: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes Reviewer #3: Yes ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: My comments were adequately addressed. I think this manuscript will make an interesting, incremental addition to the literature on metacognition. Reviewer #3: I was Reviewer #3 on the previous version of the manuscript. In my initial review, I expressed enthusiasm about the authors' experimental paradigm and its ability to yield multiple informative measures relevant to metacognition. I also appreciated the authors' commitment to open-science practices. My concerns had been relatively small. I had expressed some concern that, if setting reminders involved moving the special-colored circles nearer to their eventual target location, participants might doing so in order to partially solve the eventual task ("pre-crastination"), in addition to acting as a reminder per se. I also noted another explanation of the reminder bias is that participants might want to avoid "looking stupid" by making errors, even if that results in a suboptimal payout. Lastly, I felt that the authors could do more to situate the work within broader theory. I am pleased to report that the revision has addressed all of my concerns. The authors have clarified the methodological details that I and the other reviewers have asked about; in particular, the addition of a link to a demo version of the experiment is a fantastic decision that I hope to see implemented in more papers in the future. The authors also now use their Discussion section to acknowledge several potential limitations and counter-explanations that I and the other review mentioned. What I think is particularly encouraging about these revisions is that the authors have tied these other factors back into their "minimal-memory" theoretical account; for instance, pre-crastination may actually be part of the explanation as to *why* there is a reminder bias insofar as removing an intention from mind saves mental effort. That is, not only have the authors acknowledged this possibility, but have made a persuasive argument that it in fact advances their theoretical view. Fantastic. One small comment (which should not preclude acceptance of the manuscript): Based on comments from the editor, the authors have clarified that one "trial" refers to the entire sequence of 25 circles. That is fine with me; however, I notice that the alternate term "sequence" is used in several places to refer to an entire trial (e.g., lines 146, 151, 157). In their final submission, the authors may want to change these usages to "trial" as well so that a reader is not misled into thinking that there is a difference between a "trial" and a "sequence". ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No Reviewer #3: No |
Open letter on the publication of peer review reports
PLOS recognizes the benefits of transparency in the peer review process. Therefore, we enable the publication of all of the content of peer review and author responses alongside final, published articles. Reviewers remain anonymous, unless they choose to reveal their names.
We encourage other journals to join us in this initiative. We hope that our action inspires the community, including researchers, research funders, and research institutions, to recognize the benefits of published peer review reports for all parts of the research system.
Learn more at ASAPbio .