We report a study examining the role of ‘cognitive miserliness’ as a determinant of poor performance on the standard three-item Cognitive Reflection Test (CRT). The cognitive miserliness hypothesis proposes that people often respond incorrectly on CRT items because of an unwillingness to go beyond default, heuristic processing and invest time and effort in analytic, reflective processing. Our analysis (N = 391) focused on people’s response times to CRT items to determine whether predicted associations are evident between miserly thinking and the generation of incorrect, intuitive answers. Evidence indicated only a weak correlation between CRT response times and accuracy. Item-level analyses also failed to demonstrate predicted response-time differences between correct analytic and incorrect intuitive answers for two of the three CRT items. We question whether participants who give incorrect intuitive answers on the CRT can legitimately be termed cognitive misers and whether the three CRT items measure the same general construct.
Citation: Stupple EJN, Pitchford M, Ball LJ, Hunt TE, Steel R (2017) Slower is not always better: Response-time evidence clarifies the limited role of miserly information processing in the Cognitive Reflection Test. PLoS ONE 12(11): e0186404. https://doi.org/10.1371/journal.pone.0186404
Editor: Alessandro Antonietti, Universita Cattolica del Sacro Cuore, ITALY
Received: January 27, 2017; Accepted: August 16, 2017; Published: November 3, 2017
Copyright: © 2017 Stupple et al. This is an open access article distributed under the terms of the Creative Commons Attribution License, which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.
Data Availability: All relevant data are available from the figshare repository at the following DOI: 10.6084/m9.figshare.5522131.
Funding: This work was supported by the University of Derby and Lancaster University. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
Competing interests: The authors have declared that no competing interests exist.
Dual-process theories of thinking [1–4] propose the existence of a dissociation between intuitive, heuristic thought (which typically arises rapidly) and more effortful, analytic thought (which is typically deployed more slowly). The convention following Evans  is to refer to heuristic processes as ‘Type 1’ and to analytic processes as ‘Type 2’. In this paper, we use the Type 1/Type 2 and heuristic/analytic distinctions interchangeably. In testing the assumptions of dual-process theories a variety of tasks have been devised that have the potential to engender within-participant conflicts between heuristic and analytic processes. Such ‘conflict’ tasks are often associated with increased response times relative to equivalent tasks that do not engender heuristic/analytic conflicts (e.g., [5–7]). Furthermore, evidence has also clarified that the sensitivity to conflict items that is manifested in response times is often correlated with an individual’s propensity to apply analytic or calculative thinking (e.g., [8–11]).
Key examples of heuristic/analytic conflict tasks are the three items that make up the Cognitive Reflection Test (CRT ; see Table 1), which has taken the reasoning literature by storm over the past decade as a test-bed for examining dual-process theories. The CRT was devised to assess the ability of participants to resist tempting heuristic or ‘intuitive’ answers and to engage in analytic or ‘reflective’ reasoning to reach correct responses that conflict with intuitive responses. The most well-known item from the CRT is the ‘bat and ball’ problem, which reads as follows: “A bat and a ball cost $1.10 in total. The bat costs $1 more than the ball. How much does the ball cost? ____cents”. The common but incorrect answer to this problem is 10 cents, with this response being believed to be generated automatically by intuitive processes. In contrast, arriving at the correct response of 5 cents is generally assumed to require the inhibition of the intuitive response in favour of more careful and deliberative checking and analysis (see  for evidence of transcranial direct current stimulation of the dorsolateral prefrontal cortex disabling this inhibitory mechanism).
The CRT has a number of key advantages for examining reasoning performance over many other tasks, which perhaps underscore its popularity. First, there are no controversies over the appropriate normative (i.e., arithmetic) standards against which to judge performance (cf. ), unlike the situation for most other tasks in the reasoning, judgment and decision making literature (see [15,16]). Second, the CRT is a very easy test to administer to participants and is not time-consuming for participants to complete. Third, the CRT appears to be a very consistent predictor of normative responding across many other measures of judgment and choice (e.g., see ).
Overall, people tend to perform poorly on the three items of the CRT, with Pennycook, Cheyne, Koehler, and Fugelsang  noting that both web-based and typical university samples produce means in the range of 0.5 to 1 items correct out of 3, whilst students at elite universities such as Princeton and the Massachusetts Institute of Technology yield higher means in the range of 1.5 to 2 items correct. The difficulty of the CRT is consistent with the assumptions of dual-process theories in that low scores on the test suggest that rapidly-accessible, intuitive answers derived through Type 1 processing dominate responding for a majority of reasoners, with a sufficient level of Type 2 processing only being engaged for a minority of respondents so as to enable the derivation of effective solutions by means of effortful reflection.
Cognitive miserliness, rational thinking and the CRT
In examining the link between performance on the CRT and other cognitive abilities, researchers have frequently demonstrated positive correlations between CRT scores and normatively accurate performance on a wide range of tasks from the judgment and decision-making literature (e.g., [12,19–27]). Furthermore, Toplak et al.  have observed that not only does the CRT predict substantial variance in judgment and decision-making tasks that require ‘rational thinking’, but that it does so independently of measures of intelligence, executive functioning and thinking dispositions. This observation led Toplak et al.  to claim that poor scores on the CRT are an index of cognitive miserliness , that is, an unwillingness to go beyond default, heuristic processing and invest the requisite cognitive effort to solve the problem (see also ).
As Toplak et al. [27,28] note, the theme of cognitive miserliness in information processing has dominated judgment and decision-making research for over 40 years. As a case in point, Toplak et al.  refer to Kahneman and Frederick’s  discussion of ‘attribute substitution’ as a common mechanism used to lighten cognitive load. Such attribute substitution occurs when a person needs to assess attribute A but finds that assessing attribute B (correlated with attribute A) is cognitively easier such that they default to using attribute B instead. In other words, an easier question is substituted for a harder one. De Neys, Rossi, and Houdé  have explicitly evoked attribute substitution as an explanation of the bat and the ball problem on the CRT, suggesting that when participants complete tasks such as the bat and ball problem they substitute the given problem for an alternative, simpler version, in this case understanding the task as saying that “The bat costs $1” rather than what is actually stated (i.e., “The bat costs $1 more than the ball”).
In alignment with the cognitive miserliness hypothesis, De Neys et al.  demonstrated diminished confidence ratings on the bat and ball problem by those participants who gave the 10 cents response (see  for similar evidence using the bat and the ball problem and a ‘feeling of error’ measure). De Neys et al.  interpret this finding as showing that participants are not simply providing intuitive but incorrect responses to this CRT item in blissful ignorance, but rather offer the intuitive response despite having some awareness of its questionable nature and their possible misrepresentation of the problem. This failure to consider alternative responses or engage analytic thinking even in the face of reduced confidence in one’s answer does seem to be indicative of cognitive miserliness.
Thinking dispositions and the CRT
Although there is reasonable evidence triangulating on the view that poor scores on the CRT index cognitive miserliness, there are also suggestions in the literature that the case is more complex than this. Indeed, despite Toplak et al.’s  evidence that the CRT predicts rational thinking performance independent of self-reported thinking dispositions (i.e., the willingness to engage Type 2 thinking), some researchers nevertheless claim that thinking dispositions may play an important causal role in CRT success. One view (e.g., [21,32–34]), is that successful CRT performance relies, at least in part, on Actively Open-Minded Thinking (e.g., [35–37]), including the search for alternative responses. For example, Campitelli and Gerrans  present a model of CRT performance in which both Actively Open-Minded Thinking and numeric ability are key determinants of CRT success.
Another view (e.g., ) is that success on the CRT indexes an ‘analytic cognitive style’, that is, a propensity to think analytically—although it is not entirely clear to what extent this idea is conceptually distinct from Toplak et al.’s [27,28] view that success on the CRT is highly predictive of ‘rational thinking’. Arguably, however, the notion of an analytic cognitive style is more generic and encompassing than that of the somewhat narrower concept of rational thinking, which adds to its appeal. Such conceptual breadth also resonates with recent evidence showing that the CRT is predictive of a wide range of outcome measures that relate to beliefs, values and skills that are not conventionally associated with research on judgment and decision making. Such outcome measures encompass religious disbelief, paranormal disbelief, less traditional moral values, enhanced scientific understanding and reasoning, belief in evolution, improved creative problem solving, less reliance on Smartphone technology as an information source and reduced receptivity to pseudo-profound bullshit (see  for a review).
Numeracy skills, cognitive ability and the CRT
Toplak and colleagues’ emphasis on the value of the CRT in providing a measure of cognitive miserliness means that these researchers have, at times, seemed to downplay the extent to which the CRT might also assess so-called mindware gaps (e.g., the lack of the necessary cognitive rules, strategies, or belief systems to behave rationally ), with such mindware gaps being distinct from miserly information processing. For example, in their earlier research on the CRT, Toplak et al. [27,28] appear to view the CRT’s predictive power for rational thinking as being largely separable from constructs such as cognitive ability and intelligence as well as from underpinning mechanisms such as executive functioning and working memory. We note, for example, that whilst Toplak et al.  agree that mindware gaps represent an important class of reasoning error in judgment and decision-making tasks they nevertheless state that, “The potency of the CRT as a predictor of performance on heuristics-and-biases tasks certainly does not derive from its ability to assess knowledge gaps, because it clearly does no such thing” (p. 1284, emphasis added).
However, because the CRT consists of mathematics problems it would be curious if numeric ability was not important for successful performance. Indeed, we have already noted that the model presented by Campitelli and Gerrans  demonstrated that numeric ability plays a significant role in CRT performance in addition to other factors. Recent research by Sinayev and Peters  likewise claims that numeric ability seems to be a key mechanism that explains the observed association between CRT performance and normatively successful decision making, although they additionally note that the ability to detect and correct intuitions is also relevant to explaining the way in which the CRT is a predictor of effective decision making.
This emerging body of evidence for the ‘numeracy hypothesis’ in relation to the CRT suggests that mindware gaps may be more important for poor performance on the test than previously considered. For example, someone with poor numeracy skills would be unlikely to perform well on the CRT irrespective of their cognitive effort. The possession of appropriate mindware may not only be consequential for CRT success but also for its capacity to predict judgment and decision-making performance. These points have been supported by Stanovich, West, and Toplak  in their recent writing, where they directly acknowledge that CRT measures have at least moderate mindware dependence. As they note, “…even the simple bat-and-ball problem will be affected by the differential instantiation of numeracy skills. That some people find math calculations to be second nature while others do not will affect how easy the problem is” ( pp. 115–116).
Consistent with the view that cognitive ability and mindware are important in the CRT we also note Primi, Morsanyi, Chiesi, Donati, and Hamilton’s  recent proposal that the standard CRT may only be an effective measure of cognitive reflection in highly educated adults, whereas a wider range of item difficulty is needed to measure cognitive reflection in more heterogeneous samples. Thus, we suggest, the CRT may mislabel participants as cognitive misers when they struggle on the CRT when instead of being miserly they have poor levels of available working memory capacity or limited cognitive inhibition.
Thompson et al.  have also examined issues relating to cognitive ability and the CRT in a series of studies examining the possibility of priming deliberative thinking on CRT items by means of a ‘processing disfluency’ manipulation (see  ). In line with previous disfluency research (e.g., ), Thompson et al.  demonstrated that a degraded presentation of the CRT slowed participants down, which is suggestive of efforts at increased analytic thinking and the converse of cognitive miserliness. However, such slower responses were not generally associated with enhanced correct responding, with evidence in two experiments instead indicating that only certain sub-groups comprised of more cognitively able participants (as indexed by their SAT scores) were capable of benefitting from the disfluency manipulation. Thompson et al.’s  observations therefore implicate a role for cognitive ability in relation to the CRT, with higher-ability participants benefiting from the triggering of Type 2, reflective processing, although Meyer et al.  advance a more cautious position on this evidence.
Notwithstanding the uncertainty regarding the interaction between disfluency manipulations and SAT ability for enhanced performance on the CRT there are other recent lines of research that separately implicate cognitive ability metrics as being important for CRT success. For example, Stupple, Gale, and Richmond  found in two experiments that variation in working memory capacity (as measured using a composite score derived from Operation Span, Symmetry Span and Reading Span; see ) was a strong predictor of CRT performance, whereas the variation in response times to syllogistic reasoning problems that had been devised specifically to evoke a heuristic/analytic conflict was seen to be non-predictive of CRT success. Stupple et al.’s  findings run counter to a cognitive miserliness hypothesis in relation to the CRT inasmuch as increased processing times for syllogisms have previously been seen to be associated with more normative responding on these deductive arguments (e.g., [10,11]) and would therefore also be expected to be predictive of successful CRT performance.
Stupple et al.  argue that while cognitive miserliness may play a role in CRT performance, in order to solve the items correctly participants must nevertheless possess both the requisite working memory capacity and the relevant mindware, including numeracy skills. Stupple et al.  also suggest that some participants tackling the CRT might be better characterised as cognitive wastrels rather than cognitive misers because they appear to expend considerable cognitive effort engaged in a misdirected strategy that does not yield the correct response and that may not yield the incorrect intuitive response either.
In line with Stupple et al.’s  observations we also note that some researchers have recently recognised the value of scoring ‘other’ incorrect responses that are distinct from the standard ‘intuitive’ incorrect responses. For example, Pennycook et al.  derive various performance measures for the CRT, including the following: (1) CRT-Reflective, which is a participant’s total number of correct responses out of a maximum of 3; (2) CRT-Intuitive, which is a participant’s total number of incorrect intuitive responses out of a maximum of 3; and (3) CRT-PI, which is the proportion of intuitive incorrect answers out of all incorrect responses made by a participant, some of which may be non-intuitive incorrect responses (i.e., ‘other’ incorrect answers).
This latter measure has been claimed to help address statistical issues that might otherwise structurally confound CRT-Reflective and CRT-Intuitive scoring (see  for details). In their analysis, Pennycook et al.  observed that the correlation between CRT-Intuitive and a self-report measure of intuitiveness (i.e., Faith in Intuition) was not especially robust. In addition, the correlation between the CRT-PI and Faith in Intuition was not reliable as either an aggregate measure or at the item level. These results call into question the CRT as a measure of people’s tendency to rely on ‘intuitive’ responding (see  for further discussion, and , for related evidence). By implication, these findings seem to raise additional concerns about the validity of using the CRT as a measure of miserly information processing.
Aims of the experiment
In summary, we argue that while it is likely that cognitive miserliness plays a role in CRT performance (cf. ), there is a growing body of evidence that the cognitive miserliness construct offers neither a sufficient account of performance on the test nor an effective explanation of the strength of the test as a predictor of what Toplak et al. ([27,28] refer to as rational thinking, or of what Pennycook et al.  designate as an analytic cognitive style. Furthermore, we suggest that although much work has been done to examine the predictive power of the CRT as well as modeling its underpinning cognitive factors, it is nevertheless the case that, to date, a more basic analysis has tended to be neglected. More specifically, we note that the question of whether CRT items are a good index of cognitive miserliness can be examined directly through an analysis of people’s response times to the problems and the extent to which such response times are associated with solution success.
In these latter respects we assume that a fairly pure cognitive miserliness account of the CRT that acknowledges only minimal mindware dependence (e.g., relating to numeracy skills) would predict that: (1) participants who devote the shortest times to solving CRT problems should also generally be those who minimize their task engagement and opt for intuitive but incorrect answers; and (2) participants who devote the longest times to solving the CRT problems would typically be those who avoid relying on miserly, intuitive processing and instead expend effort in analytic processing in an often successful attempt to derive correct solutions. Under these assumptions the expectation would be for a robust positive correlation between the time taken on the three CRT problems and the total number of correct solutions, with such a correlation clearly indicating convergent validity with the cognitive miserliness construct.
In contrast to these predictions, if the cognitive miserliness account of the CRT is much more limited in its explanatory scope then we would expect only a weak correlation between response times and CRT success. We note that the breaking of the link between longer response times and solution success could arise for three key reasons. First, as noted in our introductory review, it is possible that people are motivated to try to solve CRT items and not simply opt for the initial intuitive response, but that they lack the requisite mindware (e.g., numerical skills) to compute the correct answer. These individuals will have relatively long response times but might either default to an intuitive response or compute an incorrect response, either way diluting the correlation between response times and CRT success. Second, some highly numerate participants might respond quickly and accurately to most or all CRT items, this time weakening the correlation because their numeracy skills allow them to calculate quickly and accurately. Third, people might engage in time-consuming analytic processing that merely rationalises an incorrect, intuitive response. Because rationalisation is analytic in its intent we would not view it as being miserly information processing, yet because such rationalisation takes time whilst leading to an incorrect, intuitive response it would again reduce the correlation between response times and CRT success. We note that there is extensive evidence for rationalisation increasing response times whilst also leading to default heuristic responding for reasoning problems such as the Wason selection task (e.g., ; but see for a more nuanced view of the evidence).
We have noted above that some participants who lack the requisite mindware to solve the CRT items might offer answers that are not consistent with either the incorrect intuitive or the correct analytic response. To investigate this issue, we aimed to conduct item level analyses for each CRT problem that contrasted the response times for the correct analytic and the incorrect intuitive responses to that problem with a third category of answer, that is, incorrect non-intuitive responses (cf. ). We hypothesised that these incorrect non-intuitive responses should again take longer than the incorrect intuitive responses.
Data were collected from 391 participants, comprising 304 undergraduates studying various degree courses at Lancaster University (220 female, 84 male), and 87 undergraduates studying for a psychology degree at the University of Derby (67 female and 16 male; 4 participants declined to provide gender information). Participants were excluded if they were familiar with the task and were replaced.
Design and materials
The three items of the Cognitive Reflection Test (Table 1) were presented to all participants via Apple Macintosh computers (Lancaster University) and PC machines (University of Derby). Responses for each item were recorded after participants had entered an answer in an on-screen field and had then clicked on a button to move to the next screen. Response times for each item reflected the total duration from its initial presentation to the submission of a response.
Ethical clearance was granted through the local ethics committees at Lancaster University and Derby University prior to commencement of data collection. We confirm that participants provided written consent and this consent procedure was approved by the local ethics committees at both Lancaster University and Derby University. Items were presented by computer one at a time in a randomized order. On completion of all of the items participants were thanked for taking part in the experiment and were given a printed debrief sheet. Any questions arising from the experiment or the debrief sheet were answered prior to participants departing the experiment.
Data were analysed using SPSS Version 24. Response time data were positively skewed and were therefore log10 transformed so as to meet the necessary requirements for parametric data analysis using regression. This regression analysis used the log10 transformed response times for each of the three separate CRT items to predict the overall CRT performance score. We further note that for the item-based comparisons of response times for the different response types the data failed to meet the assumptions for parametric analysis in terms of the absence of either normal distributions or equal group sizes. As a consequence, we analysed these data using non-parametric Kruskal Wallace tests and applied Bonferroni-adjusted Mann-Whitney follow-up tests.
The first analysis examined the association between the sum of participants’ response times (log10 transformed to correct for violations of normality) for the three CRT items and the total number of items responded to correctly (i.e., the CRT-Reflective measure described by Pennycook et al ). The correlation was positive and significant, r = .181, p < .001 (N = 391), as predicted by the cognitive miserliness account, but the relationship was relatively small, indicating that longer response times were only weakly associated with better CRT performance. Note that the correlation between participants’ mean response times for the three CRT items and the total number of items responded to correctly was equivalent to the one reported, which involved total response times for the three CRT items. This relatively weak relationship between correct performance and response times is readily apparent when CRT-Reflective scores and total response-time data are depicted graphically as a dot plot (see Fig 1).
As a follow-up analysis a multiple regression was conducted to test the contribution that log10 transformed response times for each of the CRT items made to predicting overall correct responding for the three problems (i.e., the CRT-Reflective score). The regression model was highly significant (see Table 2), but standardised regression coefficients for each problem type showed differing relationships with overall correct responding. For the bat and ball item, longer response times were associated with increased overall performance, indicating that increased response times to this item predicted correct responding overall. However, response times to the widget problem did not significantly predict correct responses. Finally, the response times for the lily pads problem indicated that reduced response times were significantly associated with more correct responses overall.
Pennycook et al.  propose that to examine miserly thinking in the CRT it is best to calculate and examine intuitiveness scores (see Fig 2 for a dot plot depicting the relationship between CRT-Intuitive scores and total response times). A further regression was conducted that was structurally equivalent to that reported above, but which used Pennycook et al.’s CRT-Intuitive score as the dependent variable, which is a participant’s total number of incorrect intuitive responses out of a maximum of 3. This model was again highly significant (see Table 3). For the bat and ball item, shorter response times were associated with increased intuitive responding overall. However, response times to the widget problem, and response times for the lily pads problem, did not significantly predict intuitive responding.
The readily apparent asymmetry between the regression models for the CRT-Reflective score and the CRT-Intuitive score suggest that incorrect but non-intuitive responses may be very important in explaining the relationship between response times and performance on the CRT. To examine this issue further, we examined the response times for each of the three CRT problems separately, having first designated a participant’s response to the item as being one of three possible types: (1) correct analytic; (2) incorrect intuitive; or (3) incorrect non-intuitive. Due to uneven sample sizes across groups non-parametric ANOVAs were conducted. Median response times for each type of answer for the CRT items are shown in Table 4, whilst Table 5 shows the frequency of each type of response across the CRT items.
For the bat and ball problem incorrect intuitive answers were the most common (259 responses) and incorrect non-intuitive answers were the least common (just 15 responses). In terms of the response-time analysis, this demonstrated a highly significant difference across response types, Kruskal Wallace H = 49.99, df = 2, p < .001. Post hoc Mann Whitney tests (Bonferroni adjusted alpha = .017) indicated that the difference in response-times for correct analytic and incorrect intuitive responses was highly significant and in line with cognitive miser predictions (p < .001). Mann Whitney tests showed that there was no significant difference either between correct analytic and incorrect non-intuitive responses (p = .09) or between incorrect intuitive and incorrect non-intuitive responses (p = .86), although these analyses were potentially compromised by the small number of responses in the incorrect non-intuitive category. In sum, the data for the bat and ball problem indicate that correct analytic responses took longer to derive than incorrect intuitive responses (cf. Fig 3), which aligns with predictions relating to cognitive miserliness dominating responding on this CRT item given that the incorrect intuitive answers were also more frequent.
For the widget problem incorrect intuitive answers were again the most common (209 responses) and incorrect non-intuitive answers were the least common (63 responses), although proportionally the latter were more prevalent for the widget problem than for other CRT items, suggesting that with this problem a fairly sizeable subset of participants were attempting to derive an analytic solution rather than defaulting to an intuitive one, but were nevertheless unable to derive a correct response. The response-time analysis for the widget problem demonstrated a significant difference across response types, Kruskal Wallace H = 7.25, df = 2, p = .027 (see also Fig 4). However, post hoc Mann Whitney tests (Bonferroni adjusted threshold alpha = .017) showed that this difference was not significant between correct analytic and incorrect intuitive responses (p = .09), which runs counter to cognitive miserliness predictions. There was also no difference in response times between correct analytic and incorrect non-intuitive responses (p = .29), although incorrect non-intuitive responses did take reliably longer than incorrect intuitive responses (p = .013).
For the lily pads problem, incorrect intuitive answers were again the most common (179 responses), although correct analytic responses were fairly frequent (163 responses) and combined with incorrect non-intuitive responses (49 responses) indicated that overall a majority of participants were attempting to derive analytic solutions for this item. Other researchers have likewise observed that the lily pads problem tends to be evoke more analytic responding that the other two items and is somewhat easier (e.g., [18,33]). The response-time analysis for the lily pads problem again revealed a significant difference across response types, Kruskal Wallace H = 31.61, df = 2, p < .001 (see also Fig 5). Post hoc Mann Whitney tests (Bonferroni adjusted threshold alpha = .017) demonstrated that this difference was not significant between correct analytic and incorrect intuitive responses (p = .25). However, incorrect non-intuitive responses took reliably longer than both incorrect intuitive responses (p < .001) and correct analytic responses (p < .001).
Finally, we examined the internal reliability of the CRT. The items showed an 'unacceptable' level of reliability, Cronbach’s ά = .47.
The results of this study reveal a weak but reliable correlation between CRT response times and overall CRT accuracy. Although this evidence does not support a pure cognitive miserliness account of CRT performance, it does, nevertheless, suggest that miserly information processing is part of an explanation of solution errors on the CRT. If cognitive miserliness was the primary explanation of CRT performance then one would expect a stronger pattern of positive association between CRT response times and accuracy rates, with correct responses taking time to be produced via analytic, Type 2 processing, and with incorrect intuitive responses arising quickly via heuristic, Type 1 processing. Instead, our data suggest that several factors are at play when participants complete the CRT.
Further evidence against a pure cognitive miserliness hypothesis emerged from examining individual CRT items. Regression analyses showed that although response times for the bat and ball problem predicted overall successful CRT performance and were therefore consistent with a cognitive miser account, this was not so for the widget problem. Most notably, faster responses to the lily pads problem were predictive of better overall CRT performance, in opposition to what might be expected from a cognitive miser account. Furthermore, the item-level analyses contrasting response times across response types (incorrect intuitive, incorrect non-intuitive and correct analytic) also failed to demonstrate predicted differences between correct analytic responses and incorrect intuitive responses for two of the three CRT problems, with the exception again being the bat and ball problem.
Because these latter comparisons are not contaminated by responses times arising when people respond non-intuitively but incorrectly, they represent a more clear-cut test of the cognitive miserliness hypothesis than the weak correlation between response times and accuracy rates. The observation that the predicted response-time difference between correct analytic responses and incorrect intuitive responses was only evident for the bat and ball problem raises questions about the validity of a pure cognitive miserliness explanation of poor CRT performance. We note that a recent study by Travers, Rolinson, and Feeney , which examined response times using an adapted version of Primi et al.’s  eight-item CRT, found that participants took significantly longer to evaluate correct analytic answers relative to incorrect intuitive ones when response-time data were analyzed at an aggregate level. However, since an item-level analysis was not presented for this comparison it is uncertain to what extent the observed effect was driven by differences arising for perhaps just a few items, such as the bat and ball problem, as in the present study.
Travers et al.’s  study presents a fascinating and detailed analysis of responding on conflict and non-conflict versions of the CRT using a mouse-tracking methodology to capture the time-course of processing. We note, however, that there are some non-standard aspects of their mouse-tracking paradigm, which may render it difficult to draw clear-cut comparisons with the standard presentation of the CRT, including the presentation of four response options per problem. There are also more general concerns with what mouse-tracking data can legitimately reveal regarding reasoning processes, which arise because of inherent methodological artifacts that seriously bias response-time effects in ways that ultimately confound theoretical interpretation, even producing effects that are opposite to these found using more sensitive attentional measures such as eye-tracking analysis (for relevant discussion see [51,54–56].
Our data not only question whether participants who give incorrect intuitive answers to the CRT can simply be referred to as cognitive misers, but also indicate that a sizeable minority of incorrect responders (especially on the widget and lily pads problems) did not produce predicted intuitive answers. This observation again suggests that such individuals cannot be labeled as cognitive misers since they are presumably engaging in analytic thinking, albeit analytic thinking that is not delivering a correct solution. Interestingly, participants who generated the incorrect non-intuitive responses for the widget and lily pads problems also demonstrated the longest response times for these items and may be best categorized as cognitive wastrels , since they seem to be trying to compute a response whilst not possessing sufficient cognitive abilities—numeric or otherwise—to derive an effective solution.
In this latter respect the present evidence suggests that some aspects of poor CRT performance derive from a failure to compute the normative response because of mindware gaps (e.g., in basic mathematical skills), which means that it is possible that the CRT also has the capacity to assess such gaps on various other judgment and decision making tasks employed in heuristics-and-biases research (e.g., tasks also involving fundamental numerical skills such as ratio-bias problems; e.g., ). Indeed, we again note Sinayev and Peters’  evidence that numeric ability is one of the key factors that explains the standard association observed between CRT performance and successful responses on judgment and decision-making tasks.
Returning to our finding that two CRT items failed to engender predicted response-time differences between correct analytic responses and incorrect intuitive responses—as expected by a cognitive miserliness account—we suggest that at the very least these data raise concerns about the internal reliability of the CRT in terms of whether the items measure the same general construct. The Cronbach's alpha we observed was at an unacceptable level and not dissimilar to that seen in previous studies (e.g., Teovanović and colleagues , reported a Cronbach’s ά of .39). Indeed, Toplak et al.  have themselves commented that from the standpoint of reliability “three items is obviously too few” (p. 150), which has, in part, inspired them to validate a seven-item version of the CRT. We also note Teovanović et al.’s  research examining the factor structure of a range of cognitive biases, which found that while the standard three-item CRT correlated well with some of the individual tasks it did not load significantly onto any factors identified as explaining variance in these cognitive biases. Teovanović et al. argue that this outcome might be because of a lack of internal consistency between CRT items.
If different CRT items are measuring different cognitive and dispositional constructs, then this has major implications for how the CRT is interpreted and utilised and warrants further investigation. In this respect, we applaud the development of extended versions of the CRT ([28,43,59]) given that these have a broader focus, increase the range of possible scores, reduce the confounding influence of numeracy and include items that some less cognitively able individuals can solve. Our own data, which indicate that some participants may have engaged in misdirected analytic thinking on standard CRT items, suggest that the efforts by Primi et al.  to measure reflective thinking in less able participants are especially welcome.
In terms of the internal reliability of the standard three-item CRT, we suggest that the bat and ball problem appears to be the most convincing candidate for measuring cognitive miserliness, with the other two items appearing not to measure cognitive miserliness at all. Certainly, for the lily pads problem there is little or no calculation that needs to be done to produce the correct solution. The lily pads doubling in area each day simply requires the realization that the lily pads would be at half their final extent the day before the lake is completely covered, which contrasts markedly with the bat and ball problem, where realizing that 10 cents is not the correct answer requires the application of some (albeit basic) algebra, whereby candidate values for the ball are tested. We would argue that this is an example of the kind of sustained ‘decoupling’ that Stanovich and Toplak  describe as a necessary criterion for Type 2 processing. It is also noteworthy in relation to the lily pads problem that if participants do engage in a misdirected calculative strategy from the outset then they often invest a substantial amount of time in reaching an incorrect answer. In this respect, the lily pads problem may best be categorized as an ‘insight’ problem that hinges on a restructured representation of the given information. Indeed, an earlier version of the problem studied by Schooler, Ohlsson, and Brooks  was described in such terms.
A final question to address given our concerns with the reliability of CRT items as a measure of cognitive miserliness relates to the matter of what participants are doing when they take a median of around 19 to 25 seconds on these items only thence to generate an incorrect intuitive response. We note that this important question is not directly addressed by Travers et al.  in relation to their findings regarding the time-course of processing on the CRT. This is despite their observation that the average time to provide the incorrect intuitive response to CRT items involving a heuristic/analytic conflict was a lengthy 21 seconds, which is closely aligned with the median response time observed in the present study for incorrect intuitive answers. These relatively slow responses seem to support aspects of De Neys et al.’s (e.g., ) theoretical perspective, which proposes that although participants may be unconvinced by their initial, intuitive answer to a CRT item such that analytic Type 2 reasoning might be triggered, they may also lack the requisite mindware to solve the problem and still default to an intuitive response without engaging Type 2 analytic thinking, albeit with a sense of doubt as to the validity of their answer. Presumably, this uncertainty takes time to be resolved, thereby extending response times even though the eventual answer that is proffered in the incorrect intuitive one. It is also possible that these participants do strive to resolve the uncertainly through analytic processing (which takes time), but because of the inherent processing demands of this analytic thinking the intuitive response is defaulted to as a form of ‘computational escape hatch’ [62–64]. In explaining the long response times for incorrect intuitive responses, we also reiterate here the possibility that people might engage in time-consuming analytic processing that ends up merely rationalising the intuitive response.
We note that all of these explanations for relatively long processing times for incorrect intuitive answers on the CRT arise through analytic, Type 2 processes and not because of miserly information processing. More generally, it is important to note that the relationship between responses and response times is not always straightforward and that ‘slowness’ has been argued to be an incidental correlate of Type 2 processing rather than a necessary criterion [3,60]. However, it is difficult to argue that when a participant is investing more time in responding to an item they are being miserly, even if they are attempting to come up with a post hoc rationale for an erroneous response. Similarly, it is difficult to see how responses derived from autonomous processes might be consistently slower than those derived from the sustained decoupling of problems from hypothetical solutions. It should also be noted that the dual-process explanation of the CRT, while dominant in the literature, is not universally endorsed (see [65,66], for critical perspectives on dual-process theory).
In conclusion, we contend that although the CRT is widely assumed to be an effective measure of cognitive miserliness, our findings indicate that it is prudent not to make this assumption uncritically. Indeed, we see grounds for viewing two of the three items used on the CRT as being poor measures of cognitive miserliness, which may undermine the internal reliability of the CRT as a whole. We would suggest that researchers using the CRT to measure miserliness would do well to find some approach to eliminate spurious variance arising from responses that are neither correct analytic answers nor incorrect intuitive answers, since in many cases these incorrect non-intuitive responses are not miserly. In this respect, we welcome the emergence of alternative scoring approaches in the CRT literature [18,41]. Moreover, we consider it to be judicious for researchers to continue to use measures of cognitive disposition such as the Need for Cognition Scale , the Rational Experiential Inventory  and the Actively Open Minded Thinking Scale [36,37], since these self-report measures afford useful insights into how participants experience their own cognitive processes. Without such measures we may conflate cognitive wastrels, whose performance could be ameliorated through training, with genuine cognitive misers, who elect not to be analytic and who may not be amenable to training interventions.
- 1. De Neys W. Bias and conflict: A case for logical intuitions. Perspect Psychol Sci. 2012;7: 28–38. pmid:26168420
- 2. Evans JStBT. Thinking twice: Two minds in one brain. Oxford: Oxford University Press; 2010.
- 3. Evans JStBT, Stanovich KE. Dual-process theories of higher cognition: Advancing the debate. Perspect Psychol Sci. 2013;8: 223–241. pmid:26172965
- 4. Kahneman D. Thinking Fast and Slow London: Penguin. 2011;
- 5. Bonner C, Newell BR. In conflict with ourselves? An investigation of heuristic and analytic processes in decision making. Mem Cognit. 2010;38: 186–196. pmid:20173191
- 6. Stupple EJN, Ball LJ. Belief–logic conflict resolution in syllogistic reasoning: Inspection-time evidence for a parallel-process model. Think Reason. 2008;14: 168–181.
- 7. Thompson VA, Striemer CL, Reikoff R, Gunter RW, Campbell JI. Syllogistic reasoning time: Disconfirmation disconfirmed. Psychon Bull Rev. 2003;10: 184–189. pmid:12747506
- 8. De Neys W, Glumicic T. Conflict monitoring in dual process theories of thinking. Cognition. 2008;106: 1248–1299. pmid:17631876
- 9. Rand DG, Greene JD, Nowak MA. Spontaneous giving and calculated greed. Nature. 2012;489: 427–430. pmid:22996558
- 10. Stupple EJN, Ball LJ, Ellis D. Matching bias in syllogistic reasoning: Evidence for a dual-process account from response times and confidence ratings. Think Reason. 2013;19: 54–77.
- 11. Stupple EJN, Ball LJ, Evans JStBT, Kamal-Smith E. When logic and belief collide: Individual differences in reasoning times support a selective processing model. J Cogn Psychol. 2011;23: 931–941.
- 12. Frederick S. Cognitive reflection and decision making. J Econ Perspect. 2005;19: 25–42.
- 13. Oldrati V, Patricelli J, Colombo B, Antonietti A. The role of dorsolateral prefrontal cortex in inhibition mechanism: A study on cognitive reflection test and similar tasks through neuromodulation. Neuropsychologia. 2016;91: 499–508. pmid:27647553
- 14. Bourgeois-Gironde S, Van Der Henst J-B. How to open the door to System 2: Debiasing the bat-and-ball problem. In Watanabe S, Bloisdell AP, Huber L, & Young A, editors. Rational animals, irrational humans. Tokyo: Keio University Press. 2009.
- 15. Elqayam S, Evans JStBT. Subtracting “ought” from “is”: Descriptivism versus normativism in the study of human thinking. Behav Brain Sci. 2011;34: 233–248. pmid:22000212
- 16. Stupple EJN, Ball LJ. The intersection between Descriptivism and Meliorism in reasoning research: further proposals in support of ‘soft normativism’. Front Psychol. 2014;5.
- 17. Weber EU, Johnson EJ. Mindful judgment and decision making. Annu Rev Psychol. 2009;60: 53–85. pmid:18798706
- 18. Pennycook G, Cheyne JA, Koehler DJ, Fugelsang JA. Is the cognitive reflection test a measure of both reflection and intuition? Behav Res Methods. 2016;48: 341–348. pmid:25740762
- 19. Baldi P. L., Iannello P., Riva S., & Antonietti A. Socially biased decisions are associated to individual differences in cognitive reflection. Stud Psychol. 55: 265–271.
- 20. Calvillo DP, Burgeno JN. Cognitive reflection predicts the acceptance of unfair ultimatum game offers. Judgm Decis Mak. 2015;10: 332.
- 21. Campitelli G, Labollita M. Correlations of cognitive reflection with judgments and choices. Judgm Decis Mak. 2010;5: 182.
- 22. Cokely ET, Kelley CM. Cognitive abilities and superior decision making under risk: A protocol analysis and process model evaluation. Judgm Decis Mak. 2009;4: 20.
- 23. Hoppe EI, Kusterer DJ. Behavioral biases and cognitive reflection. Econ Lett. 2011;110: 97–100.
- 24. Koehler DJ, James G. Probability matching and strategy availability. Mem Cognit. 2010;38: 667–676. pmid:20852231
- 25. Oechssler J, Roider A, Schmitz PW. Cognitive abilities and behavioral biases. J Econ Behav Organ. 2009;72: 147–152.
- 26. Simonovic B, Stupple EJN, Gale M, Sheffield D. Stress and Risky Decision Making: Cognitive Reflection, Emotional Learning or Both. J Behav Decis Mak. 2017;30: 658–665.
- 27. Toplak ME, West RF, Stanovich KE. The Cognitive Reflection Test as a predictor of performance on heuristics-and-biases tasks. Mem Cognit. 2011;39: 1275. pmid:21541821
- 28. Toplak ME, West RF, Stanovich KE. Assessing miserly information processing: An expansion of the Cognitive Reflection Test. Think Reason. 2014;20: 147–168.
- 29. Kahneman D, Frederick S. Representativeness revisited: Attribute substitution in intuitive judgment. Heuristics Biases Psychol Intuitive Judgm. 2002;49: 49–81.
- 30. De Neys W, Rossi S, Houdé O. Bats, balls, and substitution sensitivity: Cognitive misers are no happy fools. Psychon Bull Rev. 2013;20: 269–273. pmid:23417270
- 31. Gangemi A, Bourgeois-Gironde S, Mancini F. Feelings of error in reasoning—in search of a phenomenon. Think Reason. 2015;21: 383–396.
- 32. Baron J, Scott S, Fincher K, Metz SE. Why does the cognitive reflection test (sometimes) predict utilitarian moral judgment (and other things)? J Appl Res Mem Cogn. 2015;4: 265–284.
- 33. Campitelli G, Gerrans P. Does the cognitive reflection test measure cognitive reflection? A mathematical modeling approach. Mem Cognit. 2014;42: 434–447. pmid:24132723
- 34. Liberali JM, Reyna VF, Furlan S, Stein LM, Pardo ST. Individual differences in numeracy and cognitive reflection, with implications for biases and fallacies in probability judgment. J Behav Decis Mak. 2012;25: 361–381. pmid:23878413
- 35. Baron J. Thinking and deciding 4th ed. New York: Cambridge University Press; 2008.
- 36. Stanovich KE, West RF. Reasoning independently of prior belief and individual differences in actively open-minded thinking. J Educ Psychol. 1997;89: 342.
- 37. Stanovich KE, West RF. Natural myside bias is independent of cognitive ability. Think Reason. 2007;13: 225–247.
- 38. Pennycook G, Ross RM. Commentary: Cognitive reflection vs. calculation in decision making. Front Psychol. 2016;7.
- 39. Pennycook G, Fugelsang JA, Koehler DJ. Everyday consequences of analytic thinking. Curr Dir Psychol Sci. 2015;24: 425–432.
- 40. Perkins D. Outsmarting IQ: The emerging science of learnable intelligence. Simon and Schuster; 1995.
- 41. Sinayev A, Peters E. Cognitive reflection vs. calculation in decision making. Front Psychol. 2015;6.
- 42. Stanovich KE, West RF, Toplak ME. The rationality quotient: toward a test of rational thinking. MIT Press; 2016.
- 43. Primi C, Morsanyi K, Chiesi F, Donati MA, Hamilton J. The development and testing of a new version of the cognitive reflection test applying item response theory (IRT). J Behav Decis Mak. 2016;29: 453–469.
- 44. Thompson VA, Turner JAP, Pennycook G, Ball LJ, Brack H, Ophir Y, et al. The role of answer fluency and perceptual fluency as metacognitive cues for initiating analytic thinking. Cognition. 2013;128: 237–251. pmid:23158572
- 45. Alter AL, Oppenheimer DM, Epley N, Eyre RN. Overcoming intuition: metacognitive difficulty activates analytic reasoning. J Exp Psychol Gen. 2007;136: 569. pmid:17999571
- 46. Thompson VA, Turner JAP, Pennycook G. Intuition, reason, and metacognition. Cognit Psychol. 2011;63: 107–140. pmid:21798215
- 47. Meyer A, Frederick S, Burnham TC, Guevara Pinto JD, Boyer TW, Ball LJ, et al. Disfluent fonts don’t help people solve math problems. J Exp Psychol Gen. 2015;144: e16. pmid:25844628
- 48. Stupple EJN, Gale M, Richmond C. Working Memory, Cognitive Miserliness and Logic as Predictors of Performance on the Cognitive Reflection Test. In Knauff M, Pauen M, Sebanz N, & Wachsmuth I, editors. Proceedings of the 35th Annual Conference of the Cognitive Science Society. Austin, TX: Cognitive Science Society; 2013.
- 49. Unsworth N, Heitz RP, Schrock JC, Engle RW. An automated version of the operation span task. Behav Res Methods. 2005;37: 498–505. pmid:16405146
- 50. Erceg N, Bubić A. One test, five scoring procedures: different ways of approaching the cognitive reflection test. J Cogn Psychol. 2017;29: 381–392.
- 51. Ball LJ, Lucas EJ, Miles JNV, Gale AG. Inspection times and the selection task: What do eye-movements reveal about relevance effects? Q J Exp Psychol Sect A. 2003;56: 1053–1077.
- 52. Evans JStBT, Ball LJ. Do people reason on the Wason selection task? A new look at the data of Ball et al.(2003). Q J Exp Psychol. 2010;63: 434–441.
- 53. Travers E, Rolison JJ, Feeney A. The time course of conflict on the Cognitive Reflection Test. Cognition. 2016;150: 109–118. pmid:26896724
- 54. Roberts MJ. Inspection times and the selection task: Are they relevant? Q J Exp Psychol Sect A. 1998;51: 781–810.
- 55. Roberts MJ. How should relevance be defined? What does inspection time measure? A reply to Evans. Q J Exp Psychol Sect A. 1998;51: 815–817.
- 56. Roberts MJ, Newton EJ. Inspection times, the change task, and the rapid-response selection task. Q J Exp Psychol Sect A. 2001;54: 1031–1048.
- 57. Mevel K, Poirel N, Rossi S, Cassotti M, Simon G, Houdé O, et al. Bias detection: Response confidence evidence for conflict sensitivity in the ratio bias task. J Cogn Psychol. 2015;27: 227–237.
- 58. Teovanović P, Knežević G, Stankov L. Individual differences in cognitive biases: Evidence against one-factor theory of rationality. Intelligence. 2015;50: 75–86.
- 59. Thomson KS, Oppenheimer DM. Investigating an alternate form of the cognitive reflection test. Judgm Decis Mak. 2016;11: 99.
- 60. Stanovich KE, Toplak ME. Defining features versus incidental correlates of Type 1 and Type 2 processing. Mind Soc. 2012;11: 3–13.
- 61. Schooler JW, Ohlsson S, Brooks K. Thoughts beyond words: When language overshadows insight. J Exp Psychol Gen. 1993;122: 166.
- 62. Ball LJ, Quayle JD. Alternative task construals, computational escape hatches, and dual-system theories of reasoning. Behav Brain Sci. 2000;23: 667–668.
- 63. Quayle JD, Ball LJ. Working memory, metacognitive uncertainty, and belief bias in syllogistic reasoning. Q J Exp Psychol Sect A. 2000;53: 1202–1223.
- 64. Stanovich KE, West RF. Individual differences in reasoning: Implications for the rationality debate? Behav Brain Sci. 2000;23: 645–665. pmid:11301544
- 65. Keren G, Schul Y. Two is not always better than one: A critical evaluation of two-system theories. Perspect Psychol Sci. 2009;4: 533–550. pmid:26161732
- 66. Osman M. An evaluation of dual-process theories of reasoning. Psychon Bull Rev. 2004;11: 988–1010. pmid:15875969
- 67. Cacioppo JT, Petty RE, Kao CF, Rodriguez R. Central and peripheral routes to persuasion: An individual difference perspective. J Pers Soc Psychol. 1986;51: 1032.
- 68. Epstein S. Integration of the cognitive and the psychodynamic unconscious. Am Psychol. 1994;49: 709. pmid:8092614