- 1School of Psychology, University of East London, London, United Kingdom
- 2Leeds University Business School, Leeds, United Kingdom
- 3Department of High Performance Computing, Simula Research Laboratory, Oslo, Norway
- 4Leeds University Business School, Leeds, United Kingdom
The attempts to mitigate the unprecedented health, economic, and social disruptions caused by the COVID-19 pandemic are largely dependent on establishing compliance to behavioral guidelines and rules that reduce the risk of infection. Here, by conducting an online survey that tested participants’ knowledge about the disease and measured demographic, attitudinal, and cognitive variables, we identify predictors of self-reported social distancing and hygiene behavior. To investigate the cognitive processes underlying health-prevention behavior in the pandemic, we co-opted the dual-process model of thinking to measure participants’ propensities for automatic and intuitive thinking vs. controlled and reflective thinking. Self-reports of 17 precautionary behaviors, including regular hand washing, social distancing, and wearing a face mask, served as a dependent measure. The results of hierarchical regressions showed that age, risk-taking propensity, and concern about the pandemic predicted adoption of precautionary behavior. Variance in cognitive processes also predicted precautionary behavior: participants with higher scores for controlled thinking (measured with the Cognitive Reflection Test) reported less adherence to specific guidelines, as did respondents with a poor understanding of the infection and transmission mechanism of the COVID-19 virus. The predictive power of this model was comparable to an approach (Theory of Planned Behavior) based on attitudes to health behavior. Given these results, we propose the inclusion of measures of cognitive reflection and mental model variables in predictive models of compliance, and future studies of precautionary behavior to establish how cognitive variables are linked with people’s information processing and social norms.
Introduction
Behavioral Measures to Control COVID-19
Countries world-wide are currently considering how to guide and change people’s behavior in order to maintain or ease COVID-related measures such as social distancing, increased hand washing, self-isolation, etc. These behavioral guidelines and the degree of their uptake are important to reduce the spread of the disease, prevent potentially very costly recurring waves of infections (Lauerman, 2020), and indeed mitigate likely future epidemics (or pandemics). Although compliance is not uniform, little is known about the psychosocial determinants of compliance with COVID guidelines (Bogg and Milad, 2020), and the current advice provided from scientists to policy-makers is based on general principles from pre-pandemic behavioral research (Bavel et al., 2020). Epidemiologists admit the lack of much-needed knowledge about the heterogeneity of behavioral responses (Weston et al., 2018). The famous Imperial College model (Ferguson et al., 2020), which altered the United Kingdom’s strategy, assumed 25% non-compliance on social distancing for people aged over 70, apparently without any specific empirical basis.
Nevertheless, studies conducted both before (Keizer et al., 2019) and during (Xie et al., 2020) the COVID pandemic have shown that various factors influence compliance with officially recommended health measures, which in turn should increase prevention success, including cognitive ability or disposition to pay attention, understand, memorize, or enact official guidelines. Thus far, however, no single study has investigated a comprehensive range of COVID-related precautionary behaviors and their dependence on multiple cognitive factors (see Xie et al. for the effect of working memory). It is possible that, when measured at a granular level (e.g., use of face masks and tracing apps), other cognitive factors may predict compliance with COVID-19-related precautionary measures. Consequently, the use of more fine-grained cognitive-behavioral predictions should enable better adherence estimates and allow adjustments of policies and guidelines (Anderson et al., 2020; Webster et al., 2020).
The current research investigates three specific cognitive variables – cognitive failures, cognitive reflection, and thinking disposition – and their potential role in precautionary measures during the COVID pandemic. These variables – together with knowledge about the new disease – were chosen because they relate to an important and often referred to theoretical framework, the so-called dual-processing theories (see, for reviews, Kahneman, 2011; Evans and Stanovich, 2013). The dual-processing theories propose that human judgment and decision behavior is driven by automatic and unconscious mental processes as well as by controlled and reflective thinking. While dual-processing theories seek to account for human thinking performance, another well-established theory offers potential for understanding and predicting COVID-19-related behaviors based on attitudinal differences is the Theory of Planned Behavior (TPB; Ajzen, 1991).
Theory of Planned Behavior
The TPB has been applied to an extensive range of health-related behaviors (Armitage and Conner, 2001) and is the most influential social cognition model for predicting and explaining health behavior. TPB stipulates that a person’s behavioral, social, and control beliefs affect the intention for behavior change. For example, people who think that one cannot transmit the disease in the absence of observable symptoms will have behavioral beliefs (“Will this be effective?”), social beliefs (“Are others doing it?”), and control beliefs (“Am I able to do this?”), which make it less likely for them to adopt prevention measures.
TPB predicts an impressive 30–40% of the variance in health (prevention) behaviors (Armitage and Conner, 2001). TPB-related variables (attitudes and norms) can in principle be applied to COVID-19-related behavior (Sætrevik, 2020). Indeed, recent, but pre-COVID, research found that, in a Chinese sample, social norms, perceived behavioral control, and attitudes all predicted willingness to socially isolate in the face of a pandemic (Zhang et al., 2019).
Dual-Process Theories of Thinking and Decision-Making
TPB notably assumes that attitudes and beliefs about actions are explicit, that is, they are given as a considered reflective account. However, within cognitive psychology, the dual-process theories of thinking and decision-making have become influential. They propose the workings in the mind consisting of both explicit reasoning and qualitatively different implicit judgment processes. The latter “Type 1” processing is thought to be fast, intuitive, and automatic, relying on heuristics (i.e., mental shortcuts) or “gut feelings” (Tversky and Kahneman, 1974); while the former explicit “Type 2” processing is considered to be slow, reflective, and effortful (Evans, 2003, 2010), encompassing logical and rational reasoning (Evans, 2008). Thus, unlike Type 2 processing, intuitive Type 1 processes are considered to be not under conscious cognitive control (Lowe Bryan and Harter, 1899; Shiffrin and Schneider, 1977; Evans, 2008), although the outputs from Type 1 thinking may or may not get overturned by conscious Type 2 processing (Kahneman and Frederick, 2002, 2007). While there are some critics of this notion (Gigerenzer and Regier, 1996; Osman, 2004; Keren and Schul, 2009; Kruglanski and Gigerenzer, 2011), the distinction between so-called Type 1 and Type 2 processes is supported by considerable empirical evidence (Evans, 1977; Evans et al., 1983; Klauer et al., 2000; Evans and Stanovich, 2013; Stanovich et al., 2019).
A more recent theory of dual-processing proposes a tripartite model that specifies two layers responsible for Type 2 processing: (1) the “algorithmic mind” and (2) the “reflective mind” (Stanovich, 2009). The performance of the algorithmic mind can be specified as the ability to override intuitive Type 1 responses and to respond with the correct analytical Type 2 responses (Toplak et al., 2011; see also Kahneman and Frederick, 2002, 2005). Its operations, therefore, should be related to attentional processes as well as mental simulation abilities (being able to separate and manipulate mental representational content, Stanovich, 2012). For example, the oft-used Cognitive Reflection Test (CRT, Frederick, 2005) presents a small series of brief math puzzles. Each of these CRT items evidently prompts an intuitively obvious – but incorrect – Type 1 answer; but, when applying reflective thinking, people are more likely to inhibit this first thought and produce the correct Type 2 answer by using basic algorithmic thinking. The CRT is thought to largely reflect the algorithmic layer processing in different ways (Stanovich, 2012): (1) it inhibits and overrides Type 1 (autonomous) processes and (2) it generates the correct answers by being able to symbolically manipulate representations (for which it needs attentional and working memory processing). For our dependent variable then, people may need to inhibit the automatic responses (i.e., it is easier not washing your hands so often, not wearing face masks, and not keeping extra distance). Furthermore, people may also need extra attentional and working memory processes (to remind oneself to wash one’s hands when coming into the house or when having touched surfaces, to memorize to stock and then find anti-bacterial gel, etc.). In fact, for some people (e.g., of older age), even the operation of different types of face masks may require instructions and significant efforts (Lee et al., 2020). Although the CRT is thought to be associated with a range of cognitive constructs (e.g., Toplak et al., 2011), including thinking dispositions and numeracy (e.g., Cokely and Kelley, 2009; Campitelli and Gerrans, 2014), recent evidence shows that working memory is the strongest single predictor of CRT performance (Stupple et al., 2017; Gray and Holyoak, 2020).
The reflective mind is the second layer within Type 2 processes and comprises higher-level cognitive styles, thinking dispositions, and metacognitive beliefs (Stanovich, 2011), which explain additional variance in thinking performance beyond the workings of the algorithmic mind (Stanovich and West, 2008; Stanovich, 2012). The reflective mind is responsible for the degree to which one thinks extensively about problems before responding, the amount of information one collects before making decisions, whether one integrates others’ points of view into one’s decisions or whether one adjusts beliefs according to the quality of the evidence (Baron, 2008). High actively open-minded thinking (AOT, Stanovich and West, 1998) scores have been shown to have a positive correlation with performance in the CRT (Baron et al., 2015) and belief bias syllogistic reasoning tasks (Macpherson and Stanovich, 2007). Thus, two people may have the same level of cognitive ability, but one may be more inclined than the other to engage their algorithmic mind because of their disposition to open-mindedly employ reflective thinking by taking in new information and be prepared to change their judgments based on it – a property of the reflective mind.
Based on the dual-process framework, and, in particular, Stanovich’s tripartite model, we hypothesized that people with higher cognitive reflection tendency (AOT) and ability (the algorithmic-level processing, measured with CRT) will engage more in thinking about, and therefore be more likely to employ, precautionary measures than people with lower cognitive reflection tendencies. Adopting new tasks, or performing them in a new context or with greater frequency (such as remembering to wash hands frequently and putting on face masks) should tax cognitive resources linked with the algorithmic mind, such as inhibition, attention, and working memory capacity (Stanovich, 2012). Indeed, cognitive reflection (measured with the CRT) has been shown to correlate positively with the ability to inhibit impulsive actions (Oechssler et al., 2009; Jimenez et al., 2018) and recently with causal learning task performance (Don et al., 2016).
In addition, people who perform better on the CRT have been found to be less susceptible to holding paranormal beliefs (Pennycook et al., 2012) and less prone to “unusual experiences” (generally linked to “jumping to conclusions”; Broyd et al., 2019). People with higher CRT scores also perform better at distinguishing fake from real news reports (Bronstein et al., 2019; Pennycook and Rand, 2019). Accordingly, since individual differences in willingness to engage effortful and reflective cognitive processes seem to be linked to propensity for irrational beliefs, then one can predict that people scoring low on the CRT will tend to be more likely to believe that the COVID-19 pandemic is a hoax, that risks are exaggerated, or that aspects of the guidelines are not to be believed. Consequently, they should be less likely to engage in precautionary behaviors, such as social distancing, wearing face masks, isolating, hand washing, etc. Indeed, Xie et al. (2020) have shown an effect of working memory on social distancing behavior – and since working memory performance is highly correlated with the CRT (e.g., Toplak et al., 2011; Gray and Holyoak, 2020), these data also strongly suggest a link between algorithmic thinking and precautionary behavior.
Furthermore, because adopting this range of behavior is effortful as one needs to change routines drastically, precautionary behavior should be observed more frequently when the underlying reasons are clear to the person (Bavel et al., 2020; Webster et al., 2020). More reflective people with a tendency to open-minded thinking (AOT) – a higher likelihood to inform themselves and adapt their judgments about the pandemic-related behaviors – are therefore predicted to take in new information about the pandemic and follow the official guidelines. There may, however, be a further reason why AOT would correlate with the uptake and compliance of precautionary behavior. This is the suggestion that people low on AOT scores tend to be politically more conservative, which is in some contexts (e.g., in the United States) associated with skepticism in government policies and official guidelines (Price et al., 2015; Baron, 2019). Allcott et al. (2020) employed United States geo-location data from smart phones and showed that republican-voting areas engage in less social distancing (controlling for other factors, including population density and local COVID cases). We therefore predict that people with more conservative leanings would score lower on the AOT and potentially also be less willing to adopt precautionary measures. To further disentangle cognitive inhibition performance (which the CRT measures) and thinking dispositions from general tendencies for impulsive behavior, we measure impulsivity and risk-taking tendencies separately.
Attention and Cognitive Failures Questionnaire
Dual-process theories often refer to processes that are demanding of attentional and working memory resources when describing Type 2 thinking. However, attention and working memory are hardly ever tested directly in judgment and decision surveys.
People sometimes make mistakes even with rather mundane and familiar tasks, and common everyday failures can be measured by Broadbent’s cognitive failures questionnaire (CFQ; Broadbent et al., 1982, 1986a,b). The CFQ asks people to self-rate their propensity for slips of the mind that lead them to forget names, faces, or certain tasks. There is good evidence that the CFQ correlates with both self-reported and independently recorded errors and accidents (Wallace et al., 2003; Wallace and Chen, 2005; van Doorn et al., 2010; Day et al., 2012) and is associated with absentmindedness (Ishigami and Klein, 2009). Indeed, a recent systematic review of how CFQ self-report scores correlate with objective measures of executive function domains shows that CFQ is mainly associated with performance in selective attention (Carrigan and Barkus, 2016), rather than working memory or inhibition performance. Following the tripartite model of Stanovich (2011) and its explicit mention of attentional processes (Stanovich et al., 2019, p. 1118 and 1123), we included the CFQ as a measure of attentional capacity contributing to the algorithmic layer (Type 2) processing in addition to the measure of inhibition and simulation processes provided by the CRT.
That measuring differences in attention as an additional factor for predicting self-reported uptake of precautionary measures is reasonable is corroborated by evidence from field studies. A recent, but pre-COVID, review found that minimal hand-hygiene interventions at workplaces were effective in reducing the incidence of employee illness (Zivich et al., 2018). Almost all the interventions included in the review that effectively increased compliance involved drawing attention to hand washing and/or diminishing the load on people’s working memory.
Mental Models
A further factor in predicting health behavior is the degree of knowledge and understanding of a disease. A recent review found that limited or insufficient health literacy was associated with reduced adoption of protective behaviors such as getting vaccinated (Castro-Sánchez et al., 2016). Sax and Clack (2015) also reviewed work showing that poor mental models affect uptake of hand hygiene in hospitals.
Mental models are representations of the world and its objects, the relationships between its various parts, and include perceptions about one’s own actions and their consequences. Mental models are distinct from mere knowledge or images, as they can contain abstract elements (Johnson-Laird and Byrne, 1991). Simply presenting people with scientific evidence does not mean that they fully understand – in a scientific sense – mechanisms of transmissions, prevention, and course of a (infectious) disease, because people form their own mental models about the biological and physical world influenced by their experiences and background knowledge. These conceptions often deviate substantially from scientific models (Legare and Gelman, 2008; Jee et al., 2015). For example, Sigelman (2012) found that when asked about the origins of the common cold, United States eighth graders (13–14 year olds) assigned cold weather explanations greater importance than germ-based explanations.
In our study questionnaires measure knowledge of COVID-19, asking which symptoms are related to COVID-19 (compared to common flu), questions probing the quality of the mental model of disease (transmission and immunity – again, compared to the common flu), prevention behavior (past and intended). Our predictions are that greater knowledge about COVID-19 symptoms and a better mental model about the disease transmission and prevention will correlate with better uptake of suggested precautionary measures. The logic of the study in terms of cognitive processes is summarized in Figure 1.
Figure 1. A tripartite model of thinking processes (A) adapted from Stanovich (2011) and its application in the current study (B).
The Current Study
Our survey measures cognitive variables related to dual-process frameworks, risk-taking, the knowledge or mental models people have developed about COVID-19, people’s understanding of the disease, and how these predict compliance with official prevention behaviors (including hand washing, wearing face masks, etc.). In addition, demographic (including political leanings) and experiential variables (such as media usage during the pandemic) were measured.
Following the dual-framework model, we hypothesize that cognitive reflection measures of the algorithmic level (CRT and CFQ) will predict uptake of the officially suggested COVID-19 prevention measures independently of demographic variables (age, sex, and concern about the pandemic) and impulsivity-related individual differences (risk-taking and behavioral inhibition tendencies). We also hypothesize that the amount of AOT, symptom knowledge and quality of the mental model of the disease will predict reported uptake of precautionary measures. To provide a baseline for assessing the explanatory power of the dual-process model, we will compare results with those from a simplified TPB model (which uses measures of different types of beliefs about the suggested behaviors) for predicting adherence to official behavioral guidelines.
Method
Procedure
The study used a cross-sectional quantitative design. Participants completed an online survey created using Qualtrics (2018). The data were collected on April 28, 2020. The order of questions is shown in Table 1. The data were analyzed with R 4.0.1 (R Core Team, 2020). The full questionnaire, datasets, R code and full results including additional analyses are openly accessible at the Open Science Foundation.1
Participants
We collected data from 300 participants surveyed online using Prolific Academic (female: N = 206; age: M = 33.89 years, SE = 0.72; see OSF for a post-hoc bootstrap power analysis). Only participants who were currently resident in the United Kingdom and had English as their first language were allowed to participate, using Prolific Academic’s pre-screening database. Participants were paid £2 and completed the questionnaire in an average time of 16.41 min (SE = 0.47). Two of the 300 respondents were omitted from the analyses due to missing age entries. Table 1 presents the means of all the measures captured.
Measures
Demographics
Participants’ age, gender, and employment status were identified automatically from Prolific Academic’s database. Political leanings were assessed with one simple question: “Please choose the option that best represents your political views on a 7-point scale” from “Strongly left wing” to “Strongly right wing.”
COVID-19-Related Questions
We asked participants whether they were currently staying in their main home or somewhere else and whether this was in a city or the countryside. Two more brief questions established whether they were self-isolating during the last few weeks since the start of the pandemic and whether anybody in the household had tested positive for the virus. We also asked them questions about how many hours they spent on consuming news (traditional via papers, radio and TV, or online) before the pandemic and now during the pandemic to generate a score reflecting the self-reported change (News.Diff) in news consumption (after the pandemic minus before the pandemic). Finally, we asked “How concerned are you about your own personal safety and that of people close to you in terms of the virus?,” measuring respondents’ concern with a 5-point Likert scale (from “A great deal” to “not at all”).
Mental Models, Symptoms Knowledge, and Prevention Behavior
In order to evaluate participants’ knowledge of symptoms and their mental models related to COVID-19, we asked them two sets of questions related to symptoms and attributes of COVID-19. These items probed knowledge about the disease which was, in the period during and prior to the data collection, broadly disseminated by the national and international health organizations (Centers for Disease Control, 2020; National Health Service United Kingdom, 2020; Public Health England, 2020; Robert Koch Institute, 2020; World Health Organisation, 2020a) as well as official news media (Gallagher, 2020). As a part of the information about COVID-19 directed to the public, the differences between COVID-19 and flu have been highlighted (World Health Organisation, 2020b). Our participants were asked about symptoms and attributes (mental model) of flu in the same set of items which related to COVID-19.
Knowledge of Symptoms
Participants were provided a list of eight disease symptoms (fever, shortness of breath, dry cough, headaches, aches and pains, sore throat, fatigue, and runny or stuffy nose) and were asked to evaluate how frequently they occur in cases of COVID-19 and, separately, flu (answer options were “none,” “rare,” “sometimes,” and “common”). The correctness of their answers was evaluated according to the status of knowledge disseminated by media (e.g., CBS News, 2020; Woodward and Gal, 2020) and health authorities (Centers for Disease Control, 2020; National Health Service United Kingdom, 2020; Public Health England, 2020; Robert Koch Institute, 2020; World Health Organisation, 2020a) in March and April 2020. The symptoms score (S.Diff) was calculated as the difference between summed scores. Respondents were scored one point for each correct response, zero otherwise. The symptoms score (S.Diff) was calculated as the difference between summed scores for flu symptoms and COVID-19 symptoms.
Mental Models
In order to understand participants’ mental models of COVID-19 and flu, we listed eight statements pertaining to each disease, e.g., “there is a vaccine available”2 and “the symptom onset is gradual (rather than abrupt)” and asked participants to evaluate (yes/no) whether they apply to (a) COVID-19 and (b) flu. Again, the mental models score (M.Diff) was calculated as the difference between correct sum of scores for correct flu statements and the sum of COVID-19 knowledge. As with the symptoms above, our rationale was that the difference score would be more informative, assessing how much more (or less) people would know about COVID-19 compared to the well-known flu.
Prevention Behavior
To measure participants’ self-reported prevention behavior, we used a set of 17 items referring to COVID-19 prevention measures recommended by the authorities (e.g., “avoid touching surfaces in public” and “reduce using public transport”). For each of these items, participants reported dichotomously (yes/no) whether they (a) “currently do this or have recently (in the last two months)” and (b) “plan to do this from now on.”
Additionally, participants rated (from 1 = strongly disagree to 5 = strongly agree; a) perceived effectiveness of prevention behavior (“Do you agree that the actions mentioned above are effective?”), (b) its feasibility (“Do you agree with the following statement: ‘It will be easy to do these actions’”?), and (c) its application by significant others (“Do you agree with the statement: ‘In general, people important to you are following these actions’”?)
Two different presentation-orders of these three measures were randomly employed in the online questionnaire: (1) symptoms knowledge, (2) mental model of the disease, (3) prevention behavior and (1) prevention behavior, (2) symptoms knowledge, (3) mental model of the disease. There were no significant differences between these two conditions consequently in further analyses, the data were pooled.
Impulsivity and Risk-Taking
In order to control for the potential moderating effects of impulsivity, the Barratt Impulsiveness Scale (BIS) was administered (Patton et al., 1995). In the original version, participants respond to 30 items [on a four-point Likert scale from 1 (never/rarely) to 4 (almost always/always)]. We used an abbreviated scale of eight items based on the brief version of the scale by Steinberg et al. 2013. A sample item is “I don’t pay attention.”
The Risk Propensity Scale or Risk-Taking Index (RTI) was designed to assess risk preferences through a short self-report (Nicholson et al., 2005). Participants were asked to use five-point ratings (from 1 = never to 5 = very often) for six categories of risks: Recreational, Health, Career, Financial, Safety, and Social. These had to be rated twice: one for now and one for in the past, e.g., “We are interested in everyday risk-taking. Please could you tell us if any of the following have ever applied to you, now or in your adult past? – recreational risks (e.g., rock-climbing and scuba diving).”
Cognitive Reflection
The six CRT items were taken from two articles (Frederick, 2005; Toplak et al., 2014) excluding the “bat-and-ball” problem, due to its now high level of familiarity. A “decoy” item consisting of a simple mathematical problem (with no “lure” response) was shown as the first item (the “cargo ship problem”; Thomson and Oppenheimer, 2016), but did not contribute to CRT performance score. Respondents were asked to enter the correct number using their keyboard. Correct responses were scored with 1, while incorrect responses were given 0, and so the maximum total score was 6.
Sample Item
If it takes five machines 5 min to make five widgets, how long would it take 100 machines to make 100 widgets? ____ min (Correct answer: 5 min and intuitive answer: 100 min).
Cognitive Failures
The original CFQ consists of 25 items (Broadbent et al., 1982) arranged on a 5-point Likert scale (0 = never to 4 = always). Possible total scores range from 0 to 100 and Cronbach’s alpha for the scale has been found to be 0.90 and above, and it has been reported to have a test-retest reliability of 0.82 over a 2-month interval (Vom Hofe et al., 1998). We used a short form of the CFQ by Wassenaar et al. (2018), which retained 14 out of the original 25 items.
An example item is “Do you find you forget whether you’ve turned off a light or a fire or locked the door?”
Actively Open-Minded Thinking
AOT questionnaire (Baron, 1993; Haran et al., 2013) measures the willingness to consider new information and remain “open-minded.” Participants responded to items (e.g., Changing your mind is a sign of weakness) on a scale from 1 (completely disagree) to 7 (completely agree).
Results
Two participants did not provide their age, so we omitted their data, leaving 298 respondents whose demographics and related background are summarized in Table 1.
The Cronbach’s alphas for the main psychological scales (AOT = 0.73, RTI = 0.76, BIS = 0.78, CFQ = 0.88, CRT = 0.74, and Prevention-Not-Now (P.Not.Now) = 0.87) ranged from acceptable to good (breakdown data for each question are available online at https://osf.io/8ahs5/).
We evaluated how well two different models predicted the extent of preventive behavior: a dual-process theory (DPT) model, and the TPB model. The dependent variable (DV) for each model was how many preventive measures against infection individuals reported as currently not doing, which was measured by the variable P.Not.Now. This is a count of “not” or “negative” answers and was coded as 1 for every “no” answer and 0 for every “yes” answer. The variable total score was calculated as the count of the 17 individual preventive methods and ranged from 0 (providing zero “not” answers, i.e., currently doing all the preventive methods) to 17 (providing 17 “not” answers, i.e., not currently doing any of the preventive methods).
Both models (dual-process, TPB) were evaluated using hierarchical regressions, with grouped blocks of independent variables being included sequentially. All the independent variables used in both models are shown in Table 2, and the correlation between them (we excluded potentially COVID-related variables that showed no significant association with the DV or the modeled predictors, such as News.Diff, living at home, political leanings, positive test of COVID, and employment situation – see OSF for the full correlation analysis).
In both models, we included demographic predictors (including “concern for the virus”) in Block 1 and impulsivity and risk-taking indices (BIS and RTI, respectively) in Block 2. In the dual-process model cognitive variables related to algorithmic processing (CFQ and CRT) were tested in Block 3, and AOT and mental models [symptoms (S) and disease (M) – each as difference scores from flu, S.Diff and M.Diff, respectively], in Block 4. In the TPB regression, Block 3 contained the variables relating to beliefs about behavior. Table 2 (the last two rows) identifies the variables included in each model.
Dual-Process Thinking
We started the analysis with a linear regression model (see OSF for additional results). However, P.Not.Now did not follow a normal distribution, and the fitted values from the linear model did not reflect the observed data (see OSF for histograms of observed and fitted data). In particular, the model did not predict any responses at zero (i.e., those with zero “not” answers, which equals full compliance with the list of preventive measures). This was in fact the most common answer.
Because of this excess (inflation) of answers at zero, and P.Not.Now being a count variable, we proceeded to fit the hierarchical model with a Zero-Inflated-Poisson (ZIP) model instead. While a standard Poisson model with the same average as our observed data would predict very few zero observations, the ZIP model attempts to better explain the excess observations at zero. It achieves this by using two separate processes to predict the final count of “not” answers: (1) a Poisson count model and (2) a binomial zero-inflated model. The main count model (1), which assumes a Poisson distribution, predicts the count of “not” answers (i.e., 0,1,2,3, etc.). This model mostly predicts a positive non-zero count (i.e., 1,2,3, etc.), with few zeroes; not enough to fit the observed data, which had an inflation of answers at zero. The excess of observations at zero is predicted by the zero-inflated model (2), which assumes a binomial distribution. This model predicts a binary outcome: it determines the probability of an individual answering with zero “not” responses or non-zero (i.e., one or more – the actual count is predicted by the Poisson count model). According to adjusted R2 and Akaike Information Criterion (AIC), the ZIP model fitted the data much better than the linear model (see OSF for a model fit analysis).
For the dual-process thinking analysis, the independent variables were added to the regression in sequential blocks, as shown in Table 3. The omnibus test of each additional block is also shown in Table 3, with every additional introduction of independent variables significant (p < 0.05).
Table 3. Independent variables included in each sequential stage of the dual-process thinking hierarchical regression analysis.
The results of the models are shown in Table 4. The two processes can be interpreted separately. First, in the zero-inflated part of the model (which predicts zero or non-zero “not” answers), there was a significant effect of age, with older participants more likely to provide zero “not” answers (i.e., adopting all preventive methods), but no significant difference according to gender, in Block 1. More concerned participants were also more likely to provide zero “not” responses. In Block 4, there was a significant effect of the mental models of the virus (M.Diff). M.Diff measures how well participants understood the characteristics of the virus (compared to their understanding of the common flu virus). Participants who were more knowledgeable of the virus were more likely to provide zero “not” answers – i.e., adopt all preventive behaviors.
Table 4. Coefficients for the independent variables from each of the dual-process thinking hierarchical regressions.
Second, in the count part of the model (which predicts the count of “not” answers), there was again a significant effect of concern in Block 1, with a negative coefficient; participants who were more concerned responded with fewer “not” answers (i.e., adopted more of the preventive behaviors). In Block 2, there was a significant effect of RTI, with more risk-taking participants who scored higher on RTI adopting fewer preventive behaviors, but no significant effect of BIS. In Block 3, there was a significant effect of CRT, with participants who scored higher on CRT adopting fewer preventive behaviors. There was no significant effect of CFQ. Overall, the observed R2 of the model was 0.46.
We also conducted a factor analysis, in order to better understand the relationship between the underlying individual responses which comprised P.Not.Now. We identified five factors based on shared correlations and common themes [(1) social distancing, (2) cleanliness, (3) mask usage, (4) sneezing protection, and (5) isolation]. We were particularly interested in the unusual correlation found with CRT. We found that the only factor which was positively correlated with CRT (i.e., the higher the CRT score, the higher the count of “not” answers) was factor 2 (cleanliness), with a correlation r(297) = 0.20, p < 0.001. This was confirmed by running the DPT models above on the biggest factors, factor 1 (social distancing – CRT is not a significant predictor, p = 0.116) and factor 2 (cleanliness – CRT is a significant predictor, p = 0.005; see OSF for more details on the factor analysis).
Theory of Planned Behavior Analysis
We also evaluated a TPB model using a ZIP analysis, with the independent variables as shown in Table 5. All the individual steps of the analysis led to a significant improvement of model fit in comparison to the previous step.
Table 5. Independent variables included in each sequential stage of the TPB hierarchical regression analysis.
The results of the TPB analysis are shown in Table 6. Similarly to the previous model, in among the demographics included in Block 1 in the zero-inflated model, there was a significant effect of age, with older participants more likely to respond with zero “not” answers, but no significant difference according to gender. There was also a significant effect of concern, with more concerned participants also more likely to respond with zero “not” answers.
In the count model, in Block 2, there was also a significant effect of RTI, with more risk-taking participants who scored higher on RTI adopting fewer preventive behaviors, but no significant effect of BIS.
In Block 3, there was a significant effect of P.Easy, with participants adopting more preventive behaviors when they reported finding them easier. There was no significant of the P.Effect (how effective the behaviors were rated) or P.Follow (the extent to which their friends and relatives were also following the preventive measures). Overall, the observed R2 of the model was 0.49.
We then compared the two models according to AIC. The TPB model showed a slightly lower AIC (1550) than the DPM (1555), but the difference is small. Both models have a much better AIC than the linear regression model (see OSF for a model comparison analysis). Figure 2 illustrates the correlations (for coefficients r > 0.10) in both regression analyses (DPTM and TPB) between predictors and between the predictors and criterion (P.NotNow) in a network plot.
Figure 2. Visual representation of the correlations between the DV and all the IVs in both models (dual-process theory, DPT and Theory of Planned Behavior, TPB), similar to a network plot. Only correlations greater than 0.1 are plotted. Black lines indicate positive correlations and red lines indicate negative correlations. The darkness and thickness of the lines represent the strength of the correlation. The spatial location and proximity of the variables are determined by classical multidimensional scaling based on the absolute values of the correlations.
Discussion
This online study is the first to our knowledge to test predictions from the DPT in the field of judgment and decision-making in relation to precautionary behavior in response to, and during, a pandemic. We found that cognitive factors, such as cognitive reflection and the quality of mental models (knowledge about the disease mechanism), predicted the amount of self-reported precautionary behaviors (including hand washing, wearing face masks, etc.) and hence compliance with official prevention guidelines.
The results from the first-order correlation analysis and subsequent hierarchical regression modeling are relatively clear: demographic factors previously associated with health behavior (Pack et al., 2001; Deeks et al., 2009), such as age (but not sex, see Branas-Garza et al., 2020 for a similar result regarding COVID-related donation behavior) as well as felt concern about the virus, explained a significant proportion of the variance on the DV, as did the RTI: older participants, respondents who were more concerned about the virus, and those self-reporting as less risk-taking in normal life, reported greater adherence to precautionary measures.
Interestingly, the cognitive reflection performance as measured by the CRT (even after accounting for thinking disposition, AOT) and measures of cognitive failures – which have not been used in the context of pandemic behavior, and hardly at all in the health behavior literature in general – correlated with preventive behavior: people reporting greater incidences of cognitive failures reported less behavioral adherence (although the individual contribution of CFQ observed in the first-order correlations is not significant anymore in the regression analysis). This would be predicted by standard cognitive theories, based on the notion that cognitive failures – as a proxy measure of attentional capacity – is linked to working memory (Heitz et al., 2005; Unsworth and Spillers, 2010; Oberauer, 2019) and hence to performance on tasks relying on such executive functions (McCabe et al., 2010; Xie et al., 2020).
Cognitive reflection performance as measured by the CRT uniquely predicted a portion of the variance in precautionary behavior. However, counter to our hypothesis, this correlation was negative – that is, people scoring lower on the CRT (and presumably leaning toward heuristics, fast judgments, and decisions) were more likely to engage in the recommended distance and hygiene measures to prevent the spread of COVID-19. In line with dual-process models as well as general conceptions about relevant health behavior tested in a pandemic (e.g., Bavel et al., 2020), we expected more reflective individuals to be more compliant, as, for them, the need for engaging in such demanding tasks – involving working memory and prospective memory (Xie et al., 2020) – should be easier to understand, plan, and adhere to. We discuss further possible explanations for this surprising finding below.
Finally, AOT and knowledge about the symptoms of the new disease did not predict reported behavior. AOT did, however, correlate positively with CRT – meaning that actively open-minded people are more prone to cognitive reflection, which is of course in line with the tripartite model (Stanovich, 2011).
TPB and Cognition
The results of the current study show that TPB as a model of health-related behavior also predicted the uptake and maintenance of current precautionary behaviors at the first height of the COVID-19 pandemic. However, of the three behavioral attitudes only the variable of perceived behavioral control (the item measuring how easy it was to follow the behavioral advice) was a significant predictor (although the three attitudinal variables correlated with another). It is likely that, in this current pandemic, subjective norms were already at ceiling and that the vast majority was following the guidelines (early indications point to 83% compliance in the United Kingdom, Weinberg, 2020). Moreover, behavior compliance was relatively enforced (police checks on unnecessary travel) and alternative behavioral opportunities were already heavily curtailed (work places, entertainment venues, shops etc., closed).
Finally, it should be acknowledged that our TPB model was highly simplified, measuring behavioral attitudes with only three questions (perceived control, social norms, and effectiveness). Nevertheless, TPB predicted a substantial proportion of variance in precautionary behavior, explaining a similar amount of variance than the dual-process model.
Explanations for CRT Correlation
Against our expectations, the correlation between CRT scores and avoidance of precautionary behavior was positive (i.e., the correlation between CRT and P.Not.Now was positive, with higher CRT scores correlated to more “not” answers to precautionary actions); more reflective people adopted fewer preventive behaviors. Our original expectation was based on the general notion that the tasks in the heuristics and biases literature are deliberately constructed to induce a heuristically triggered response, which needs to be overridden by a normative response generated by the analytic system. According to Stanovich’s concept of “cognitive decoupling,” the CRT measures the ability to inhibit automatic responses and simulate alternative responses (Stanovich, 2011). Our premise was that this ability would be needed if people were to adhere to precautionary measures, as they would need to override automatic responses, such as relying on their previous default behavior (Johnson and Goldstein, 2003) and in addition use mental simulation to employ the correct measures at the correct time, in the correct order. Similar reasoning has been invoked to explain why high CRT scorers are less likely to believe in conspiracy theories and fake news (Pennycook and Rand, 2019).
Concerning other correlations with the CRT, previous work has also reported effects of gender (Frederick, 2005; Campitelli and Gerrans, 2014; Thomson and Oppenheimer, 2016; Branas-Garza et al., 2020) using the classic three-item version, with male participants usually outperforming females. One reason given for this observation is that males have higher numeracy (Baron et al., 2015), though Campitelli and Gerrans cite both numeracy and rational thinking ability, whereas others think the difference could be due to higher anxiety or lower self-assessment on numerical aptitude (e.g., Zhang et al., 2016). Note that we omitted the notorious “bat-and-ball” question from the classic three-item test, which may have contributed significantly to the previously reported association with numeracy (Sinayev and Peters, 2015) and added four items from Toplak et al. (2014), which arguably are less reliant on numeracy. Less frequently reported are associations with age, with some authors finding no correlation (Campitelli and Gerrans, 2014; Thoma et al., 2015).
According to Baron (2017), the CRT is largely a measure of reflection/impulsivity: the willingness to take more time in order to be more accurate on judgment tasks, and CRT scores should therefore correlate with other normative responses. Clearly, this was not the case here for our type of responses, precautionary measures. Some commentators see the dual-system approach as only valid in well-structured environments such as psychological laboratory settings (Dane and Pratt, 2007; Hogarth, 2010; Magnusson et al., 2014). A similar argument is made by Risen (2016) who argues that Type 2 processing can be indeed differentiated as error detection and correction but adds the notion that error correction does not necessarily follow when an error is detected – and hence “acquiescence” is a possible System 2 response. This arguably explains why even “smart” people believe in magical thinking and superstition.
But although this approach may explain why we did not find a negative correlation between CRT and P.Not.Now, it does not explain why we still see a significant positive correlation between CRT and P.Not.Now. It is generally assumed that the CRT measures heuristic processing, and heuristics are thought to work through “attribute substitution”: when asked to answer a hard question (i.e., make numerical judgments) people substitute it with an easier one (e.g., “how easy does the answer come to mind?”; Kahneman and Frederick, 2002), which causes judgment biases. Able individuals’ Type 2 processing – measured with the CRT – will however, intervene and stop this substitution of a hard-to-evaluate characteristic for an easy one and usually improve judgment performance. However, according to West et al. (2012), when it comes to judgments about risks, Type 2 processing may do the opposite: “For example, people will substitute the less effortful attributes of vividness or salience for the more effortful retrieval of relevant facts. But when we are evaluating important risks—such as the risk of certain activities and environments for our children—we do not want to substitute vividness for careful thought about the situation. In such situations, we want to employ Type 2 override processing to block the attribute substitution.” (p. 508).
A different possible explanation could be that people with high CRT score thought more than others about the different guidelines and associated behavior, and in turn queried them critically to the point of higher non-adherence. For example, there is evidence that during an Ebola epidemic health professionals in quarantined villages were less likely to adhere to the quarantine than (presumably less knowledgeable) volunteers (see Webster et al., 2020). We originally hypothesized that the (perceived) effort of compliance with precautions would make less reflective people reluctant to adopt precautionary measures. However, conceivably, the effort of compliance may also spur the more reflective to think of reasons to override the prescribed behaviors; following precautionary guidelines, while effortful, may be cognitively simpler than generating reasons to dissent. If so then, accordingly, the non-compliant might conceivably be a mixture of two types: thoughtless recalcitrants (low on CRT) and thoughtful sceptics (high on CRT). The blend of each – and so the observed relationship between CRT and compliance – may depend on such things as the strength of social norms to comply (including how consistently experts endorse the measures) and how many other like-minded and/or critical people one is proximal to.
So could one have predicted these results if one assumes that irrational behavior (as measured by the CRT) depends on the perceived rationality or irrationality of the suggested measures by policy-makers and governments (for example, if people thought the measures were too drastic or even counterproductive then may be the positive CRT correlations express rational thinking)? Given that the data were collected at the height of the pandemic’s first wave (not only in the United Kingdom but also across Europe) and the measures (i.e., behavioral guidelines) we asked about apparently had a drastic effect in reducing infections, we think we rightly assumed that rational thinking and precautionary behavior were indeed linked at that time (in the first wave). Compliance in the population was very high then, and of course hygiene measures are widely accepted to be effective (although we now know that social distancing is even more important). Also, during the pandemic’s first wave many people have died, a strong argument for the rationality of these behavioral measures. Finally, the variable measuring concern did correlate positively with uptake of these measures.
Yet another possible explanation for the positive correlation between CRT and P.Not.Now is the negative association between CRT and prosocial acting. According to the recent study by Campos-Mercade et al. (2021), prosociality predicted health behaviors during the COVID-19 pandemic. According to Capraro et al. (2017), intuition is connected with concern for relative shares (which could be not only egalitarian but also spiteful), whereas deliberation is associated with individuals’ focus on social efficiency. In the context of economic games (e.g., the dictator game, the ultimatum game, and the prisoner’s dilemma), it was found that high cognitive reasoning and intelligence are negatively associated with cooperation and prosociality (Yamagishi et al., 2014) particularly in situations when the participants’ lack of cooperation did not have any negative consequences for them such as in one-shot games (Barreda-Tarrazona et al., 2017; Inaba et al., 2018). This association disappears in situations when cooperation has no or very low cost for the individual (Ponti and Rodriguez-Lara, 2015; Corgnet et al., 2016).
Based on these findings, prosociality was proposed to be connected with intuitive processes and the findings led to the social heuristics’ hypothesis, according to which intuition increases prosociality for people used to cooperative interactions (Rand et al., 2014; but see Chen et al., 2013; Verkoeijen and Bouwmeester, 2014). Clearly there is a need for further research to disentangle the significance of CRT scores from other psychological variables and contextual effects.
Cognitive Failures Questionnaire
The CFQ correlated negatively with precautionary behavior, in the first-order correlations, although there was not a unique contribution of cognitive failures in the regression model. CFQ scores are related in the literature to variables, such as selective attention, multi-tasking, worry, stress, and boredom (Robertson et al., 1997; Wallace et al., 2003; Linden et al., 2005) – all factors that can be expected to play a major role in a lockdown situation in which many of the respondents will have found themselves in the United Kingdom. The main reason for including the CFQ was to enable us to disentangle cognitive reflection (CRT: cognitive inhibition and mental stimulation) from other cognitive processes (e.g., selective attention). Therefore, we cannot currently pinpoint a potential link between cognitive failures and precautionary behavior but given its association with a range of psychological factors, further research should be conducted to elucidate its role in preventative behavior.
Mental Models
Knowledge of COVID-19 symptoms (S.Diff – comparing with knowledge of common flu symptoms) did not predict uptake of prevention behavior; however, the quality of the mental model around disease transmission and infection (M.Diff) did – similar to what was found, for example, for hospital staff (Sax and Clack, 2015). Regarding the lack of effects from symptoms knowledge, one possible reason could be a floor effect (median for P.Not.Now was 3, meaning that participants were doing 82% of all the possible actions) and that people were already well-informed at the height of the pandemic. Indeed, we did not find a correlation between P.Not.Now and additional (since the pandemic) news consumption r(298) = −0.02, p = 0.73. Future research will need to address the cause-and-effect relationship between cognitive reflection, mental models, and preventative behavior, but our results make it clear that the quality of information and their uptake by the population have a significant effect on compliance.
Strengths and Limitations
Although based on theory, this study was necessarily exploratory to some degree, simply because of the novel nature of human actions it was investigating: the first global pandemic for 100 years. There are a number of variables that may have shed more light on our findings, e.g., perceived behavioral barriers (influences that discourage adoption of the behavior); also including an explicit measure self-efficacy (as often used within TPB) and measures of altruistic tendencies could help to find explanations for the patterns observed here. Nevertheless, the current research has some significance and originality, as it combines variables from two major theoretical strands of health-related research, the dual-process framework and TPB and demonstrates how these theoretical ideas could help to predict precautionary behaviors, and by extension, save human lives in future.
A further limitation is that we have not included further cognitive control variables – such as numeracy or math skills, which may explain part of the variance in CRT (e.g., Cokely and Kelley, 2009) – to better disentangle the analytic processes associated with predicting precautionary behavior. Furthermore, other variables could have made a contribution to the behavioral scores such as level of education. Another limitation is of course the time frame, as we could not trace changes in perceptions and actions over time during the COVID-19 crisis. Our survey captured the United Kingdom respondents at the height of the first lockdown (end of April 2020), only after which (from May to June 2020) there was an easing of both the pandemic and behavioral guidelines in the United Kingdom. It is possible that certain correlations between cognitive factors and precautionary behavior may be dependent on the length of time in which the measures have been already implemented. For example, it is possible that there would be a negative – instead of the observed positive – correlation between CRT and P.Not.Now in the early days of lockdown, when more reflective individuals may have assessed the situation as graver than the non-reflective.
Conclusions
In a recent Nature Human Behavior perspective article (Bavel et al., 2020) by over forty behavioral scientists reviewing how insights from the social and behavioral sciences can be used to help align human behavior with the recommendations of epidemiologists and public health experts, the authors stressed the need for prosocial messages: e.g., “Leaders and the media might try to promote cooperative behavior by emphasizing that cooperating is the right thing to do and that other people are already cooperating.:…” Messages that (i) emphasize benefits to the recipient, (ii) focus on protecting others, (iii) align with the recipient’s moral values, (iv) appeal to social consensus or scientific norms, and/or (v) highlight the prospect of social group approval tend to be persuasive.” However, these authors did not mention cognitive reflection (or any other cognitive variables) as relevant factors.
In conclusion, our results demonstrate that individual differences in general cognitive abilities (cognitive reflection) and knowledge about the disease (mechanisms about transmission and infectiousness, but not knowledge about symptoms) are significant predictors for behavioral adherence to precautionary behavior in a pandemic, beyond known factors such as age or risk-taking. These variables appear to be as or even more predictive than differences in impulsivity, people’s political views, or where they live (town vs. country). This finding promises to close a gap in understanding compliance with precautionary behavior left by social norms approaches such as TPB.
People were more likely to adhere to official guidelines during the extraordinary COVID-19 pandemic when they were, in general, less reflective in their judgment and decision-making style, possibly due to them following heuristics or simple rules as this was an easier cause of action, they overly criticized the rationality of the guidelines or because they were following social norms. At the same time respondents were also more likely to follow these guidelines when they had a better understanding of the infection mechanism. Future research on cognitive factors in health-prevention behaviors should better establish how cognitive variables are linked with people’s information processing and social norms in order to improve predictions of precautionary behavior.
Data Availability Statement
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found online at: https://osf.io/8ahs5/.
Ethics Statement
The studies involving human participants were reviewed and approved by the City University of London. The patients/participants provided their written informed consent to participate in this study.
Author Contributions
VT and PA conceived of the study, developed the theoretical background, planned and supervised the study, and wrote the article. VT, PA, and PF developed and designed the materials. LW-C ran the survey and analyzed the data and was the main contributor for the results section. PF contributed to literature search, theoretical discussions, write-up, and checking of the manuscript. All authors contributed to the article and approved the submitted version.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Footnotes
2. At the time of the study there was no vaccine available for COVID-19.
References
Ajzen, I. (1991). The theory of planned behavior. Organ. Behav. Hum. Decis. Process. 50, 179–211. doi: 10.1016/0749-5978(91)90020-T
Allcott, H., Boxell, L., Conway, J., Gentzkow, M., Thaler, M., and Yang, D. (2020). Polarization and public health: partisan differences in social distancing during the coronavirus pandemic. J. Public Econ. 191:104254. doi: 10.1016/j.jpubeco.2020.104254
Anderson, R. M., Heesterbeek, H., Klinkenberg, D., and Hollingsworth, T. D. (2020). How will country-based mitigation measures influence the course of the COVID-19 epidemic? Lancet 395, 931–934. doi: 10.1016/S0140-6736(20)30567-5
Armitage, C. J., and Conner, M. (2001). Efficacy of the theory of planned behaviour: a meta-analytic review. Br. J. Soc. Psychol. 40, 471–499. doi: 10.1348/014466601164939
Baron, J. (1993). Why teach thinking?-An essay. Appl. Psychol. 42, 191–214. doi: 10.1111/j.1464-0597.1993.tb00731.x
Baron, J. (2017). Comment on Kahan and Corbin: Can polarization increase with actively open-minded thinking? Res. Polit. 4:205316801668812. doi: 10.1177/2053168016688122
Baron, J. (2019). Actively open-minded thinking in politics. Cognition 188, 8–18. doi: 10.1016/j.cognition.2018.10.004
Baron, J., Scott, S., Fincher, K., and Emlen Metz, S. (2015). Why does the Cognitive Reflection Test (sometimes) predict utilitarian moral judgment (and other things)? J. Appl. Res. Mem. Cogn. 4, 265–284. doi: 10.1016/j.jarmac.2014.09.003
Barreda-Tarrazona, I., Jaramillo-Gutiérrez, A., Pavan, M., and Sabater-Grande, G. (2017). Individual characteristics vs. experience: an experimental study on cooperation in prisoner’s dilemma. Front. Psychol. 8:596. doi: 10.3389/fpsyg.2017.00596
Bavel, J. J. V., Baicker, K., Boggio, P. S., Capraro, V., Cichocka, A., Cikara, M., et al. (2020). Using social and behavioural science to support COVID-19 pandemic response. Nat. Hum. Behav. 4, 460–471. doi: 10.1038/s41562-020-0884-z
Bogg, T., and Milad, E. (2020). Slowing the spread of COVID-19: demographic, personality, and social cognition predictors of guideline adherence in a representative U.S. sample. PsyArXiv. doi: 10.31234/osf.io/yc2gq
Branas-Garza, P., Jorrat, D. A., Alfonso, A., Espin, A. M., García, T., and Kovarik, J. (2020). Exposure to the Covid-19 pandemic and generosity. PsyArXiv. doi: 10.31234/osf.io/6ktuz
Broadbent, D. E., Broadbent, M. H., and Jones, J. L. (1986a). Performance correlates of self-reported cognitive failure and of obsessionality. Br. J. Clin. Psychol. 25, 285–299. doi: 10.1111/j.2044-8260.1986.tb00708.x
Broadbent, D. E., Cooper, P. F., FitzGerald, P., and Parkes, K. R. (1982). The Cognitive Failures Questionnaire (CFQ) and its correlates. Br. J. Clin. Psychol. 21, 1–16. doi: 10.1111/j.2044-8260.1982.tb01421.x
Broadbent, D. E., FitzGerald, P., and Broadbent, M. H. P. (1986b). Implicit and explicit knowledge in the control of complex systems. Br. J. Psychol. 77, 33–50.
Bronstein, M. V., Pennycook, G., Bear, A., Rand, D. G., and Cannon, T. D. (2019). Belief in fake news is associated with delusionality, dogmatism, religious fundamentalism, and reduced analytic thinking. J. Appl. Res. Mem. Cogn. 8, 108–117. doi: 10.1016/j.jarmac.2018.09.005
Broyd, A., Ettinger, U., and Thoma, V. (2019). Thinking dispositions and cognitive reflection performance in schizotypy. Judgm. Decis. Mak. 14, 80–90.
Campitelli, G., and Gerrans, P. (2014). Does the cognitive reflection test measure cognitive reflection? A mathematical modeling approach. Mem. Cogn. 42, 434–447. doi: 10.3758/s13421-013-0367-9
Campos-Mercade, P., Meier, A. N., Schneider, F. H., and Wengström, E. (2021). Prosociality predicts health behaviors during the COVID-19 pandemic. J. Public Econ. 195:104367. doi: 10.1016/j.jpubeco.2021.104367
Capraro, V., Corgnet, B., Espín, A. M., and Hernán-González, R. (2017). Deliberation favours social efficiency by making people disregard their relative shares: evidence from USA and India. R. Soc. Open Sci. 4:160605. doi: 10.1098/rsos.160605
Carrigan, N., and Barkus, E. (2016). A systematic review of cognitive failures in daily life: healthy populations. Neurosci. Biobehav. Rev. 63, 29–42. doi: 10.1016/j.neubiorev.2016.01.010
Castro-Sánchez, E., Chang, P. W. S., Vila-Candel, R., Escobedo, A. A., and Holmes, A. H. (2016). Health literacy and infectious diseases: why does it matter? Int. J. Infect. Dis. 43, 103–110. doi: 10.1016/j.ijid.2015.12.019
CBS News (2020). Think you have coronavirus symptoms? Here’s what to do. Available at: https://www.cbsnews.com/news/coronavirus-symptoms-covid-19-fever-cough-hospital-home/ (Accessed July 29, 2020).
Centers for Disease Control (2020). Coronavirus Disease 2019 (COVID-19) – Prevention and Treatment. Centers for Disease Control and Prevention. Available at: https://www.cdc.gov/coronavirus/2019-ncov/prevent-getting-sick/prevention.html (Accessed July 27, 2020).
Chen, C. C., Chiu, I. M., Smith, J., and Yamada, T. (2013). Too smart to be selfish? Measures of cognitive ability, social preferences, and consistency. J. Econ. Behav. Organ. 90, 112–122. doi: 10.1016/j.jebo.2013.03.032
Cokely, E. T., and Kelley, C. M. (2009). Cognitive abilities and superior decision making under risk: a protocol analysis and process model evaluation. Judgm. Decis. Mak. 4, 20–33.
Corgnet, B., Espín, A. M., Hernán-González, R., Kujal, P., and Rassenti, S. (2016). To trust, or not to trust: cognitive reflection in trust games. J. Behav. Exp. Econ. 64, 20–27. doi: 10.1016/j.socec.2015.09.008
Dane, E., and Pratt, M. G. (2007). Exploring intuition and its role in managerial decision making. Acad. Manag. Rev. 32, 33–54. doi: 10.5465/amr.2007.23463682
Day, A. J., Brasher, K., and Bridger, R. S. (2012). Accident proneness revisited: the role of psychological stress and cognitive failure. Accid. Anal. Prev. 49, 532–535. doi: 10.1016/j.aap.2012.03.028
Deeks, A., Lombard, C., Michelmore, J., and Teede, H. (2009). The effects of gender and age on health related behaviors. BMC Public Health 9:213. doi: 10.1186/1471-2458-9-213
Don, H. J., Goldwater, M. B., Otto, A. R., and Livesey, E. J. (2016). Rule abstraction, model-based choice, and cognitive reflection. Psychon. Bull. Rev. 23, 1615–1623. doi: 10.3758/s13423-016-1012-y
Evans, J. S. B. T. (1977). Toward a statistical theory of reasoning. Q. J. Exp. Psychol. 29, 621–635. doi: 10.1080/14640747708400637
Evans, J. S. B. T. (2003). In two minds: dual-process accounts of reasoning. Trends Cogn. Sci. 7, 454–459. doi: 10.1016/j.tics.2003.08.012
Evans, J. S. B. T. (2008). Dual-processing accounts of reasoning, judgment, and social cognition. Annu. Rev. Psychol. 59, 255–278. doi: 10.1146/annurev.psych.59.103006.093629
Evans, J. S. B. T. (2010). Intuition and reasoning: a dual-process perspective. Psychol. Inq. 21, 313–326. doi: 10.1080/1047840X.2010.521057
Evans, J. S. B. T., Barston, J., and Pollard, P. (1983). On the conflict between logic and belief in syllogistic reasoning. Mem. Cogn. 11, 295–306. doi: 10.3758/BF03196976
Evans, J. S. B. T., and Stanovich, K. E. (2013). Dual-process theories of higher cognition: advancing the debate. Perspect. Psychol. Sci. 8, 223–241. doi: 10.1177/1745691612460685
Ferguson, N. M., Laydon, D., Nedjati-Gilani, G., Imai, N., Ainslie, K., Baguelin, M., et al. (2020). Impact of non-pharmaceutical interventions (NPIs) to reduce COVID-19 mortality and healthcare demand. Imperial College COVID-19 Response Team, London. Available at: https://www.imperial.ac.uk/media/imperial-college/medicine/sph/ide/gida-fellowships/Imperial-College-COVID19-NPI-modelling-16-03-2020.pdf
Frederick, S. (2005). Cognitive reflection and decision making. J. Econ. Perspect. 19, 25–42. doi: 10.1257/089533005775196732
Gallagher, J. (2020). What symptoms should I look out for? BBC News. Available at: https://www.bbc.com/news/health-51048366 (Accessed July 27, 2020).
Gigerenzer, G., and Regier, T. (1996). How do we tell an association from a rule? Comment on Sloman (1996). Psychol. Bull. 119, 23–26. doi: 10.1037/0033-2909.119.1.23
Gray, M. E., and Holyoak, K. J. (2020). Individual differences in relational reasoning. Mem. Cogn. 48, 96–110. doi: 10.3758/s13421-019-00964-y
Haran, U., Ritov, I., and Mellers, B. A. (2013). The role of actively open-minded thinking in information acquisition, accuracy, and calibration. Judgm. Decis. Mak. 8, 188–201. doi: 10.1017/CBO9781107415324.004
Heitz, R. P., Unsworth, N., and Engle, R. W. (2005). “Working memory capacity, attention control, and fluid intelligence” in Handbook of understanding and measuring intelligence. eds. O. Wilhelm and R. W. Engle (Sage Publications, Inc.), 61–77.
Hogarth, R. M. (2010). Intuition: a challenge for psychological research on decision making. Psychol. Inq. 21, 338–353. doi: 10.1080/1047840X.2010.520260
Inaba, M., Inoue, Y., Akutsu, S., Takahashi, N., and Yamagishi, T. (2018). Preference and strategy in proposer’s prosocial giving in the ultimatum game. PLoS One 13:e0193877. doi: 10.1371/journal.pone.0193877
Ishigami, Y., and Klein, R. M. (2009). Are individual differences in absentmindedness correlated with individual differences in attention? J. Individ. Differ. 30, 220–237. doi: 10.1027/1614-0001.30.4.220
Jee, B. D., Uttal, D. H., Spiegel, A., and Diamond, J. (2015). Expert-novice differences in mental models of viruses, vaccines, and the causes of infectious disease. Public Underst. Sci. 24, 241–256. doi: 10.1177/0963662513496954
Jimenez, N., Rodriguez-Lara, I., Tyran, J. -R., and Wengström, E. (2018). Thinking fast, thinking badly. Econ. Lett. 162, 41–44. doi: 10.1016/j.econlet.2017.10.018
Johnson, E. J., and Goldstein, D. (2003). Medicine: do defaults save lives? Science 302, 1338–1339. doi: 10.1126/science.1091721
Johnson-Laird, P. N., and Byrne, R. M. J. (1991). Essays in cognitive psychology. Deduction. Hillsdale, USA: Lawrence Erlbaum Associates, Inc.
Kahneman, D., and Frederick, S. (2002). “Representativeness revisited: attribute substitution in intuitive judgment” in Heuristics and biases: The psychology of intuitive judgment. eds. T. Gilovich, D. Griffin, and D. Kahneman (New York: Cambridge University Press), 49–81.
Kahneman, D., and Frederick, S. (2005). “A model of heuristic judgment” in The Cambridge handbook of thinking and reasoning. eds. K. J. Holyoak and R. G. Morrison (New York: Cambridge University Press), 267–293.
Kahneman, D., and Frederick, S. (2007). Frames and brains: elicitation and control of response tendencies. Trends Cogn. Sci. 11, 45–46. doi: 10.1016/j.tics.2006.11.007
Keizer, A. G., Tiemeijer, W., and Bovens, M. (2019). Why knowing what to do is not enough: A realistic perspective on self-reliance. Springer Nature.
Keren, G., and Schul, Y. (2009). Two is not always better than one: a critical evaluation of two-system theories. Perspect. Psychol. Sci. 4, 533–550. doi: 10.1111/j.1745-6924.2009.01164.x
Klauer, K. C., Musch, J., and Naumer, B. (2000). On belief bias in syllogistic reasoning. Psychol. Rev. 107, 852–884. doi: 10.1037/0033-295X.107.4.852
Kruglanski, A. W., and Gigerenzer, G. (2011). Intuitive and deliberate judgments are based on common principles. Psychol. Rev. 118, 97–109. doi: 10.1037/a0020762
Lauerman, J. (2020). Covid-19 pandemic likely to last two years, report says. Available at: https://www.bloomberg.com/news/articles/2020-05-01/covid-19-pandemic-likely-to-last-two-years-report-says (Accessed July 25, 2020).
Lee, L. Y., Lam, E. P., Chan, C., Chan, S., Chiu, M., Chong, W., et al. (2020). Practice and technique of using face mask amongst adults in the community: a cross-sectional descriptive study. BMC Public Health 20:948. doi: 10.1186/s12889-020-09087-5
Legare, C., and Gelman, S. (2008). Bewitchment, biology, or both: the co-existence of natural and supernatural explanatory frameworks across development. Cogn. Sci. Multidiscip. J. 32, 607–642. doi: 10.1080/03640210802066766
Linden, D. V. D., Keijsers, G. P. J., Eling, P., and Schaijk, R. V. (2005). Work stress and attentional difficulties: an initial study on burnout and cognitive failures. Work Stress 19, 23–36. doi: 10.1080/02678370500065275
Lowe Bryan, W., and Harter, N. (1899). Studies on the telegraphic language: the acquisition of a hierarchy of habits. Psychol. Rev. 6, 345–375. doi: 10.1037/h0073117
Macpherson, R., and Stanovich, K. E. (2007). Cognitive ability, thinking dispositions, and instructional set as predictors of critical thinking. Learn. Individ. Differ. 17, 115–127. doi: 10.1016/j.lindif.2007.05.003
Magnusson, P. R., Netz, J., and Wästlund, E. (2014). Exploring holistic intuitive idea screening in the light of formal criteria. Technovation 34, 315–326. doi: 10.1016/j.technovation.2014.03.003
McCabe, D. P., Roediger, H. L. III, McDaniel, M. A., Balota, D. A., and Hambrick, D. Z. (2010). The relationship between working memory capacity and executive functioning: evidence for a common executive attention construct. Neuropsychology 24:222. doi: 10.1037/a0017619
National Health Service United Kingdom (2020). Coronavirus (COVID-19). Available at: https://www.nhs.uk/conditions/coronavirus-covid-19/ (Accessed July 27, 2020).
Nicholson, N., Soane, E., Fenton-O’Creevy, M., and Willman, P. (2005). Personality and domain-specific risk taking. J. Risk Res. 8, 157–176. doi: 10.1080/1366987032000123856
Oberauer, K. (2019). Working memory and attention – a conceptual analysis and review. J. Cogn. 2:36. doi: 10.5334/joc.58
Oechssler, J., Roider, A., and Schmitz, P. W. (2009). Cognitive abilities and behavioral biases. J. Econ. Behav. Organ. 72, 147–152. doi: 10.1016/j.jebo.2009.04.018
Osman, M. (2004). An evaluation of dual-process theories of reasoning. Psychon. Bull. Rev. 11, 988–1010. doi: 10.3758/BF03196730
Pack, R. P., Crosby, R. A., and Lawrence, J. S. S. (2001). Associations between adolescents’ sexual risk behavior and scores on six psychometric scales: impulsivity predicts risk. J. HIV/AIDS Prev. Educ. Adolesc. Child. 4, 33–47. doi: 10.1300/J129v04n01_04
Patton, J. H., Stanford, M. S., and Barratt, E. S. (1995). Factor structure of the Barratt impulsiveness scale. J. Clin. Psychol. 51, 768–774. doi: 10.1002/1097-4679(199511)51:6<768::AID-JCLP2270510607>3.0.CO;2-1
Pennycook, G., Cheyne, J. A., Seli, P., Koehler, D. J., and Fugelsang, J. A. (2012). Analytic cognitive style predicts religious and paranormal belief. Cognition 123, 335–346. doi: 10.1016/j.cognition.2012.03.003
Pennycook, G., and Rand, D. G. (2019). Lazy, not biased: susceptibility to partisan fake news is better explained by lack of reasoning than by motivated reasoning. Cognition 188, 39–50. doi: 10.1016/j.cognition.2018.06.011
Ponti, G., and Rodriguez-Lara, I. (2015). Social preferences and cognitive reflection: evidence from a dictator game experiment. Front. Behav. Neurosci. 9:146. doi: 10.3389/fnbeh.2015.00146
Price, E., Ottati, V., Wilson, C., and Kim, S. (2015). Open-minded cognition. Pers. Soc. Psychol. Bull. 41, 1488–1504. doi: 10.1177/0146167215600528
Public Health England (2020). Coronavirus (COVID-19). Available at: https://www.gov.uk/government/organisations/public-health-england (Accessed July 27, 2020).
R Core Team (2020). R: A language and environment for statistical computing. R Foundation for Statistical Computing. Austria: Vienna.
Rand, D. G., Peysakhovich, A., Kraft-Todd, G. T., Newman, G. E., Wurzbacher, O., Nowak, M. A., et al. (2014). Social heuristics shape intuitive cooperation. Nat. Commun. 5, 1–12. doi: 10.1038/ncomms4677
Risen, J. L. (2016). Believing what we do not believe: acquiescence to superstitious beliefs and other powerful intuitions. Psychol. Rev. 123, 182–207. doi: 10.1037/rev0000017
Robert Koch Institute (2020). Coronavirus SARS-CoV-2 – Informationen zum Erreger (Stand: 15.5.2020). Available at: https://www.rki.de/SharedDocs/FAQ/NCOV2019/FAQ_Liste.html?nn=13490888 (Accessed July 27, 2020).
Robertson, I. H., Manly, T., Andrade, J., Baddeley, B. T., and Yiend, J. (1997). Oops!’: performance correlates of everyday attentional failures in traumatic brain injured and normal subjects. Neuropsychologia 35, 747–758. doi: 10.1016/S0028-3932(97)00015-8
Sætrevik, B. (2020). Realistic expectations and pro-social behavioural intentions to the early phase of the COVID-19 pandemic in the Norwegian population. PsyArXiv. doi: 10.31234/osf.io/uptyq
Sax, H., and Clack, L. (2015). Mental models: a basic concept for human factors design in infection prevention. J. Hosp. Infect. 89, 335–339. doi: 10.1016/j.jhin.2014.12.008
Shiffrin, R. M., and Schneider, W. (1977). Controlled and automatic human information processing: II. Perceptual learning, automatic attending and a general theory. Psychol. Rev. 84, 127–190. doi: 10.1037/0033-295X.84.2.127
Sigelman, C. K. (2012). Age and ethnic differences in cold weather and contagion theories of colds and flu. Health Educ. Behav. 39, 67–76. doi: 10.1177/1090198111407187
Sinayev, A., and Peters, E. (2015). Cognitive reflection vs. calculation in decision making. Front. Psychol. 6:532. doi: 10.3389/fpsyg.2015.00532
Stanovich, K. E. (2009). “Distinguishing the reflective, algorithmic and autonomous minds: is it time for a tri-process theory?” in In two minds: Dual processes and beyond. eds. J. S. B. T. Evans and K. Frankish (Oxford: Oxford University Press), 55–58.
Stanovich, K. (2012). “On the distinction between rationality and intelligence: implications for understanding individual differences in reasoning” in The Oxford handbook of thinking and reasoning. eds. K. J. Holyoak and R. G. Morrison (Oxford: Oxford University Press), 433–455.
Stanovich, K. E., Toplak, M. E., and West, R. F. (2019). “Intelligence and rationality” in The Cambridge handbook of intelligence. ed. R. J. Sternberg (Cambridge University Press), 1106–1139.
Stanovich, K. E., and West, R. F. (1998). Individual differences in rational thought. J. Exp. Psychol. Gen. 127, 161–188. doi: 10.1037/0096-3445.127.2.161
Stanovich, K., and West, R. (2008). On the relative independence of thinking biases and cognitive ability. J. Pers. Soc. Psychol. 94, 672–695. doi: 10.1037/0022-3514.94.4.672
Steinberg, L., Sharp, C., Stanford, M. S., and Tharp, A. T. (2013). New tricks for an old measure: the development of the Barratt Impulsiveness Scale–Brief (BIS-Brief). Psychol. Assess. 25:216. doi: 10.1037/a0030550
Stupple, E. J. N., Pitchford, M., Ball, L. J., Hunt, T. E., and Steel, R. (2017). Slower is not always better: response-time evidence clarifies the limited role of miserly information processing in the Cognitive Reflection Test. PLoS One 12:e0186404. doi: 10.1371/journal.pone.0186404
Thoma, V., White, E., Panigrahi, A., Strowger, V., and Anderson, I. (2015). Good thinking or gut feeling? Cognitive reflection and intuition in traders, bankers and financial non-experts. PLoS One 10:e0123202. doi: 10.1371/journal.pone.0123202
Thomson, K. S., and Oppenheimer, D. M. (2016). Investigating an alternate form of the Cognitive Reflection Test. Judgm. Decis. Mak. 11:99.
Toplak, M. E., West, R. F., and Stanovich, K. E. (2011). The Cognitive Reflection Test as a predictor of performance on heuristics-and-biases tasks. Mem. Cogn. 39, 1275–1289. doi: 10.3758/s13421-011-0104-1
Toplak, M. E., West, R. F., and Stanovich, K. E. (2014). Assessing miserly information processing: an expansion of the Cognitive Reflection Test. Think. Reason. 20, 147–168. doi: 10.1080/13546783.2013.844729
Tversky, A., and Kahneman, D. (1974). Judgment under uncertainty: heuristics and biases. Science 185, 1124–1131. doi: 10.1126/science.185.4157.1124
Unsworth, N., and Spillers, G. J. (2010). Working memory capacity: attention control, secondary memory, or both? A direct test of the dual-component model. J. Mem. Lang. 62, 392–406. doi: 10.1016/j.jml.2010.02.001
van Doorn, R. R. A., Lang, J. W. B., and Weijters, T. (2010). Self-reported cognitive failures: a core self-evaluation? Personal. Individ. Differ. 49, 717–722. doi: 10.1016/j.paid.2010.06.013
Verkoeijen, P. P., and Bouwmeester, S. (2014). Does intuition cause cooperation? PLoS ONE 9:e96654. doi: 10.1371/journal.pone.0096654
Vom Hofe, A., Mainemarre, G., and Vannier, L.-C. (1998). Sensitivity to everyday failures and cognitive inhibition: are they related? Eur. Rev. Appl. Psychol. 48, 49–56.
Wallace, J. C., and Chen, G. (2005). Development and validation of a work-specific measure of cognitive failure: implications for occupational safety. J. Occup. Organ. Psychol. 78, 615–632. doi: 10.1348/096317905X37442
Wallace, J. C., Vodanovich, S. J., and Restino, B. M. (2003). Predicting cognitive failures from boredom proneness and daytime sleepiness scores: an investigation within military and undergraduate samples. Personal. Individ. Differ. 34, 635–644. doi: 10.1016/S0191-8869(02)00050-8
Wassenaar, A., de Reus, J., Donders, A. R. T., Schoonhoven, L., Cremer, O. L., de Lange, D. W., et al. (2018). Development and validation of an abbreviated questionnaire to easily measure cognitive failure in ICU survivors: a multicenter study. Crit. Care Med. 46, 79–84. doi: 10.1097/CCM.0000000000002806
Webster, R. K., Brooks, S. K., Smith, L. E., Woodland, L., Wessely, S., and Rubin, G. J. (2020). How to improve adherence with quarantine: rapid review of the evidence. Public Health 182, 163–169. doi: 10.1016/j.puhe.2020.03.007
Weinberg, J. (2020). Coronavirus lockdown: fresh data on compliance and public opinion. The Conversation. Available at: http://theconversation.com/coronavirus-lockdown-fresh-data-on-compliance-and-public-opinion-135872 (Accessed July 27, 2020).
West, R. F., Meserve, R. J., and Stanovich, K. E. (2012). Cognitive sophistication does not attenuate the bias blind spot. J. Pers. Soc. Psychol. 103, 506–519. doi: 10.1037/a0028857
Weston, D., Hauck, K., and Amlôt, R. (2018). Infection prevention behaviour and infectious disease modelling: a review of the literature and recommendations for the future. BMC Public Health 18:336. doi: 10.1186/s12889-018-5223-1
Woodward, A., and Gal, S. (2020). How coronavirus symptoms compare with those of the flu, allergies, and the common cold. Business Insider. Available at: https://www.businessinsider.com/coronavirus-symptoms-compared-to-flu-common-cold-and-allergies-2020-3 (Accessed July 29, 2020).
World Health Organisation (2020a). Technical Guidance Publications. Available at: https://www.who.int/emergencies/diseases/novel-coronavirus-2019/technical-guidance-publications (Accessed July 27, 2020).
World Health Organisation (2020b). Coronavirus disease (COVID-19): Similarities and differences with influenza. Available at: https://www.who.int/news-room/q-a-detail/coronavirus-disease-covid-19-similarities-and-differences-with-influenza
Xie, W., Campbell, S., and Zhang, W. (2020). Working memory capacity predicts individual differences in social-distancing compliance during the COVID-19 pandemic in the United States. Proc. Natl. Acad. Sci. 117, 17667–17674. doi: 10.1073/pnas.2008868117
Yamagishi, T., Li, Y., Takagishi, H., Matsumoto, Y., and Kiyonari, T. (2014). In search of homo economicus. Psychol. Sci. 25, 1699–1711. doi: 10.1177/0956797614538065
Zhang, D. C., Highhouse, S., and Rada, T. B. (2016). Explaining sex differences on the Cognitive Reflection Test. Personal. Individ. Differ. 101, 425–427. doi: 10.1016/j.paid.2016.06.034
Zhang, X., Wang, F., Zhu, C., and Wang, Z. (2019). Willingness to self-isolate when facing a pandemic risk: model, empirical test, and policy recommendations. Int. J. Environ. Res. Public Health 17:197. doi: 10.3390/ijerph17010197
Keywords: COVID-19, cognitive reflection, cognitive failures, risk-taking, infection precaution, planned behavior
Citation: Thoma V, Weiss-Cohen L, Filkuková P and Ayton P (2021) Cognitive Predictors of Precautionary Behavior During the COVID-19 Pandemic. Front. Psychol. 12:589800. doi: 10.3389/fpsyg.2021.589800
Edited by:
Claudia Gianelli, University Institute of Higher Studies in Pavia, ItalyReviewed by:
Caterina Primi, University of Florence, ItalyAntonio M. Espín, University of Granada, Spain
Copyright © 2021 Thoma, Weiss-Cohen, Filkuková and Ayton. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Volker Thoma, v.thoma@uel.ac.uk