- Cognitive Neuropsychology, Department of Neurology, Hannover Medical School, Hannover, Germany
Introduction: This study explores the transformative potential of digital, theory-driven, and Bayesian paradigms in neuropsychology by combining digital technologies, a commitment to evaluating theoretical frameworks, and Bayesian statistics. The study also examines theories of executive function and cognitive flexibility in a large sample of neurotypical individuals (N = 489).
Methods: We developed an internet-based Wisconsin Card-Sorting Task (iWCST) optimized for online assessment of perseveration errors (PE). Predictions of the percentage of PE, PE (%), in non-repetitive versus repetitive situations were derived from the established supervisory attention system (SAS) theory, non-repetitive PE (%) < repetitive PE (%), and the novel goal-directed instrumental control (GIC) theory, non-repetitive PE (%) > repetitive PE (%).
Results: Bayesian t-tests revealed the presence of a robust error suppression effect (ESE) indicating that PE are less likely in repetitive situations than in non-repetitive situations, contradicting SAS theory with posterior model probability p < 0.001 and confirming GIC theory with posterior model probability p > 0.999. We conclude that repetitive situations support cognitive set switching in the iWCST by facilitating the retrieval of goal-directed, instrumental memory that associates stimulus features, actions, and outcomes, thereby generating the ESE in neurotypical individuals. We also report exploratory data analyses, including a Bayesian network analysis of relationships between iWCST measures.
Discussion: Overall, this study serves as a paradigmatic model for combining digital technologies, theory-driven research, and Bayesian statistics in neuropsychology. It also provides insight into how this integrative, innovative approach can advance the understanding of executive function and cognitive flexibility and inform future research and clinical applications.
1 Introduction
Neuropsychology needs innovation (Casaletto and Heaton, 2017; Loring et al., 2018; Parsons et al., 2018; Bilder and Reise, 2019; Kessels, 2019). In this article, we propose three main pillars for the necessary transformation: digitization, theory-driven research, and Bayesian statistics. The empirical study presented here illustrates how these three innovative trends can be combined to advance neuropsychology. It provides an example of how the incorporation of digital tools; the explicit formulation of hypotheses based on existing theories and the use of these theories to guide assessment design, data collection, and analysis; and the application of Bayesian methods to the statistical analysis of neuropsychological data provide opportunities for the future development of the discipline.
Digital approaches to neuropsychology, including computerized cognitive assessments, especially unsupervised online assessments, are a growing trend in neuropsychological research (Feenstra et al., 2017; Germine et al., 2019; Parsons and Duffield, 2020; Libon et al., 2022). These tools have the potential to provide measures of cognitive function in a variety of settings, and demonstrating their psychometric quality in terms of validity, reliability, and diagnostic utility is an important step in the digital transformation that also offers new opportunities for standardized neuropsychological testing (Brooks et al., 2009; Kiselica et al., 2023). In the present study, we present an internet-based version of the well-known Wisconsin Card Sorting Test (WCST; originally developed by Berg, 1948, and Grant and Berg, 1948), which we refer to as the internet-based Wisconsin Card-Sorting Task (iWCST), and which is described in detail in the Materials and Methods section below (see also Figure 1). Standardized (Heaton, 1981; Heaton et al., 1993; Kongs et al., 2000) and computerized (e.g., Barceló, 2003; Heaton and PAR Staff, 2003a,b; Kopp and Lange, 2013) versions of the WCST, as well as a variety of additional WCST versions (Modified Card Sorting Test, MCST, Nelson, 1976; Modified Wisconsin Card Sorting Test, M-WCST, Schretlen, 2010), are widely used in clinical and research settings. Please note that the computerized versions of the WCST (Heaton and PAR Staff, 2003a,b) are also available as a commercial product for online administration on PARiConnect.1
Figure 1. The figure on the left shows the typical layout of computerized Wisconsin card-sorting tasks. Wisconsin stimulus cards can be described in a three-dimensional space (with the dimensions COLOR/SHAPE/NUMBER of the items represented). Each dimension is instantiated by one of four features (with COLOR features: red, green, yellow, blue/SHAPE features: triangle, asterisk, cross, circle/NUMBER features: #1, #2, #3, #4). On each trial, an action is required that matches the current target card to one of four consistently presented keycards (i.e., outer left keycard: ‘1 red triangle’, inner left keycard: ‘2 green asterisks’, inner right keycard: ‘3 yellow crosses’, outer right keycard: ‘4 blue circles’). The task is to select, on each trial, the keycard that shares the feature with the target feature to be prioritized (be it the COLOR, SHAPE, or NUMBER feature). This is illustrated on trial 1 by selecting the left-most keycard ‘1 red triangle’, which shares the SHAPE feature ‘triangle’ with the current target card, by pressing the Y key. The feedback CORRECT indicates that sorting by SHAPE on trial 1 was correct. Trial 2 shows the selection of the inner left keycard ‘2 green asterisks’, which shares the SHAPE feature ‘asterisk’ with the target card, by pressing the C key. However, the feedback INCORRECT indicates that sorting by SHAPE was no longer correct on trial 2; this feedback is therefore an informative error signal. Trials 3 and 4 illustrate perseveration errors (PE) in repetitive vs. non-repetitive situations. Trial 3 shows a repetitive PE in which, despite the informative error signal obtained on trial 2, sorting by SHAPE is applied, and the action-relevant SHAPE feature (asterisks on trials 2 and 3) is repeated on the two successive trials (hence the label repetitive PE or rPE). Trial 4 shows a non-repetitive PE, where sorting by SHAPE is applied despite the informative error signal obtained on trial 3, but the action-relevant SHAPE feature (asterisk on trial 3, circle on trial 4) changes between the two successive trials (hence the label non-repetitive PE or nPE).
All of these different WCST versions are commonly used to assess executive function and cognitive flexibility in individuals with neurological diseases and psychological disorders. Cognitive flexibility, the ability to switch between different cognitive sets, is a topic of central interest in neuropsychology (see Kopp, in press, for a review). Milner (1963) pioneered this extensive line of WCST-based neuropsychological research on cognitive flexibility, which has been the subject of several reviews and meta-analyses (e.g., Demakis, 2003; Nyhus and Barceló, 2009; Lange et al., 2017, 2018; Miles et al., 2021).
Theory-driven approaches remain critical in science because they provide a framework to guide hypothesis formulation and study design, to better understand data, and to ensure that analyses are theoretically relevant. Theory falsification remains important to scientific inquiry, underscoring its central role in advancing scientific understanding (Popper, 1934/35). However, theory-driven investigations in neuropsychology can be difficult when theoretical frameworks lack precise quantitative predictions that can be rigorously tested. The present study illustrates a method for subjecting neuropsychological theories to falsifiability tests, even in the absence of quantitative predictions, with the goal of providing a way to ultimately distill the most valid theoretical framework.
The Supervisory Attentional System (SAS) theory explains how executive control operates in goal-directed tasks (Norman and Shallice, 1981; Shallice, 1982; Shallice and Burgess, 1991, 1996; Shallice and Cooper, 2011). The SAS plays a critical role in managing and coordinating lower-level cognitive processes, called schemas, to achieve higher-order goals. The SAS is responsible for supervisory control, intervening when the schemas appear inadequate to achieve a goal, for example, after errors and in non-routine situations. Impairments in the SAS can manifest as cognitive inflexibility, which affects an individual’s ability to adapt to changing environmental demands, presumably due to a difficulty in shifting cognitive sets.
As shown in Figure 1, the WCST asks individuals to sort cards according to changing dimensions, such as color, shape, or number. Perseveration errors (PE) occur when individuals continue to apply a dimension that was previously correct but has since become incorrect. PE are thus a behavioral expression of a difficulty in adapting to new environmental demands, presumably indicating an inability to shift cognitive sets (Kopp, in press). From the perspective of SAS theory, the SAS, as an integral part of executive control, is critically involved in monitoring and adjusting cognitive sets. Impairments in the SAS may therefore increase PE percentages in the WCST, highlighting associations between SAS impairments and deficits in cognitive flexibility.
In recent years, we have identified two subtypes of PE that differ in their prevalence, a phenomenon referred to as the error suppression effect (ESE; Kopp et al., 2019, 2023; Steinke et al., 2020a). Within this framework, two distinct types of PE have been recognized: those that occur in non-repetitive situations (referred to as non-repetitive PE or nPE throughout this article) and those that occur in repetitive situations (referred to as repetitive PE or rPE throughout this article), as shown in Figure 1. As can be seen, non-repetitive PE (nPE) occur when the same incorrect response strategy is used despite the provision of informative feedback indicating the need for change, but the relevant feature changes between successive trials. Repetitive PE (rPE) occur when the same incorrect response strategy is repeated despite the provision of this type of informative feedback, but the relevant feature remains constant across successive trials. Importantly, repetitions can be viewed as retrieval cues, as will be discussed later. Non-repetitive situations, which involve changes in both action-relevant features and actions, are non-routine conditions. In contrast, repetitive situations, which involve repetition in both action-relevant features and actions, are routine conditions. Contrary to the prediction of the SAS theory that non-repetitive PE should be less frequent than repetitive PE due to a stronger involvement of the SAS in non-routine situations, our studies led to the identification of the ESE where repetitive PE are less frequent than non-repetitive PE. Initially discovered in a sample of neurological inpatients (Kopp et al., 2019), subsequent studies confirmed the ESE in neurotypical individuals, with estimated effect sizes (Cohen’s d) of 0.53 (Steinke et al., 2020a; N = 375) and 0.50 (Kopp et al., 2023, Study 1; N = 40).
In the aforementioned investigation (Kopp et al., 2023), which involved multiple experiments, we found that goal-directed instrumental learning (Dickinson, 1994; De Wit and Dickinson, 2009; Balleine and O’Doherty, 2010) plays a central role in generating the ESE. We developed the theory of goal-directed instrumental control (GIC), according to which repetitive conditions for PE facilitate the retrieval of goal-directed instrumental memory. These memory traces include associative bindings between action-relevant features, actions and their outcomes on recent WCST trials. As a result of retrieving these goal-directed, instrumental memory traces, which on switch trials include bindings between action-relevant features, actions, and negative outcomes (i.e., error feedback), repetitive PE should be less frequent than non-repetitive PE.2 Taken together, the manifestation of perseverative behavior in the WCST may shed light on the interplay between cognitive flexibility and routinization, as affected by repetition, and on the explanatory power of the two neuropsychological theories under consideration (i.e., SAS theory and GIC theory), as illustrated in Figure 2.
Figure 2. Illustration of the supervisory attentional system (SAS) theory and the goal-directed instrumental control (GIC) theory. The top graphs show the relative engagement (y-axis) of the conceptualized transient control processes as a function of time (x-axis) following an informative error signal. The figure on the left shows that SAS theory suggests that the supervisory control system is more involved in non-repetitive, relatively non-routinized than in repetitive, relatively routinized situations, with the effect that set repetition is more strongly inhibited in non-repetitive than in repetitive situations, leading to the prediction non-repetitive PE (%) < repetitive PE (%). The figure on the right shows that GIC theory proposes dissociable pathways to set inhibition, with feedback-based control activated equally strongly by informative error signals in both non-repetitive and repetitive situations. Retrieval-based control comes into the play primarily on repetitive trials, where feature repetition (see Figure 1) facilitates set switching through retrieving goal-directed, instrumental memory traces that associate action-relevant features, actions, and outcomes. The involvement of this pathway on repetitive trials leads to the prediction of an error suppression effect (ESE), i.e., non-repetitive PE (%) > repetitive PE (%). See Figure 11 for more details on the GIC theory. Arrowheaded lines (↑) indicate facilitative effects, and T-shaped lines (T) indicate inhibitory effects.
There is a growing recognition in psychology that traditional statistical methods, such as null hypothesis significance testing (NHST), may need to be revised. NHST evaluates the significance of observed effects by comparing data to a predetermined threshold (e.g., α = 0.05), generating a p-value that indicates the likelihood of results under the null hypothesis. However, NHST has been criticized for its rigid thresholds, binary outcomes, potential misinterpretation of p-values, and selective reporting (e.g., Meehl, 1978; Szucs and Ioannidis, 2017).
Bayesian statistical methods (Gelman et al., 2013; Gelman and Shalizi, 2013; Kruschke, 2014) have gained popularity in psychology and neuroscience as an alternative approach. Bayesian methods integrate prior knowledge with observed data, allowing for a cumulative evaluation of evidence (Rozeboom, 1960; Benjamin et al., 2018; Wagenmakers et al., 2018a,b; van de Schoot et al., 2021). Bayesian analysis generates probability distributions that provide metrics such as posterior means or credible intervals. In addition, Bayesian hypothesis testing evaluates posterior model probabilities, which helps to directly assess the strength of theories. In this sense, Bayesian statistics is more in line with Popper’s (1934/35) approach to science.
In summary, we have identified three key areas of innovation in neuropsychology: the proliferation of online assessment tools, the turn to theory-driven research, and the shift from frequentist NHST to Bayesian statistics. The present study illustrates how these three innovative trends can be combined to advance neuropsychology. Using the unsupervised online assessment of perseverative behavior on the iWCST, we analyzed the evidence for conflicting predictions about PE derived from two neuropsychological theories, with the goal of rejecting the theory that received less empirical support. As explained above, the manifestation of PE on the iWCST may serve to elucidate the interplay between cognitive flexibility and routinization, as affected by repetition. SAS theory, which focuses on goal-directed control of cognitive processes, predicts more frequently occurring PE in routine situations compared to non-routine situations, i.e., non-repetitive PE (%) < repetitive PE (%). GIC theory, which focuses on retrieval of goal-directed, instrumental memory traces, predicts less frequently occurring PE in repetitive compared to non-repetitive situations, i.e., non-repetitive PE (%) > repetitive PE (%). While the two theories under consideration do not produce numerical point values, they do provide ordinal, informative hypotheses (Hoijtink, 2011). Specifically, HSAS, derived from SAS theory, predicts nPE% < rPE%, whereas HGIC, derived from GIC theory, predicts nPE% > rPE%. Because the Bayesian statistical framework offers the aforementioned advantages over NHST, we used a Bayesian statistical method called BAIN (BAyesian INformative hypotheses evaluation; Hoijtink et al., 2019) to formally assess the validity of these two theories of the dynamic modulation of cognitive flexibility in light of the observed sample evidence.
The decision to test only GIC versus SAS is based on two considerations. First, SAS is the most established theory and provides a robust benchmark, while GIC is a novel theory developed specifically for this phenomenon. Comparing these two theories allows for a rigorous evaluation of the validity of the GIC theory against the gold standard. Second, by first comparing GIC with SAS and then testing the prevailing theory against others, we envision a thorough evaluation process that integrates insights from multiple theories over time, provided that appropriate diagnostic experimental designs are developed. This strategy of successive pairwise theory tests aims to systematically narrow down to the most accurate theory.
2 Materials and methods
Study preparation included the following components: (1) programming the behavioral card-sorting task (iWCST), (2) setting up the online server to host the study, including data storage, and (3) recruiting participants for the study. The schematic workflow of the process of bringing the iWCST online as part of the present study is presented in Supplementary Figure S1.1.
2.1 Inviting study participants
Supplementary Figure S1.1 shows that the researchers approached several inviting organizations, such as student offices and student associations, at a number of universities in the southeastern part of the German state of Lower Saxony. The participating universities include the Hannover Medical School (MHH), the Hannover University of Music, Drama and Media (HMTMH), the Hannover University of Veterinary Medicine (TiHo), and the psychology departments of the Technical University of Braunschweig (TU BS) and the University of Hildesheim (U HI).
The collaborating inviting organizations distributed emails prepared by the researchers to their students (see Supplementary Figure S1.2). It contained a hyperlink that connected participants with the server of the online study. Table 1 clearly shows that the invitation emails were more or less successful in terms of the number of participants that could be recruited. Overall, email distribution through official faculty channels (student offices, e.g., MHH) seemed to be much more effective than distribution through more informal channels (student associations, e.g., U HI).
Table 1. Number of participants who opened the study hyperlink, number of early and late dropouts, and number of complete and valid records per participating university.
Table 1 shows that a total of N = 826 students opened the study hyperlink; we refer to them as neurotypical individuals because they are presumed to have no neurological conditions. Of these 826 individuals, 503 individuals (N = 359 females/N = 143 males; one preferred not to say) completed the data collection fully (approximately 60.9%). Of the 323 individuals who dropped out, the vast majority (289 individuals, or approximately 89.5 percent of all dropouts) decided not to continue during the task instructions (early dropouts), while only 34 individuals (or approximately 10.5 percent of all dropouts) actually started but did not complete the data collection (late dropouts). Of these 34 late dropouts, 31 did not fully complete the iWCST, while the remaining three late dropouts completed the iWCST fully but did not complete the questionnaires.
Data quality was very good among those who completed the iWCST and questionnaires in full, as only 14 out of 503 records (i.e., no more than approximately 2.8%) had to be excluded based on our pre-specified criteria for iWCST data quality (see below). A total of N = 489 participants (N = 350 females/N = 139 males) provided complete records that fulfilled the inclusion criteria (see below).
The general study instructions are documented in Supplementary Figure S2.1. The basic framing of the study objective was the assessment of “ability to concentrate.” Participants were informed that the task would take approximately half an hour, and that the validity of the study required that they perform the task alone in a quiet environment to allow for undisturbed task performance, without any help of others.
Participants had the opportunity to receive a financial compensation of 10€, which could be claimed by sending an email to the researchers as shown in Supplementary Figure S1.1. A unique 8-digit alpha-numeric code generated by each individual participant for anonymous identification was checked against the records to ensure actual study participation.
2.2 Instruments
The online study consisted of the iWCST and a series of demographic questions/psychological questionnaires. These two study components were imported separately into the open source software JATOS 3.8.5 (Lange et al., 2015). Once combined, the MindProbe server3 hosted the study and stored data.
JATOS creates several types of hyperlinks. We used the ‘general single’ hyperlink type, which can be distributed to an unlimited number of recipients at once (hence the qualifier ‘general’). However, this general hyperlink can only be used once in the same browser (hence the qualifier ‘single’), thus restricting participants from participating in the study more than once. Each participating university was given a unique ‘general single’ hyperlink, which allowed us to determine the number of participants from each participating university, as shown in Table 1. Details of this study workflow are shown in Supplementary Figure S1.1.
2.2.1 Questionnaires
The demographic questions (age, years of education, and gender) and four questionnaires were embedded in JATOS as HTML files. Table 2 shows the corresponding data. Handedness was assessed using the four-item short form of the Edinburgh Handedness Inventory (EHI-SF; Veale, 2014). The 10-item short form of the Big Five Inventory was used to assess extroversion, conscientiousness, openness, agreeableness, and neuroticism (BFI-10; Rammstedt et al., 2014). Participants also completed the 11-item short form of the Behavioral Inhibition System/Behavioral Approach Inventory (BIS/BAS-11; Studer et al., 2016). Depression and anxiety were assessed using the four-item short form of the Physical Health Questionnaire (PHQ-4; Löwe et al., 2010).
2.2.2 Internet-based Wisconsin card-sorting task
The iWCST was programmed using OpenSesame version 4.0 and OSWeb, which is the JavaScript implementation of OpenSesame (Mathôt and March, 2022). The OpenSesame/OSWeb iWCST program was imported into JATOS 3.8.5 (Lange et al., 2015) installed on the MindProbe server (see text footnote 3, respectively). The authors will provide access to the iWCST program to interested researchers upon reasonable request, preferably by providing additional hyperlinks. These hyperlinks will be sufficient for conducting additional studies using the iWCST in collaboration with the authors.
The iWCST is closely modeled after the M-WCST (Schretlen, 2010), which is an established, commercially available WCST variant. The M-WCST, in turn, shares many features with Nelson’s (1976) MCST variant, such as using only the 24 unambiguous target cards and changing dimensions after only six consecutive correct sorts. Dimensions had to be performed in a fixed order throughout the iWCST, i.e., COLOR, SHAPE, NUMBER, COLOR, SHAPE, etc., in contrast to the standard M-WCST and MCST instructions, but consistent with Heaton’s (1981) and Heaton et al.’s (1993) variant of the WCST.
As shown in Supplementary Figure S1.3, participants match target cards on a trial-by-trial basis to one of the four consistently appearing keycards. All stimuli are presented on a gray background. On each trial, participants encounter four keycards that appear along the horizontal axis at consistent spatial locations on the computer screen. The card-sorting responses can be based on any of the three dimensions of COLOR, SHAPE, and NUMBER of the objects depicted. Each target card selected from the set of 24 unambiguous target cards can be assigned to three different keycards based on the COLOR, SHAPE, or NUMBER of the objects depicted; hence the qualifier ‘unambiguous’. On each trial, the selection of the fourth keycard represents an odd error (OE) because it has no common feature with the target card.
The iWCST differs from the manual versions of the WCST in four main ways. First, participants in the iWCST respond by pressing the Y, C, B, or M keys on a QWERTZ keyboard (their spatial arrangement corresponds to Z, C, B, or M on a QWERTY keyboard). This keyboarding requires the use of desktop or laptop computers, excluding the use of touchscreen devices such as tablets or smartphones. Supplementary Figure S1.3 shows how these four letter keys map to the four keycards. To facilitate stimulus–response mapping, the keycards are presented together with their corresponding letter keys.
Second, while manually sorted WCST target cards remain visible for longer periods of time, iWCST target cards disappear after trial-by-trial feedback, as shown in Figure 1 and described below. Each trial begins with the presentation of a target card underneath the keycards, followed by the participant’s response via key press. Upon response, the target card visually moves from its original position to under the selected keycard. Participants then receive feedback indicated by the words “correct” or “incorrect” displayed for 1,200 milliseconds. The target card and feedback then disappear, followed by an 800-millisecond intertrial interval before the next trial begins. These relatively short temporary appearances, which are unique to the iWCST, place a relatively higher demand on working memory than traditional manual versions of the WCST (Lange et al., 2016).
A third major departure from manual card-sorting tasks concerns the trial-by-trial sequence of target cards. While manual WCST versions typically follow a predetermined order, the iWCST uses a pseudo-random selection of target cards from the set of 24 different target cards. This pseudo-random order ensures that the same target card never appears on two consecutive trials. The pseudo-random order also maintains a 50% chance of encountering non-repetitive versus repetitive opportunities to commit PE. This guarantees an equal chance of encountering trials in which the action-relevant stimulus features are repeated or not repeated. Thus, the pseudo-random sequence provides an equal opportunity for non-repetitive and repetitive PE occurrences. Note that the distinction between non-repetitive and repetitive PE excludes exact repetitions of target cards on successive trials and focuses on the alternation or repetition of current action-relevant features on trials following informative error feedback as shown in Figure 1.
Fourth, the iWCST is adapted to optimize the assessment of PE, which is a primary rationale for the study. The termination condition in the iWCST sets a maximum of 240 trials, but allows for early termination if the participant encounters 20 trials each of potential non-repetitive and repetitive PE before reaching this limit. This approach strikes a balance between obtaining adequate data for reliable cognitive assessment (Kopp et al., 2021) and preventing undue prolongation of the test or participant fatigue. The iWCST’s flexibility in stopping the assessment optimizes duration while ensuring sufficient data collection.
Table 3 shows the resulting number of trials and the number of trials for potential non-repetitive and repetitive PE. Inspection of Table 3 shows that the average number of trials is more than 200 trials, and that the average number of opportunities to perform non-repetitive and repetitive PE is slightly more than 20 trials each. The latter aspect of the iWCST is important because it allows the main measures, non-repetitive PE (%) and repetitive PE (%), to be assessed with comparable psychometric quality. To date, it has not been possible to achieve psychometrically matched measures of non-repetitive and repetitive PE with manual card-sorting tasks such as the M-WCST (Schretlen, 2010), due to the inherent imbalance between opportunities for non-repetitive and repetitive PE (Kopp et al., 2019).
Participants receive detailed task instructions prior to the start of the iWCST. The formulation of appropriate task instructions is very important because there is no other way to comprehensively instruct the card-sorting task in the context of an unsupervised online assessment. Supplementary Figures S2.2–S2.23 provide the complete iWCST instructions, which include a general description of the sorting task at hand, interactive elements to ensure that participants have a good understanding of the basics of the card-sorting task, and information about occasional dimension changes. Supplementary Figures S2.1–S2.22 provide detailed information on how these dimension changes were instructed in the iWCST. Note that the iWCST instructions never explicitly mention the three dimensions underlying the sorting of the cards, i.e., the COLOR, SHAPE, and NUMBER of objects depicted, leaving room for cognitive processes involved in categorizing information. Thus, the iWCST requires abstraction and concept formation, much like the original WCST. Abstraction involves extracting general properties from sensory input, while conceptualization goes beyond direct sensory experience and involves the formation of abstract concepts for higher-level organization of information.
The iWCST is administered immediately following the completion of the task instructions. The average time to complete the iWCST and questionnaires is less than 30 min (Mdn = 22.65 min, IQR = 7.15 min).
2.2.3 Internet-based Wisconsin card-sorting task measures
As shown in Table 3, the iWCST measures include the number of trials completed, the number of categories achieved, where a category is defined as six consecutive correct sorts, and the percentage of categories among the number of completed trials. PE are the main measures of the iWCST, notably non-repetitive PE (%) and repetitive PE (%). Therefore, the iWCST records the number of occasions for committing PE, non-repetitive PE, and repetitive PE, as well as the actual number of PE, non-repetitive PE, and repetitive PE, as defined in Figure 1, allowing the percentage of PE, non-repetitive PE, and repetitive PE to be calculated.
Figure 3 shows two additional types of errors, namely integration errors (Lange et al., 2016) and set-loss errors (Kopp, in press). An integration error (IE) occurs after receiving informative error feedbacks on two consecutive iWCST trials that inform about the current incorrectness of two dimensions (such as the second and third trials in Figure 3). Successful integration of the previous two informative error feedbacks results in the selection of the remaining viable dimension, whereas faulty integration of the previous two informative error feedbacks results in the repetition of the dimension excluded by the first informative error feedback. In the example in Figure 3, sorting by SHAPE on Trial 4 is an IE because the SHAPE dimension was already eliminated by the informative error feedback on Trial 2. Successful integration of the previous two informative error feedbacks would have eliminated both the SHAPE and COLOR dimensions, resulting in sorting by NUMBER on Trial 4. An example of successful integration is shown on Trial 5, where sorting by NUMBER correctly follows from integrating the previous two informative error feedbacks, which eliminated both the SHAPE and COLOR dimensions.
Figure 3. This figure illustrates the typical layout of computerized Wisconsin card-sorting tasks and two additional types of errors, i.e., integration errors and set-loss errors. Each row represents a new trial on which participants must match a trial-specific target card with one of the four consistently appearing keycards. After applying the dimension SHAPE on the first trial and receiving the CORRECT feedback, the sorting by the dimension SHAPE is repeated on trial 2. However, the INCORRECT informative error signal indicates that the relevant dimension has changed. Trial 3 shows the application of the dimension COLOR, but the INCORRECT feedback indicates that COLOR is also currently incorrect. Trial 4 shows the application of the dimension SHAPE, even though informative error signals have already been issued regarding SHAPE (on trial 2) and COLOR (on trial 3), illustrating an integration error (after excluding SHAPE and COLOR, NUMBER remains the only viable dimension). Trial 5 shows the application of the dimension NUMBER, and the CORRECT feedback indicates that NUMBER is indeed the current correct dimension. However, trial 6, which shows the application of the dimension SHAPE, illustrates a set-loss error because the dimension NUMBER was correctly applied on the previous trial.
Set-loss errors (SLE) are comparatively simple types of iWCST errors, as shown on Trial 6 in Figure 3. SLE occur when participants fail to maintain a cognitive set on an iWCST trial despite receiving feedback that sorting according to that set was correct on the previous trial. For both types of errors, IE and SLE, we calculated the number of occasions for them, as well as their actual number, which in turn allowed us to calculate the percentage of IE and SLE (Table 3). Finally, the number of OE was calculated.
2.3 Data analysis
Data preprocessing was performed in Microsoft Excel 2016 (Microsoft Corporation, 2016). Participants with complete records were excluded if they scored more than three standard deviations (a) below the average number of categories achieved, or (b) above the average number of OE (see also Steinke et al., 2020a). In both cases, the presence of these unrepresentative measures may indicate insufficient effort on the task. The application of this criterion resulted in the exclusion of 12 participants. In addition, two participants were excluded because, contrary to our intent, they had fewer than eight occasions of either non-repetitive or repetitive PE. Inspection of Table 1 shows that only 14 out of 503 records had to be excluded based on these criteria, leaving a sample of N = 489 valid records.
Bayesian data analysis was performed using JASP version 0.18.3.4 Confirmatory data analyses included Bayesian paired samples t-tests (Faulkenberry et al., 2020), which provided informal evidence regarding the fit of SAS theory and GIC theory. We finally conducted Bayesian informative hypotheses evaluation via BAIN paired-samples t-tests, implemented in the JASP-BAIN module (Hoijtink et al., 2019), to calculate posterior model probabilities of hypothesis HSAS, derived from SAS theory and predicting nPE% < rPE%, and of hypothesis HGIC, derived from GIC theory and predicting nPE% > rPE%.
Exploratory data analyses included Bayesian correlation and network (Huth et al., 2023) analyses to examine relationships between the iWCST measures. Bayesian regression analyses (van den Bergh et al., 2021) examined relationships between the iWCST measures (as dependent variables) and questionnaire scores (as covariates).
3 Results
3.1 Descriptive iWCST results
Table 3 shows the descriptive sample statistics for all of the iWCST variables for the participants who were included in the final analysis (N = 489). Participants completed a mean of 206.3 iWCST trials, ranging from 90 to 240 trials. Participants achieved a mean of 24.0 iWCST categories, ranging from 7 to 34 categories.
For iWCST error scores, the description of the sample statistics for PE is deferred to the confirmatory analyses. Participants made an average of 6.6 iWCST SLE, ranging from 0 to 24 SLE. The number of SLE occasions averaged 137.3 iWCST trials, ranging from 41 to 187 trials. This resulted in an average of 5.1 SLE%, ranging from 0 to 19.6 percent. Participants made an average of 1.9 iWCST IE, ranging from 0 to 10 IE. The number of IE occasions averaged 11.2 iWCST trials, ranging from 0 to 24 trials. This resulted in an average of 14.9 IE%, ranging from 0 to 66.7 percent. Participants made an average of 0.9 iWCST OE, ranging from 0 to 11 OE.
3.2 Confirmatory Bayesian data analysis
Table 3 shows that participants made an average of 2.9 iWCST PE, ranging from 0 to 14 PE. The number of PE occasions averaged 42.9 iWCST trials, ranging from 18 to 63 trials. Regarding the theoretical distinction between non-repetitive and repetitive PE, the sample statistics are that participants committed an average of 2.0 iWCST nPE, ranging from 0 to 11 nPE. The number of nPE occasions averaged 21.5 iWCST trials, ranging from 10 to 43 trials. This resulted in an average of 9.2 nPE%, ranging from 0 to 45.0 percent. Participants committed an average of 0.9 iWCST rPE, ranging from 0 to 8 rPE. The number of rPE occasions averaged 21.4 iWCST trials, ranging from 8 to 35 trials. This resulted in an average of 3.9 rPE%, ranging from 0 to 31.3 percent. The iWCST PE sample data, the main target of the present study, are also visualized separately for nPE and rPE in Figure 4.
Figure 4. Individual percentages of non-repetitive PE (nPE; in yellow) and repetitive PE (rPE; in orange) are shown on the left. Box and raincloud plots of sample statistics of nPE% and rPE% are shown on the right.
If δ is the standardized difference between nPE% and rPE%, then δ provides a measure of the effect size, indicating how large the difference between the two error percentages is relative to the variability within the data. Both theories under consideration predict that nPE% and rPE% will be different in the neurotypical population, with the SAS theory predicting δ < 0 and the GIC theory predicting δ > 0. Note that the number of opportunities to commit either nPE or rPE was nearly identical, as discussed in detail above, with nPE opportunities averaging 21.5 iWCST trials and rPE opportunities averaging 21.4 iWCST trials (see Table 3). Importantly, therefore, the possible existence of a population difference cannot be attributed to the confounding factor that there was a difference in the number of iWCST trials in which non-repetitive PE and repetitive PE were possible.
The inferential plot, including the default prior distribution and the obtained posterior distribution, is shown in Figure 5. We used a Bayesian paired samples t-test to test the credibility of two models of nPE% and rPE% in the neurotypical population (H0 (δ = 0) vs. H1 (δ ≠ 0)). For a two-tailed test of H0 (δ = 0) vs. H1 (δ ≠ 0), we used JASP’s default settings, i.e., a two-parameter Cauchy prior distribution (with (central) location δ = 0, width r = 1/√2 = 0.707), as shown by the dashed line in the graph in Figure 5. The solid line in the graph in Figure 5 shows the obtained posterior distribution.
Figure 5. Inferential plot depicting the (default) prior distribution and the obtained posterior distribution (N = 489). 95% CI = 95 percent credible interval.
We ran the test to index support for H1 over H0 and found a log Bayes factor Log(BF10) = 73.34, indicating decisive (Jeffreys, 1939) evidence in favor of H1 (δ ≠ 0). Note that the y-coordinate at δ = 0 in the posterior distribution is smaller than that of the prior distribution at δ = 0. This corresponds to the fact that our posterior belief in H0 (δ = 0) has decreased after observing the data. In fact, it has decreased by the obtained log Bayes factor of Log(BF10) = 73.34. This log Bayes factor can be represented as the ratio of the prior to the posterior on the population effect δ under H0 (Faulkenberry et al., 2020).
Figure 5 also shows estimates for the effect size parameter δ, with a median effect size of δ = 0.604, indicating that the more probable population values of δ are centered around δ = 0.604, and a 95% credible interval between 0.508 and 0.701, providing a measure of the uncertainty in the size of the population effect δ after observing the data. In other words, under the assumption that H0 is false (i.e., that δ ≠ 0), there is a 95% probability that the true value of the population effect δ falls between 0.508 and 0.701 (Faulkenberry et al., 2020). This inference is clearly inconsistent with the prediction of SAS theory (δ < 0), but it is consistent with the prediction of GIC theory (δ > 0).
The left panel of Figure 6 shows a Bayesian robustness check. The plot shows that the log Bayes factor is robust to changes in the width parameter r of the Cauchy prior distribution, with all results Log(BF10) ≥ 73.02. The invariance of the log Bayes factor to variations in the width of the Cauchy prior (max log Bayes factor, user-defined, wide and ultrawide width) indicates that the results are robust in that they do not depend in an important way on our particular user-defined r = 0.707. The right panel of Figure 6 shows a Bayesian sequential analysis. The plot shows increases in the log Bayes factor as data from more and more participants are added under the user-defined, wide, and ultra-wide widths of the Cauchy prior.
Figure 6. Bayesian robustness check (on the left) and Bayesian sequential analysis (on the right) for the N = 489 participants.
After excluding N = 59 participants who did not perform any PE, N = 430 participants who performed one or more PE remained for analysis. In this subsample, an average of 3.2 iWCST PE was found, ranging from 1 to 14 PE. The number of PE occasions averaged 43.2 iWCST trials, ranging from 27 to 63 trials. Regarding non-repetitive PE and repetitive PE, participants committed an average of 2.3 iWCST nPE, ranging from 0 to 11 nPE. The number of nPE occasions averaged 21.6 iWCST trials, ranging from 10 to 43 trials. This resulted in an average of 10.5 nPE%, ranging from 0 to 45.0 percent. Participants committed an average of 1.0 iWCST rPE, ranging from 0 to 8 rPE. The number of rPE occasions averaged 21.6 iWCST trials, ranging from 12 to 35 trials. This resulted in an average of 4.5 rPE%, ranging from 0 to 31.3 percent. The iWCST PE sample data for N = 430 are visualized separately for nPE and rPE in Figure 7.
Figure 7. Individual percentages of non-repetitive PE (nPE; in yellow) and repetitive PE (rPE; in orange) are shown on the left after exclusion of N = 59 PE-free individuals, leaving N = 430 participants. Box and raincloud plots of sample statistics of nPE% and rPE% are shown on the right.
The inferential plot, including the default prior distribution and the obtained posterior distribution, is shown in Figure 8. We again used a Bayesian paired samples t-test to test the credibility of the two models of nPE% and rPE% in N = 430 participants. The Cauchy prior distribution is shown again by the dashed line in the graph in Figure 5, while the solid line shows the obtained posterior distribution.
Figure 8. Inferential plot depicting the (default) prior distribution and the obtained posterior distribution (N = 430). 95% CI = 95 percent credible interval.
We ran the test to index support for H1 over H0 and found a log Bayes factor Log(BF10) = 75.23, again indicating decisive (Jeffreys, 1939) evidence in favor of H1 (δ ≠ 0). Figure 5 also shows that the median effect size is δ = 0.661, with a 95% credible interval between 0.557 and 0.765. Again, assuming that H0 is false (i.e., that δ ≠ 0), there is a 95% probability that the true value of the population effect δ falls between 0.557 and 0.765 (Faulkenberry et al., 2020), which is again inconsistent with what SAS theory predicts (δ < 0), but consistent with what GIC theory predicts (δ > 0).
The Bayesian robustness check on left panel of Figure 9 shows that the log Bayes factor is again robust to changes of the width of the Cauchy prior, with all results Log(BF10) ≥ 74.97. The right panel of Figure 6 shows a Bayesian sequential analysis, showing increases in the log Bayes factor as data from more and more participants are added under the user-defined, wide, and ultra-wide widths of the Cauchy prior.
Figure 9. Bayesian robustness check (on the left) and Bayesian sequential analysis (on the right) for N = 430 participants.
Taken together, the results of these confirmatory Bayesian data analyses are remarkably clear. First, we obtained conclusive evidence against H0 (δ = 0) that there is no difference between non-repetitive and repetitive PE percentages in the neurotypical population. This conclusion could be drawn both from the analysis of the full sample and from the subsample of participants who committed one or more PE. Second, assuming that H0 is false (i.e., that δ ≠ 0), the 95% probability that the true value of the population effect δ falls between 0.508 and 0.701 (the central tendency equals 0.604 based on the full sample) or 0.557 and 0.765 (the central tendency equals 0.661 based on the subsample). In both cases, the results are inconsistent with what SAS theory predicts (δ < 0), but consistent with what GIC theory predicts (δ > 0).
The formal assessment of these two theories was accomplished by BAIN Bayesian paired-samples t-tests (Hoijtink et al., 2019). BAIN analyses allow direct comparison of the posterior model probabilities of HSAS, derived from SAS theory and predicting nPE% < rPE%, and of HGIC, derived from GIC theory and predicting nPE% > rPE%. Based on the full sample (N = 489), and assuming equal prior model probabilities, the log Bayes factor in favor of the GIC theory was Log(BFGIC-SAS) = 93.65, corresponding to posterior model probabilities of p > 0.999 for the GIC theory and p < 0.001 for the SAS theory. Based on the restricted sample (N = 430), again assuming equal prior model probabilities, the log Bayes factor in favor of the GIC theory was Log(BFGIC-SAS) = 98.47, corresponding to posterior model probabilities of p > 0.999 for the GIC theory and p < 0.001 for the SAS theory. Therefore, based on the observed data, the GIC theory is probably (even almost certainly) correct in predicting the ESE, while the SAS theory is probably (even almost certainly) incorrect in describing cognitive processes involved in the dynamics of cognitive flexibility.
3.3 Exploratory Bayesian data analysis
We performed a Bayesian correlation analysis to examine the associations between iWCST measures, namely categories%, SLE%, nPE%, rPE%, and IE%. As expected, categories%, with higher values indicating better performance, and error variables, with lower values indicating better performance, are negatively correlated (Table 4). Categories% and SLE% show the strongest degree of association. The first-order correlations between error variables are positive but moderate at best, ranging from r = 0.35 (SLE% and nPE%) to r = 0.15 (IE% and rPE%).
We also performed a Bayesian network analysis on these cross-sectional iWCST multivariate data (Borsboom et al., 2021) using JASP’s Network module (Huth et al., 2023). To do this, we have used the default settings with the following exceptions: The GCGM (Gaussian copula graphical model, mixed variables) estimator was preferred over the GCM (Gaussian graphical model, continuous variables) estimator, and the iteration and burn-in sampling parameters were increased to 50,000 (instead of 10,000) and 20,000 (instead of 5,000), respectively. Finally, the prior edge inclusion probability was set to 0.10 (instead of 0.50).
Figure 10 shows the results of this Bayesian network analysis, which is based on partial correlations between the variables. As can be seen, the resulting network structure retains moderate complexity (including six edges), with one single structure assembling a posterior probability of about 0.75. The most likely network structure includes all edges between categories% and the error variables, SLE%, nPE%, rPE%, and IE%, with parameter estimates ranging from −0.87 (edge connecting categories% to SLE%) to −0.11 (edge connecting categories% to rPE%), plus two edges connecting error variables (edges connecting SLE% to nPE% and SLE% to IE%, respectively).
Figure 10. Bayesian network analysis results of categories%, SLE%, nPE%, rPE%, and IE% iWCST measures for N = 489. (A,B) Are concerned with network structure uncertainty and show the posterior probabilities of different network structures and their complexity. (A) Shows the posterior probability of different structural complexities, where the complexity is the network density. The peak posterior probability occurs at a density of six edges. (B) Shows the posterior probabilities of the visited structures, sorted from most to least probable. A single structure clearly stands out as the most probable, with a posterior probability around 0.75. (C,D) Are concerned with edge inclusion and exclusion and show details of the most probable network structure. (C) Edge evidence plot showing six edges present in blue (evidence for inclusion, BF10 > 10), three edges absent in red (evidence for exclusion, BF10 < 0.10), and one edge with insufficient evidence (0.10 < BF10 < 10) in gray. (D) The corresponding parameter estimate plot shows edges with BF10 > 10 and their model-averaged parameter estimates. Edge thickness and saturation represent the strength of the association; only negative relationships are present (shown in red). Note that an analysis of the uncertainty in the precision of the edge parameter estimates is not possible with the JASP Network module.
Overall, categories% appears as a strongly connected node (four edges), while rPE% appears as a weakly connected node (only one edge, which is also the weakest edge included). These impressions are clearly supported by the centrality measures (Supplementary Table S3.1), which show that categories% is the most central node in terms of betweenness, closeness, and strength, while rPE% is the least central node in terms of betweenness, closeness, and strength.
In addition to these analyses of internal iWCST associations, we analyzed associations between iWCST variables and the questionnaires we used (see Table 2; BFI-10, Rammstedt et al., 2014; BIS/BAS-11, Studer et al., 2016; PHQ-4, Löwe et al., 2010). The results of these Bayesian regression analyses can be found in the Supplementary Table S3.2. A notable finding reported there is that none of the questionnaire scores contributed significantly to the prediction of the iWCST measures, including PE variables. That is, neither personality variables, as indexed by two items per Big Five factor on the BFI-10, nor disposition to reward or punishment, as indexed by five items on the BIS scale and by six items on the BAS scale (with two items per subscale, i.e., fun seeking, drive, and reward responsiveness), nor current depressive or anxious mood, as indexed by two items on the PHQ-4, appear to be significant predictors of iWCST measures, including PE, in neurotypical individuals. DeYoung et al. (2005), Schretlen et al. (2010), and Stanek and Ones (2023) provide more information on possible relationships between personality and cognitive ability.
4 Discussion
The present study had three main goals in attempting to support further innovation in the field of neuropsychology. First, the results of the study clearly indicate that the iWCST is an appropriate tool for the unsupervised online assessment of cognitive flexibility in neurotypical individuals. Second, the results of this study, particularly the BAIN hypothesis evaluation (Hoijtink et al., 2019), provide conclusive evidence in favor of the novel GIC theory over the established SAS theory, suggesting that goal-directed instrumental control contributes significantly to cognitive flexibility. Third, the study exemplifies how the application of Bayesian data analysis can contribute to further advances in the field of neuropsychology.
4.1 Conclusion on the validity of the iWCST
Regarding the validity of the iWCST for assessing cognitive flexibility, it’s worth reconsidering some sample statistics. First, of the 826 individuals who responded to our email hyperlink, 503 completed the entire data collection (approximately 60.9%). Of the 323 individuals who dropped out, the majority (289 individuals, or approximately 89.5% of all dropouts) decided not to continue during the task instructions (early dropouts), while only 34 individuals (or approximately 10.5% of all dropouts) began but did not complete the data collection (late dropouts). Of these 34 late dropouts, 31 did not complete the iWCST, while the remaining three completed the iWCST but did not complete the questionnaires.
These statistics support the idea that the iWCST can collect representative samples of neurotypical individuals, since those who chose to participate after seeing the full task instructions were highly likely to complete the task. It also suggests that the cognitive load associated with the iWCST may be well tolerated by neurotypical individuals, as those who completed it appeared to do so with concentration, despite the unsupervised online data collection. Consequently, data quality was high among those who completed the iWCST and questionnaires in full, with only 14 out of 503 records (approximately 2.8%) requiring exclusion based on our pre-specified criteria for iWCST data quality.
The ESE replication, i.e., the finding that nPE% > rPE%, is another strong argument for the validity of the iWCST for assessing cognitive flexibility. This finding was previously reported for supervised paper-and-pencil (Kopp et al., 2019) and on-site computerized (Kopp et al., 2023, individual administration; Steinke et al., 2020a, group-wise administration) variants of the WCST and was replicated in this fully unsupervised online study, with slightly higher effect sizes compared to these two studies that included computerized versions of the WCST. Thus, the replicability of the ESE in neurotypical individuals has been demonstrated beyond reasonable doubt in the totality of studies that have included computerized and internet-based versions of the WCST with a total sample size of nearly 1,000 neurotypical individuals.
4.2 Conclusion on the theories considered
The robust behavioral ESE phenomenon consists of the simple empirical fact that non-repetitive PE (%) > repetitive PE (%) and is also important in the realm of neuropsychological theories. The reason the ESE is theoretically relevant is that the ESE supports the novel GIC theory (Kopp et al., 2023) over the established SAS theory (Norman and Shallice, 1981; Shallice, 1982; Shallice and Burgess, 1991, 1996; Shallice and Cooper, 2011). Specifically, as explained in more detail in the Introduction, SAS theory predicts nPE% < rPE%, due to the stronger involvement of the SAS in non-routine compared to routine situations, whereas GIC theory predicts nPE% > rPE%, due to the retrieval of goal-directed instrumental memory in repetitive situations only. Our confirmatory Bayesian data analysis yielded an estimated δ = 0.604 (point value of central tendency), with a 95 percent credible interval of [0.508–0.701], suggesting that the ordinal hypothesis nPE% > rPE% is correct. In addition, according to the BAIN hypothesis evaluation (Hoijtink et al., 2019), the posterior model probability of the SAS theory is very close to 0, while the posterior model probability of the GIC theory is very close to 1. These statistical findings are clearly inconsistent with the SAS theory, which therefore does not appear to capture well the essential cognitive processes behind the dynamic modulation of cognitive flexibility. Since a major goal of the present study was to reject the theory that received less support from the data, in accordance with the falsifiability approach, we reject the SAS theory as a viable explanation of cognitive flexibility as assessed in card-sorting tasks.
A main theoretical conclusion from the present and related data (Steinke et al., 2020a; Kopp et al., 2023) is that the GIC theory appears to provide an adequate explanation of cognitive flexibility as assessed by card-sorting tasks. Figure 11 provides a detailed representation of how the GIC theory explains the ESE. Essentially, the GIC theory posits the existence of two independent cognitive processes that follow informative error signals. First, regardless of the repetition vs. alternation of feature-action pairs on successive trials, executive control inhibits the continuation of the previously prioritized dimension following informative error signals, thereby generally preventing the occurrence of PE; indeed, we routinely find less than 10 % PE on average in neurotypical individuals in our studies. Second, goal-directed, instrumental memory traces are retrieved conditional on feature-action repetition on successive trials. In our previous study, we showed that these memory traces associate action-relevant features, actions, and feedback events from the most recent error trials (Kopp et al., 2023). The behavioral effect of retrieving traces of these instrumental episodes is to prevent the repetition of recently punished feature-action pairs, thereby producing the ESE, i.e., nPE% > rPE%. Note that the results of the exploratory Bayesian correlation/network analysis are consistent with the assumption of two independent cognitive processes following informative error signals, because categories%, SLE%, nPE%, and IE% appear to form a coherent subnetwork of iWCST measures, while rPE% may be subject to processes independent of the process governing the larger subnetwork (Figure 10).
Figure 11. The figure illustrates cognitive processes on switch trials following informative error trials, as proposed by the theory of goal-directed instrumental control (GIC). The figure uses the SHAPE dimension, the ‘asterisk’ dimensional feature, and the C key as examples. On the error trial shown on the left, a target card containing the asterisk feature was sorted by SHAPE to the keycard containing two green asterisks by pressing the C key (see Figure 1 for the complete layout of the keycards and associated keys), and the feedback INCORRECT provided an informative error signal. The GIC theory postulates the existence of two independent cognitive processes following this error signal. Details of these processes are shown on the switch trial shown on the right. First, executive function exerts dimensional control by inhibiting the continued prioritization of the SHAPE dimension, thereby generally preventing the occurrence of PE. Second, the reactivation of goal-directed instrumental memory, contingent on the repetition of an asterisk feature on the switch trial’s target card, combined with the consideration of responding again with the C key, the reactivation of goal-directed instrumental memory facilitates the retrieval of the most recent feature-action-feedback episode (Kopp et al., 2023). The behavioral effect of retrieving this goal-directed, instrumental memory trace is to prevent the repetition of that feature-action pair (here, responding with the C key to an asterisked target card), producing the ESE with non-repetitive PE (%) > repetitive PE (%), as shown in Figure 2.
The GIC theory has roots in dual-process theories of instrumental learning (Dickinson, 1994; De Wit and Dickinson, 2009; Balleine and O’Doherty, 2010). Dual-process theories distinguish between the formation of habitual and goal-directed memory during instrumental learning. Habits, defined as stimulus–response associations, are learned slowly and incrementally and are inflexible and insensitive to sudden changes in response-outcome contingencies (Dickinson and Pérez, 2018). Because it has proven surprisingly difficult to study habits in humans (de Wit et al., 2018), investigations of habitual control have used experimental paradigms that extensively train an instrumental response under consistent conditions. Such (over)training is thought to lead to a shift from goal-directed to habitual control. Goal-directed instrumental control is based on stimulus–response-outcome associations (Dickinson, 1994), allows for flexible adaptation to sudden changes in response-outcome contingencies, but is cognitively demanding. The field of reinforcement learning (Sutton and Barto, 2018), developed in computer science, is characterized by a similar dichotomy between model-based (resembling goal-directed) and model-free (resembling habitual) learning (Dolan and Dayan, 2013; Decker et al., 2016). The model-based system relies on internal models of the environment, allowing the simulation of possible actions and their potential outcomes based on the environmental model. This system is often associated with goal-directed decision-making, where decisions are made based on explicit consideration of expected outcomes and their associated values. The model-free system, on the other hand, is more habitual. It operates on the basis of learned associations between stimuli and responses that are driven by their outcomes, without explicit consideration of the structure of the environment. In model-free learning, decisions are made based on the reinforcement history of learned stimulus–response associations. Regarding this issue, our group has developed a computational model of the ESE, based on the distinction between model-based and model-free reinforcement learning (Steinke et al., 2018, 2020a,b,c; Steinke and Kopp, 2020).
From the paragraphs above, it should be clear that the GIC theory views the occurrence of PE as a behavioral signature of habit formation that can occasionally override executive control, even in neurotypical individuals, and the ESE as a behavioral signature of goal-directed instrumental control. This analysis ultimately leads to a three-level theory of cognitive flexibility, with the levels being habitual and goal-directed instrumental (alternatively, operational) control and executive (alternatively, strategic) control (see (Kopp, 2024), for a detailed account of a three-level theory of cognitive flexibility).
A cornerstone of this conceptualization is that retrieving goal-directed, instrumental memory traces depends on the repetition of feature-action conjunctions, as shown in Figure 11 and demonstrated in Kopp et al. (2023). Conjunctive retrieval on switch trials, which depends on the repetition of both features and actions, is instrumental in nature and generates the ESE. This stands in contrast to disjunctive retrieval, which on switch trials depends on the repetition of either features or actions and generates switch costs in cued card-sorting tasks (Kopp et al., 2020a,b).
Our integration of GIC theory within a broader framework of three levels of retrieval-based behavioral control (habitual, goal-directed/operational, executive/strategic) may contribute to a better understanding of cognitive flexibility (Kopp, 2024). GIC theory shares similarities not only with the dual-process theory of instrumental learning, but also with several retrieval-based accounts in cognitive psychology. These theories include, but are not limited to, the event coding theory (see Hommel et al., 2001), which posits that perception and action are closely linked, with representations of events encoded in a common format; the binding and retrieval theory of action control (see Frings et al., 2020), which examines how actions are bound to contexts and retrieved during action selection; learning theories of cognitive control (see Chein and Schneider, 2012), which examine how cognitive control skills are acquired through learning processes; theories linking memory and inhibition (see Anderson and Hulbert, 2021), which examine how memory processes interact with inhibitory control mechanisms; and instance theories of cognition (see Jamieson et al., 2022), which propose that cognitive processes involve the retrieval of specific instances or episodes from memory to guide current behavior. Evaluating these theories for their explanatory power may ultimately provide a comprehensive understanding of cognitive flexibility. Further neuropsychological research, particularly in patients with neurological disorders, is needed to elucidate the neuroanatomical correlates associated with repetition-dependent cognitive processes.
A related group of theories is the Bayesian view of the brain, which conceptualizes a predictive system that operates through action-perception loops using feedback-feedforward connectivity (e.g., Friston, 2010; Kolossa et al., 2015; Friston et al., 2017) and has been repeatedly applied to the Wisconsin card-sorting task (D’Alessandro et al., 2020; Barceló, 2021). From this perspective, the reduction of PE in repetitive (or stable) versus non-repetitive (or volatile) task contexts can be considered a specific example of repetition-dependent plasticity, whether simply characterized by reduced responding to repetition (e.g., Auksztulewicz and Friston, 2016) or, alternatively, by short-term memory that guides adaptation to predictable events (Jonides et al., 2008). Either way, these ideas support the notion that understanding the ESE and related phenomena requires exploring how repetition of prior stimulus–response exposures affects cognitive processes, perhaps in a manner similar to GIC theory, thereby contributing to a general understanding of cognitive flexibility.
4.3 Conclusion on the transition to Bayesian statistics
In our study, we distinguished between two approaches to Bayesian data analysis: confirmatory and exploratory. Confirmatory analysis focuses on hypothesis testing and theory evaluation. It uses posterior model probabilities to assess the plausibility of hypotheses and integrates prior beliefs and observed data to compare the plausibility of competing ideas. Exploratory Bayesian analysis, on the other hand, emphasizes iterative learning through parameter estimation, facilitating the cumulative discovery of new insights and relationships in the data. User-friendly tools such as JASP (JASP Team, 2024) and its BAIN module (Hoijtink et al., 2019) make Bayesian statistics more accessible. Throughout this article, we have argued that confirmatory Bayesian data analysis can further advance our theoretical understanding of cognitive processes important to neuropsychological research. We hope that this example of confirmatory Bayesian data analysis will encourage more frequent use of this highly targeted method in neuropsychological research.
5 Study limitations and future directions
Our study has some limitations, which are primarily related to issues of generalizability. The main limitation is that only a sample of neurotypical individuals was studied, which precludes generalization to patients with brain damage. In a previous study, the ESE was observed in consecutively sampled patients with a variety of brain diseases, but no clear disease-behavior relationship emerged (Kopp et al., 2019). Thus, understanding how the presence of neurological disease affects PE, including the ESE, is still an important area of neuropsychological research on cognitive flexibility (Dajani and Uddin, 2015). Specifically, brain-behavior relations in patients with focal brain lesions should be examined to understand which neural structures are involved in habitual, goal-directed/operational, and cognitive/strategic behavioral control for balancing cognitive flexibility against stability (Goschke and Bolte, 2014; Dreisbach and Fröber, 2019; Kopp, in press). However, the suitability of the iWCST for use in patients with neurological diseases remains to be evaluated. If the cognitive load of the iWCST is found to be appropriate for individuals with focal brain lesions, the application of voxel-based lesion-behavior mapping (Pustina and Mirman, 2022) should be the next step in elucidating the neuropsychological correlates of PE and the ESE in card-sorting tasks.
The focus of the present study is primarily on the ESE, so it would be beneficial to measure it in the presence of more prevalent PE. As mentioned above, we routinely find that neurotypical individuals have an average of less than 10 % PE, even in non-repetitive situations. A substantial proportion of the participants (N = 59; see Table 3) did not commit any PE during their entire iWCST assessment and, strictly speaking, it is not possible to study the ESE in the absence of PE. This issue is illustrated by the fact that after excluding these participants, the estimated effect size increases to δ = 0.661 (point value of central tendency), with a 95 percent credible interval of [0.557–0.765] in those participants who committed one or more PE. Follow-up studies of neurotypical individuals may well benefit from modifications of the iWCST, such as administering it in the context of dual-task management, which is known to increase the probability to commit PE, especially when the additional task administered simultaneously draws on similar cognitive resources as the card-sorting task (Cooper et al., 2012).
Most participants who provided complete records requested and received the financial compensation offered in the invitation email. This compensation appears to have been a motivating factor for participation, and it is not currently known how motivation to participate in this type of research is influenced by other motivational forces, such as contributing to the understanding of the nature of neurological diseases. The clinical application of the iWCST requires the resolution of several translational issues, such as the effects of motivational forces to participate and the appropriateness of cognitive load for individuals with brain damage. If these translational issues can be resolved, the introduction of the iWCST into clinical neuropsychology offers a number of advantages, including massively increased reach for obtaining large and more representative sample sizes and facilitated opportunities for disease-specific standardization of this important neuropsychological test (Brooks et al., 2009; Kiselica et al., 2023).
To avoid potential misunderstanding, we do not claim to have invented digital neuropsychology, theory-driven research, or Bayesian statistics within neuropsychology; indeed, there are many examples of each of these achievements in the literature. The innovation in this study lies in the unique alignment of these three building blocks, including the use of an Internet-based version of the Wisconsin card-sorting task (i.e., iWCST), the development of a diagnostic test to compare two predictions derived from different theoretical backgrounds (SAS and GIC theories), and the use of a sophisticated method of confirmatory Bayesian hypothesis testing.
6 Conclusion
Neuropsychology should take advantage of digital technologies. This technological advancement will create new opportunities for neuropsychological research by significantly expanding the reach of neuropsychological assessment. As demonstrated in the present study, the quality of behavioral data obtained through unsupervised online testing can reach high levels. Unprecedented opportunities for Bayesian hypothesis evaluation will be achieved with large sample sizes, accelerating the development of neuropsychological theories. Here we have shown that the established SAS theory does not adequately explain the dynamics of cognitive flexibility, whereas the novel GIC theory seems to provide a good starting point for its understanding. According to GIC theory, executive/strategic control, whose limits of efficiency can be measured as PE in non-repetitive situations, and goal-directed instrumental/operational control, whose efficiency can be measured as PE in repetitive situations and the ESE, work together to prevent the occurrence of habitual responses (Kopp, 2024). In addition, these digital technologies will provide one of the foundations for the future of clinical neuropsychology, with internet-based assessment and rehabilitation tools that will enable remote patient care, perhaps combined with supervised tele-neuropsychological techniques (Singh and Germine, 2021; Luxton et al., 2024). The iWCST presented here is just one example of how this adoption of digital technologies is likely to evolve in the future of neuropsychology.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
The studies involving humans were approved by Ethikkommission Medizinische Hochschule Hannover. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.
Author contributions
CS: Writing – original draft, Writing – review & editing. BK: Writing – original draft, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. Financial support for the research and publication of this article was provided to BK by a grant from the Karlheinz Hartmann Foundation, Hannover, Germany.
Acknowledgments
We would like to thank Cassandra Krause, Franziska Schütz, and Eduard Unruh for assisting the authors in preparing this paper. Thanks are also due Eckart Altenmüller (HMTMH), Volkhard Fischer (MHH), Mark Vollrath (TU BS), for distributing emails to their students.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2024.1437192/full#supplementary-material
Footnotes
1. ^https://www.parinc.com/What-is-PARiConnect
2. ^Readers may recall that the so-called Law of Effect states that behaviors followed by positive outcomes are more likely to be repeated, while behaviors followed by negative outcomes are less likely to be repeated (Thorndike, 1933). From the perspective of operant conditioning (Skinner, 1953), negative outcomes reduce the likelihood of behaviors and, by extension in GIC theory, of cognitive processes such as the continuation of a cognitive set.
3. ^https://jatos.mindprobe.eu/
4. ^The JASP Team, Amsterdam, NL, The Netherlands, https://jasp-stats.org/, accessed January 15, 2024.
References
Anderson, M. C., and Hulbert, J. C. (2021). Active forgetting: adaptation of memory by prefrontal control. Annu. Rev. Psychol. 72, 1–36. doi: 10.1146/annurev-psych-072720-094140
Auksztulewicz, R., and Friston, K. (2016). Repetition suppression and its contextual determinants in predictive coding. Cortex 80, 125–140. doi: 10.1016/j.cortex.2015.11.024
Balleine, B., and O’Doherty, J. (2010). Human and rodent homologies in action control: corticostriatal determinants of goal-directed and habitual action. Neuropsychopharmacology 35, 48–69. doi: 10.1038/npp.2009.131
Barceló, F. (2003). The Madrid card sorting test (MCST): a task switching paradigm to study executive attention with event-related potentials. Brain Res. Protocol. 11, 27–37. doi: 10.1016/s1385-299x(03)00013-8
Barceló, F. (2021). A predictive processing account of card sorting. J. Cogn. Neurosci. 33, 1636–1656. doi: 10.1162/jocn_a_01662
Benjamin, D. J., Berger, J. O., Johannesson, M., Nosek, B. A., Wagenmakers, E.-J., Berk, R., et al. (2018). Redefine statistical significance. Nat. Hum. Behav. 2, 6–10. doi: 10.1038/s41562-017-0189-z
Berg, E. A. (1948). A simple objective technique for measuring flexibility in thinking. J. Gen. Psychol. 39, 15–22. doi: 10.1080/00221309.1948.9918159
Bilder, R. M., and Reise, S. P. (2019). Neuropsychological tests of the future: how do we get there from here? Clin. Neuropsychol. 33, 220–245. doi: 10.1080/13854046.2018.1521993
Borsboom, D., Deserno, M. K., Rhemtulla, M., Epskamp, S., Fried, E. I., McNally, R. J., et al. (2021). Network analysis of multivariate data in psychological science. Nat. Rev. Methods Prim. 1:58. doi: 10.1038/s43586-021-00055-w
Brooks, B. L., Strauss, E., Sherman, E., Iverson, G. L., and Slick, D. J. (2009). Developments in neuropsychological assessment: refining psychometric and clinical interpretive methods. Can. Psychol. Psychol. Can. 50, 196–209. doi: 10.1037/a0016066
Casaletto, K. B., and Heaton, R. K. (2017). Neuropsychological assessment: past and future. J. Int. Neuropsychol. Soc. 23, 778–790. doi: 10.1017/s1355617717001060
Chein, J. M., and Schneider, W. (2012). The brain’s learning and control architecture. Curr. Dir. Psychol. Sci. 21, 78–84. doi: 10.1177/0963721411434977
Cooper, R. P., Karolina, W., and Davelaar, E. J. (2012). Differential contributions of set-shifting and monitoring to dual-task interference. Q. J. Exp. Psychol. 65, 587–612. doi: 10.1080/17470218.2011.629053
D’Alessandro, M., Radev, S. T., Voss, A., and Lombardi, L. (2020). A Bayesian brain model of adaptive behavior: an application to the Wisconsin card sorting task. PeerJ 8:e10316. doi: 10.7717/peerj.10316
Dajani, D. R., and Uddin, L. Q. (2015). Demystifying cognitive flexibility: implications for clinical and developmental neuroscience. Trends Neurosci. 38, 571–578. doi: 10.1016/j.tins.2015.07.003
De Wit, S., and Dickinson, A. (2009). Associative theories of goal-directed behaviour: a case for animal–human translational models. Psychol. Res. 73, 463–476. doi: 10.1007/s00426-009-0230-6
De Wit, S., Kindt, M., Knot, S. L., Verhoeven, A. A. C., Robbins, T. W., Gasull-Camos, J., et al. (2018). Shifting the balance between goals and habits: five failures in experimental habit induction. J. Exp. Psychol. Gen. 147, 1043–1065. doi: 10.1037/xge0000402
Decker, J. H., Otto, A. R., Daw, N. D., and Hartley, C. A. (2016). From creatures of habit to goal-directed learners: tracking the developmental emergence of model-based reinforcement learning. Psychol. Sci. 27, 848–858. doi: 10.1177/0956797616639301
Demakis, G. J. (2003). A meta-analytic review of the sensitivity of the Wisconsin card sorting test to frontal and lateralized frontal brain damage. Neuropsychology 17, 255–264. doi: 10.1037/0894-4105.17.2.255
DeYoung, C. G., Peterson, J. B., and Higgins, D. M. (2005). Sources of openness/intellect: cognitive and neuropsychological correlates of the fifth factor of personality. J. Pers. 73, 825–858. doi: 10.1111/j.1467-6494.2005.00330.x
Dickinson, A. (1994). “Instrumental conditioning” in Animal learning and cognition. ed. N. J. Mackintosh (San Diego, CA: Academic Press), 45–79.
Dickinson, A., and Pérez, O. D. (2018). “Actions and habits: psychological issues in dual system theory” in Goal-directed decision making: Computations and neural circuits. eds. R. W. Morris, A. M. Bornstein, and A. Shenhav (Orlando, FL: Academic Press), 1–37.
Dolan, R. J., and Dayan, P. (2013). Goals and habits in the brain. Neuron 80, 312–325. doi: 10.1016/j.neuron.2013.09.007
Dreisbach, G., and Fröber, K. (2019). On how to be flexible (or not): modulation of the stability-flexibility balance. Curr. Dir. Psychol. Sci. 28, 3–9. doi: 10.1177/0963721418800030
Faulkenberry, T. J., Ly, A., and Wagenmakers, E.-J. (2020). Bayesian inference in numerical cognition: a tutorial using JASP. J. Numer. Cogn. 6, 231–259. doi: 10.5964/jnc.v6i2.288
Feenstra, H. E., Vermeulen, I. E., Murre, J. M., and Schagen, S. B. (2017). Online cognition: factors facilitating reliable online neuropsychological test results. Clin. Neuropsychol. 31, 59–84. doi: 10.1080/13854046.2016.1190405
Frings, C., Hommel, B., Koch, I., Rothermund, K., Dignath, D., Giesen, C., et al. (2020). Binding and retrieval in action control (BRAC). Trends Cogn. Sci. 24, 375–387. doi: 10.1016/j.tics.2020.02.004
Friston, K. (2010). The free-energy principle: a unified brain theory? Nat. Rev. Neurosci. 11, 127–138. doi: 10.1038/nrn2787
Friston, K., FitzGerald, T., Rigoli, F., Schwartenbeck, P., and Pezzulo, G. (2017). Active inference: a process theory. Neural Comput. 29, 1–49. doi: 10.1162/NECO_a_00912
Gelman, A., Carlin, J. B., Stern, H. S., Dunson, D. B., Vehtari, A., and Rubin, D. B. (2013). Bayesian Data Analysis. 3rd Edn. Boca Raton, FL: Chapman and Hall/CRC Press.
Gelman, A., and Shalizi, C. R. (2013). Philosophy and the practice of Bayesian statistics. Br. J. Math. Stat. Psychol. 66, 8–38. doi: 10.1111/j.2044-8317.2011.02037.x
Germine, L., Reinecke, K., and Chaytor, N. S. (2019). Digital neuropsychology: challenges and opportunities at the intersection of science and software. Clin. Neuropsychol. 33, 271–286. doi: 10.1080/13854046.2018.1535662
Goschke, T., and Bolte, A. (2014). Emotional modulation of control dilemmas: the role of positive affect, reward, and dopamine in cognitive stability and flexibility. Neuropsychologia 62, 403–423. doi: 10.1016/j.neuropsychologia.2014.07.015
Grant, D. A., and Berg, E. (1948). A behavioral analysis of degree of reinforcement and ease of shifting to new responses in a Weigl-type card-sorting problem. J. Exp. Psychol. 38, 404–411. doi: 10.1037/h0059831
Heaton, R. K. (1981). A manual for the Wisconsin card sorting test. Odessa, FL: Psychological Assessment Resources.
Heaton, R. K., Chelune, G. J., Talley, J. L., Kay, G. G., and Curtiss, G. (1993). Wisconsin card sorting test manual: Revised and expanded. Lutz, FL: Psychological Assessment Resources.
Heaton, R. K.PAR Staff (2003a). Wisconsin card sorting test: Computer version 4-research edition (WCST: CV4). Lutz, FL: Psychological Assessment Resources.
Heaton, R. K.PAR Staff (2003b). Wisconsin card sorting Test-64: Computer version 2-research edition (WCST-64:CV2). Lutz, FL: Psychological Assessment Resources.
Hoijtink, H. (2011). Informative hypotheses: Theory and practice for behavioral and social scientists. New York: Chapman and Hall/CRC Press.
Hoijtink, H., Gu, X., and Mulder, J. (2019). Bayesian evaluation of informative hypotheses for multiple populations. Br. J. Math. Stat. Psychol. 72, 219–243. doi: 10.1111/bmsp.12145
Hommel, B., Müsseler, J., Aschersleben, G., and Prinz, W. (2001). The theory of event coding (TEC): a framework for perception and action planning. Behav. Brain Sci. 24, 849–878. doi: 10.1017/s0140525x01000103
Huth, K. B. S., de Ron, J., Goudriaan, A. E., Luigjes, J., Mohammadi, R., van Holst, R. J., et al. (2023). Bayesian analysis of cross-sectional networks: a tutorial in R and JASP. Adv. Methods Pract. Psychol. Sci. 6, 1–18. doi: 10.1177/25152459231193334
Jamieson, R. K., Johns, B. T., Vokey, J. R., and Jones, M. N. (2022). Instance theory as a domain-general framework for cognitive psychology. Nat. Rev. Psychol. 1, 174–183. doi: 10.1038/s44159-022-00025-3
Jonides, J., Lewis, R. L., Nee, D. E., Lustig, C. A., Berman, M. G., and Moore, K. S. (2008). The mind and brain of short-term memory. Annu. Rev. Psychol. 59, 193–224. doi: 10.1146/annurev.psych.59.103006.093615
Kessels, R. P. (2019). Improving precision in neuropsychological assessment: bridging the gap between classic paper-and-pencil tests and paradigms from cognitive neuroscience. Clin. Neuropsychol. 33, 357–368. doi: 10.1080/13854046.2018.1518489
Kiselica, A. M., Karr, J. E., Mikula, C. M., Ranum, R. M., Benge, J. F., Medina, L. D., et al. (2023). Recent advances in neuropsychological test interpretation for clinical practice. Neuropsychol. Rev. 34, 637–667. doi: 10.1007/s11065-023-09596-1
Kolossa, A., Kopp, B., and Fingscheidt, T. (2015). A computational analysis of the neural bases of Bayesian inference. NeuroImage 106, 222–237. doi: 10.1016/j.neuroimage.2014.11.007
Kongs, S. K., Thompson, L. L., Iverson, G. L., and Heaton, R. K. (2000). Wisconsin card sorting Test-64 card version (WCST-64). Odessa, FL: Psychological Assessment Resources.
Kopp, B. (2024). “A theory of dysexecutive syndromes: variations in the effectiveness of prefrontal control of behavior” in Encyclopedia of the human brain, section executive functions. ed. M. Ullsperger . 2nd ed (Amsterdam, NL: Elsevier).
Kopp, B., Al-Hafez, B., and Steinke, A. (2023). Habits, goals, and behavioral signs of cognitive perseveration on Wisconsin card-sorting tasks. Brain Sci. 13:919. doi: 10.3390/brainsci13060919
Kopp, B., and Lange, F. (2013). Electrophysiological indicators of surprise and entropy in dynamic task-switching environments. Front. Hum. Neurosci. 7:300. doi: 10.3389/fnhum.2013.00300
Kopp, B., Lange, F., and Steinke, A. (2021). The reliability of the Wisconsin card sorting test in clinical practice. Assessment 28, 248–263. doi: 10.1177/1073191119866257
Kopp, B., Steinke, A., Bertram, M., Skripuletz, T., and Lange, F. (2019). Multiple levels of control processes for Wisconsin card sorts: an observational study. Brain Sci. 9:141. doi: 10.3390/brainsci9060141
Kopp, B., Steinke, A., Meiran, N., Seer, C., and Lange, F. (2020a). Stimulus- and response-based interference contributes to the costs of switching between cognitive tasks. Psychol. Res. 84, 1112–1125. doi: 10.1007/s00426-018-1113-5
Kopp, B., Steinke, A., and Visalli, A. (2020b). Cognitive flexibility and N2/P3 event-related brain potentials. Sci. Rep. 10:9859. doi: 10.1038/s41598-020-66781-5
Kruschke, J. (2014). Doing Bayesian data analysis: A tutorial with R, JAGS, and Stan. 2nd Edn. San Diego, CA: Academic Press.
Lange, F., Brückner, C., Knebel, A., Seer, C., and Kopp, B. (2018). Executive dysfunction in Parkinson’s disease: a meta-analysis on the Wisconsin card sorting test literature. Neurosci. Biobehav. Rev. 93, 38–56. doi: 10.1016/j.neubiorev.2018.06.014
Lange, F., Kröger, B., Steinke, A., Seer, C., Dengler, R., and Kopp, B. (2016). Decomposing card-sorting performance: effects of working memory load and age-related changes. Neuropsychology 30, 579–590. doi: 10.1037/neu0000271
Lange, K., Kühn, S., and Filevich, E. (2015). “Just another tool for online studies” (JATOS): an easy solution for setup and management of web servers supporting online studies. PLoS One 10:e0134073. doi: 10.1371/journal.pone.0134073
Lange, F., Seer, C., and Kopp, B. (2017). Cognitive flexibility in neurological disorders: cognitive components and event-related potentials. Neurosci. Biobehav. Rev. 83, 496–507. doi: 10.1016/j.neubiorev.2017.09.011
Libon, D. J., Swenson, R., Lamar, M., Price, C. C., Baliga, G., Pascual-Leone, A., et al. (2022). The Boston process approach and digital neuropsychological assessment: past research and future directions. J. Alzheimers Dis. 87, 1419–1432. doi: 10.3233/jad-220096
Loring, D. W., Bauer, R. M., Cavanagh, L., Drane, D. L., Enriquez, K. D., Reise, S. P., et al. (2018). Rationale and design of the National Neuropsychology Network. J. Int. Neuropsychol. Soc. 28, 1–11. doi: 10.1017/S1355617721000199
Löwe, B., Wahl, I., Rose, M., Spitzer, C., Glaesmer, H., Wingenfeld, K., et al. (2010). A 4-item measure of depression and anxiety: validation and standardization of the patient health Questionnaire-4 (PHQ-4) in the general population. J. Affect. Disord. 122, 86–95. doi: 10.1016/j.jad.2009.06.019
Luxton, D. D., Nelson, E. L., and Maheu, M. M. (2024). A practitionerʼs guide to telemental health: How to conduct legal, ethical, and evidence-based telepractice. Washington, DC: American Psychological Association.
Mathôt, S., and March, J. (2022). Conducting linguistic experiments online with OpenSesame and OSWeb. Lang. Learn. 72, 1017–1048. doi: 10.1111/lang.12509
Meehl, P. E. (1978). Theoretical risks and tabular asterisks: sir Karl, sir Ronald, and the slow progress of soft psychology. J. Consult. Clin. Psychol. 46, 806–834. doi: 10.1037/0022-006X.46.4.806
Microsoft Corporation (2016). Microsoft excel (version 2016) [computer software]. Redmond, Washington: Microsoft Corporation.
Miles, S., Howlett, C. A., Berryman, C., Nedeljkovic, M., Moseley, G. L., and Phillipou, A. (2021). Considerations for using the Wisconsin card sorting test to assess cognitive flexibility. Behav. Res. 53, 2083–2091. doi: 10.3758/s13428-021-01551-3
Milner, B. (1963). Effects of different brain lesions on card sorting: the role of the frontal lobes. Arch. Neurol. 9, 90–100. doi: 10.1001/archneur.1963.00460070100010
Nelson, H. E. (1976). A modified card sorting test sensitive to frontal lobe defects. Cortex 12, 313–324. doi: 10.1016/s0010-9452(76)80035-4
Norman, D. A., and Shallice, T. (1981). “Attention to action: Willed and automatic control of behaviour”, in proceedings of the Lake wilderness attention conference, eds. M. Lansman and E. Hunt. Seattle, WA: Washington University.
Nyhus, E., and Barceló, F. (2009). The Wisconsin card sorting test and the cognitive assessment of prefrontal executive functions: a critical update. Brain Cogn. 71, 437–451. doi: 10.1016/j.bandc.2009.03.005
Parsons, T. D., and Duffield, T. (2020). Paradigm shift toward digital neuropsychology and high-dimensional neuropsychological assessments. J. Med. Internet Res. 22:e23777. doi: 10.2196/23777
Parsons, T. D., McMahan, T., and Kane, R. (2018). Practice parameters facilitating adoption of advanced technologies for enhancing neuropsychological assessment paradigms. Clin. Neuropsychol. 32, 16–41. doi: 10.1080/13854046.2017.1337932
Pustina, D., and Mirman, D. (2022). Lesion-to-symptom mapping: Principles and tools. New York: Springer.
Rammstedt, B., Kemper, C. J., Klein, M. C., Beierlein, C., and Kovaleva, A. (2014). Big Five Inventory (BFI-10). Zusammenstellung sozialwissenschaftlicher Items und Skalen (ZIS). Mannheim: ZIS-GESIS Leibniz Institute for the Social Sciences.
Rozeboom, W. W. (1960). The fallacy of the null-hypothesis significance test. Psychol. Bull. 57, 416–428. doi: 10.1037/h0042040
Schretlen, D. J. (2010). Modified Wisconsin card sorting test. Lutz, FL: Psychological Assessment Resources.
Schretlen, D. J., van der Hulst, E. J., Pearlson, G. D., and Gordon, B. (2010). A neuropsychological study of personality: trait openness in relation to intelligence, fluency, and executive functioning. J. Clin. Exp. Neuropsychol. 32, 1068–1073. doi: 10.1080/13803391003689770
Shallice, T. (1982). Specific impairments of planning. Philos. Trans. R. Soc. B Biol. Sci. 298, 199–209. doi: 10.1098/rstb.1982.0082
Shallice, T., and Burgess, P. (1991). Deficits in strategy application following frontal lobe damage in man. Brain 114, 727–741. doi: 10.1093/brain/114.2.727
Shallice, T., and Burgess, P. (1996). The domain of supervisory processes and temporal organization of behaviour. Philos. Trans. R. Soc. B Biol. Sci. 351, 1405–1412. doi: 10.1098/rstb.1996.0124
Shallice, T., and Cooper, R. P. (2011). The organisation of mind. New York: Oxford University Press.
Singh, S., and Germine, L. (2021). Technology meets tradition: a hybrid model for implementing digital tools in neuropsychology. Int. Rev. Psychiatry 33, 382–393. doi: 10.1080/09540261.2020.1835839
Stanek, K. C., and Ones, D. S. (2023). Meta-analytic relations between personality and cognitive ability. Proc. Natl. Acad. Sci. 120:e2212794120. doi: 10.1073/pnas.2212794120
Steinke, A., and Kopp, B. (2020). Toward a computational neuropsychology of cognitive flexibility. Brain Sci. 10:1000. doi: 10.3390/brainsci10121000
Steinke, A., Lange, F., and Kopp, B. (2020a). Parallel model-based and model-free reinforcement learning for card sorting performance. Sci. Rep. 10:15464. doi: 10.1038/s41598-020-72407-7
Steinke, A., Lange, F., Seer, C., Hendel, M. K., and Kopp, B. (2020b). Computational modeling for neuropsychological assessment of bradyphrenia in Parkinson’s disease. J. Clin. Med. 9:1158. doi: 10.3390/jcm9041158
Steinke, A., Lange, F., Seer, C., and Kopp, B. (2018). Toward a computational cognitive neuropsychology of Wisconsin card sorts: a showcase study in Parkinson’s disease. Comput. Brain Behav. 1, 137–150. doi: 10.1007/s42113-018-0009-1
Steinke, A., Lange, F., Seer, C., Petri, S., and Kopp, B. (2020c). A computational study of executive dysfunction in amyotrophic lateral sclerosis. J. Clin. Med. 9:2605. doi: 10.3390/jcm9082605
Studer, J., Baggio, S., Mohler-Kuo, M., Daeppen, J.-B., and Gmel, G. (2016). Factor structure and psychometric properties of a French and German shortened version of the Behavioural inhibition system/Behavioural activation system scales. Int. J. Methods Psychiatr. Res. 25, 44–54. doi: 10.1002/mpr.1486
Sutton, R. S., and Barto, A. G. (2018). Reinforcement learning: An introduction. 2nd Edn. Cambridge, MA: MIT Press.
Szucs, D., and Ioannidis, J. P. (2017). When null hypothesis significance testing is unsuitable for research: a reassessment. Front. Hum. Neurosci. 11:390. doi: 10.3389/fnhum.2017.00390
Thorndike, E. L. (1933). A proof of the law of effect. Science 77, 173–175. doi: 10.1126/science.77.1989.173.b
Van de Schoot, R., Depaoli, S., King, R., Kramer, B., Märtens, K., Tadesse, M. G., et al. (2021). Bayesian statistics and modelling. Nat. Rev. Methods Prim. 1:1. doi: 10.1038/s43586-020-00003-0
Van den Bergh, D., Clyde, M. A., Gupta, A. R. K. N., De Jong, T., Gronau, Q. F., Marsman, M., et al. (2021). A tutorial on Bayesian multi-model linear regression with BAS and JASP. Behav. Res. 53, 2351–2371. doi: 10.3758/s13428-021-01552-2
Veale, J. F. (2014). Edinburgh handedness inventory-short form: a revised version based on confirmatory factor analysis. Laterality 19, 164–177. doi: 10.1080/1357650X.2013.783045
Wagenmakers, E.-J., Marsman, M., Jamil, T., Ly, A., Verhagen, J., Love, J., et al. (2018a). Bayesian inference for psychology. Part I: theoretical advantages and practical ramifications. Psychon. Bull. Rev. 25, 35–57. doi: 10.3758/s13423-017-1343-3
Keywords: digital neuropsychology, online assessment, theory-driven research, Bayesian statistics, executive function, cognitive flexibility, Wisconsin card sorting, perseveration errors
Citation: Schmerwitz C and Kopp B (2024) The future of neuropsychology is digital, theory-driven, and Bayesian: a paradigmatic study of cognitive flexibility. Front. Psychol. 15:1437192. doi: 10.3389/fpsyg.2024.1437192
Edited by:
Lawrence M. Parsons, The University of Sheffield, United KingdomReviewed by:
Francisco Barceló, University of the Balearic Islands, SpainNicolas Zink, University of California, Los Angeles, United States
Copyright © 2024 Schmerwitz and Kopp. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Bruno Kopp, a29wcC5icnVub0BtaC1oYW5ub3Zlci5kZQ==