Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 20 September 2024
Sec. Neuropsychology
This article is part of the Research Topic Neuropsychological Testing: From Psychometrics to Clinical Neuropsychology View all 11 articles

Using behavior and eye-fixations to detect feigned memory impairment

Filomena Gomes,Filomena Gomes1,2Inês FerreiraInês Ferreira2Bruno RosaBruno Rosa2Ana Martins da Silva,,Ana Martins da Silva3,4,5Sara Cavaco,,,
Sara Cavaco1,2,3,4*
  • 1Neuropsychology Service, Centro Hospitalar Universitário de Santo António, Porto, Portugal
  • 2Laboratory of Neurobiology of Human Behavior, Centro Hospitalar Universitário de Santo António, Porto, Portugal
  • 3UMIB - Unit for Multidisciplinary Research in Biomedicine, ICBAS - School of Medicine and Biomedical Sciences, University of Porto, Porto, Portugal
  • 4ITR - Laboratory for Integrative and Translational Research in Population Health, Porto, Portugal
  • 5Department of Neurology, Centro Hospitalar Universitário de Santo António, Porto, Portugal

Background: Detecting invalid cognitive performance is an important clinical challenge in neuropsychological assessment. The aim of this study was to explore behavior and eye-fixations responses during the performance of a computerized version of the Test of Memory Malingering (TOMM-C) under standard vs. feigning conditions.

Participants and methods: TOMM-C with eye-tracking recording was performed by 60 healthy individuals (31 with standard instruction – SI; and 29 were instructed to feign memory impairment: 21 Naïve Simulators – NS and 8 Coached Simulators – CS) and 14 patients with Multiple Sclerosis (MS) and memory complaints performed. Number of correct responses, response time, number of fixations, and fixation time in old vs. new stimuli were recorded. Nonparametric tests were applied for group comparison.

Results: NS produced fewer correct responses and had longer response times in comparison to SI on all three trials. SI showed more fixations and longer fixation time on previously presented stimuli (i.e., familiarity preference) specially on Trial 1, whereas NS had more fixations and longer fixation time on new stimuli (i.e., novelty preference) specially in the Retention trial. MS patients produced longer response time and had a different fixation pattern than SI subjects. No behavioral or oculomotor difference was observed between NS and CS.

Conclusion: Healthy simulators have a distinct behavioral and eye-fixation response pattern, reflecting a novelty preference. Oculomotor measures may be useful to detect exaggeration or fabrication of cognitive dysfunction. Though, its application in clinical populations may be limited.

Introduction

Malingering is the volitional feigning or exaggeration of neurocognitive symptoms for the purpose of obtaining material gain or services or avoiding formal duty, responsibility, or undesirable outcome (Slick et al., 1999; Sherman et al., 2020). The detection of noncredible cognitive performance is an important clinical challenge in neuropsychological assessment. The presence of an external incentive (e.g., Social Security benefits, insurance compensation) is an important element to consider when distinguishing people with credible cognitive impairment from feigned cognitive deficits, even in clinical, non-forensic settings. The presence of an external incentive does not necessarily indicate unreliable neuropsychological test performance. However, it has been demonstrated that being in the process of applying for Social Security disability benefits increases the likelihood of noncredible performance (Schroeder et al., 2022; Horner et al., 2023). It has been estimated that between one third to two thirds of clinically referred patients with Social Security disability as an external incentive produce invalid data on performance validity tests - PVTs (Chafetz and Biondolillo, 2013; Schroeder et al., 2022), whereas less than one tenth of low-functioning Child Protection claimants who are motivated to do well failed on PVTs. Frequently, patients referred for routine clinical neuropsychological evaluation utilize the results of the examination for Social Security documentation.

PVTs are objective tests designed to detect invalid cognitive performance, i.e., feigned and/or exaggerated diminished capability (Sweet et al., 2021). PVTs usually require little effort or ability, as they typically are normally performed by a wide range of patients who have bona fide neurologic, psychiatric, or developmental problems (Heilbronner et al., 2009).

Most PVTs are forced-choice memory recognition tests and only explore accuracy to detect poor cognitive effort or malingering. Recent studies suggest that response time (Bolan et al., 2002; Kanser et al., 2019; Patrick et al., 2021b) and eye-tracking measures (Heaver and Hutton, 2011; Kanser et al., 2020; Tomer et al., 2020; Patrick et al., 2021a) may produce incremental information to the conventional accuracy responses on PVTs.

The Test of Memory Malingering (TOMM; Tombaugh, 1996) is one of the most widely used PVTs in research and clinical practice. TOMM is a forced-choice visual memory recognition test and the number of correct responses is the standard measure to discriminate between true memory impairment from noncredible performance. Response times are also able to detect feigned memory impairment on TOMM (Bolan et al., 2002; Kanser et al., 2019). The oculomotor behavior during the performance of TOMM has yet to be investigated.

This study aimed to quantify the potential information gains provided by eye fixation data in addition to behavioral response (i.e., response accuracy and response time), in the performance of a computerized version of TOMM (TOMM-C), to distinguish simulators from non-simulators. The clinical applicability of these measures was also explored in a sample of patients with multiple sclerosis and memory complaints. We hypothesized that eye-tracking metrics, in particular eye-fixations, could be an informative complement to behavioral responses on TOMM-C and that oculomotor measures are less vulnerable to coaching than behavioral responses.

Materials and methods

Subjects

Sixty healthy subjects recruited in the community were asked to perform a computerized version of TOMM (TOMM-C). The participants were distributed into two groups: 31 healthy subjects received the standard instruction (SI group) and 29 were instructed to feign memory impairment as if they were in the initial stages of dementia to benefit from Social Security disability (21 were “Naïve Simulators” – NS group, and 8 were “Coached Simulators” – CS group). Fourteen patients with diagnosis of Multiple Sclerosis (Polman et al., 2011) and with cognitive complaints on the routine neurological consultation, but without history of optic neuritis, were recruited from the outpatient clinic (MS group). All participants provided their informed written consent in accordance with the Declaration of Helsinki and the Centro Hospitalar Universitário de Santo António’s Ethical Committee (reference number 026-DEFI/049-CES; Figure 1).

Figure 1
www.frontiersin.org

Figure 1. Flowchart of the study participants and of the TOMM experiment.

Procedures

TOMM-C

TOMM-C is a computerized version of the standard TOMM (Tombaugh, 1996) adapted for eye-tracking recording. TOMM-C was presented in a Windows Based Software (Presentation® - Neurobehavioral Systems, Inc.). The stimuli were presented in a TFT Monitor 19″ with touch screen (KTMT-1921-USB/B, Keytec) with behavioral response recording. The iView X™ HiSpeed 1250 System (SensoriMotoric Instruments), with chin rest and forehead rest at 45 cm from the screen, eye-movements were recorded (i.e., monocular recording) during test performance. Similar to the standard version, TOMM-C is composed of two learning trials (Trial 1 and Trial 2) and a Retention trial. In both learning trials, there was an encoding phase and a recognition phase (Figure 2). During the encoding phase, participants were shown a series of simple line drawings (i.e., the same set of stimuli as the standard TOMM; Tombaugh, 1996) for 3 s each. Between items, a cross was displayed for 1 s on the screen followed by a blank screen for 1 s. The encoding phase was immediately followed by a two-choice recognition task. A Retention Trial, which was composed by just the two-choice recognition task, was administered following a 15-min delay (without further exposure of stimulus items). During the recognition phase of the three trials, after 3 s of free viewing of each pair of drawing, participants were cued by a buzz to respond with a touch on the screen (i.e., to identify the previously seen drawing). After the subject responded, a cross was displayed for 1 s on the screen followed by a blank screen for 1 s before the display of the next item. No feedback on the accuracy of response was provided. The pattern of eye-fixations was recorded during the free viewing of the test phase. The threshold to be considered a fixation was set at 100 ms (Manor and Gordon, 2003). Two areas of interest were identified: the “old” (i.e., drawing previously seen on the learning phase) and the “new” (i.e., foil drawing). Three behavioural measures were recorded: Number of Correct Responses, Total Response Time, and Median Response Time on Correct Responses. Three oculomotor measures were considered: % of Total Number of Fixation on “new” items, % of Total Fixation Time on “new” items, and % of Fixation Time on “new” items for correct responses.

Figure 2
www.frontiersin.org

Figure 2. (A) TOMM encoding phase at Trials 1 and 2; (B) TOMM two choice recognition phase at Trial 1, Trial 2, and Retention.

Eye-fixation data on Trial 2 and Retention Trial from two NS participants were discarded due to recording problems that resulted in extensive missing data, however their behavioral responses on those trials were analyzed. One NS participant did not produce correct responses on Trial 2, therefore the Median Response Time on Correct Responses and the % of Fixation Time on “New” for Correct Responses could not be calculated.

Standard instruction and multiple sclerosis

SI and MS participants were asked to perform the TOMM-C to the best of their ability. The MS group were also asked to perform the Nine Hole Peg Test, the Symbol Digit Modalities Test – SDMT and the Auditory Verbal Learning Test - AVLT. The SDMT (Sousa et al., 2021) and AVLT (Cavaco et al., 2015) were adjusted to the demographic characteristics of the subjects according to the available norms.

Naïve and coached simulators

Both NS and CS participants were read the a scenario in which they were asked to imagine experiencing real memory difficulties and feeling no longer competent to carry out their work; and to request disability benefits they would need to go through a neuropsychological assessment and convince the examiner of their disability by highlighting their memory difficulties in a credible way. Following the literature (Frederick and Foster, 1991; Rüsseler et al., 2008; Jones, 2017), the CS participants additionally received a series of suggestions to produce the most severe memory problems without making it too obvious to the examiner.

Statistical analyses

Descriptive statistics and nonparametric tests (Chi-square test or Fisher’s exact test and Mann–Whitney test) were used for characterization and comparison of the groups. Effect sizes were calculated and interpreted as follows: 0.2 (small), 0.5 (medium), and 0.8 (large). Receiver operating characteristic (ROC) curves were applied to differentiate SI and NS participants on each measure and trial. The area under the curve was calculated. By design, PVTs prioritize specificity over sensitivity and it is recommended that PVTs have at least 90% specificity when applied to individuals with evidence of significant cognitive dysfunction (Sherman et al., 2020; Sweet et al., 2021). Therefore, the specificity was set at 90%. The sensitivity, positive predictive value and negative predictive value were calculated. The cut-off scores were then used to identify frequency of abnormal score in MS and CS groups.

Multiple logistic regression analyses were used to explore the association between abnormal TOMM score performance and group, while taking into consideration demographic characteristics. TOMM score (recoded according to the cut-off) was the dependent variable, whereas group (i.e., SI vs. MS), sex, age, and years of education were the independent variables. No variable selection was applied and basic assumptions were verified. Simple logistic regression analyses were used to explore in MS group the association between performance on the SDMT and AVLT and some TOMM-C measures.

Results

Groups characteristics

As presented in Table 1, the SI group (n = 31) and the NS group (n = 21) had similar demographic characteristics, namely sex, age, and education. SI group was younger than the MS group (n = 14; p = 0.018) and had fewer years of education than the CS group (n = 8; p = 0.031). NS individuals were younger (p < 0.001) and had more years of education (p = 0.011) than MS patients; and were younger than CS (p = 0.029) participants. No group differences were recorded regarding sex.

Table 1
www.frontiersin.org

Table 1. Demographic characterization and TOMM-C behavioral and eye-fixation scores.

MS patients median T-score (25th, 75th percentiles) on the SDMT was 44.1 (31.0, 50.5). The median adjusted score (25th, 75th percentiles) of the AVLT-Delayed Recall was −1.0 (−1.7, 0.2). These adjusted scores correspond to the number of standard deviations below/above the mean of the participant’s normal peers with the same sex, age, and education. Three MS patients (21.4%) scored below the estimated 18th percentile on AVLT-Delayed Recognition for the demographic characteristics of each individual.

TOMM-C performance

As shown in Table 1, the NS group produced fewer correct responses (large effect sizes), had longer total response time and median response time on correct responses (small effect sizes), higher % of total number of fixations on “new” and % of total fixation time on “new” stimuli (medium effect sizes for Trials 1 and Retention, and small effects for Trial 2), and % of fixation time on “new” stimuli for correct responses (small effect sizes on all trials) in comparison to the SI group on all TOMM-C trials.

In comparison to the SI group, MS participants were slower to respond (i.e., total response time and median response time for correct responses) on all trials (p < 0.001), but had similar number of correct responses. The effect sizes for the response time measures were relatively small. On both Trial 1 (p < 0.06) and Retention Trial (p < 0.01), the % of total number of fixations, the % of total fixation time on “new,” and the % of fixation time on “new” for correct responses was higher for the MS group than the SI group, with a small effect size. The MS group only differed from the NS group on the number of correct responses (all p < 0.001 and with large effect sizes). Neither response time nor oculomotor measures differed between MS and NS participants.

NS and CS groups did not differ on any of TOMM-C behavioral and oculomotor measures. As shown in Table 1, the comparison between SI group and CS group revealed significant differences on all measures, but only partially (i.e., not on all trials). Regarding the number of correct responses, a medium effect size was observed for Trial 1 and large effect sizes were recorded for Trials 2 and Retention. For both Total Response Time and Median Response Time for Correct Responses, the effect sizes were relatively small. Medium effect sizes were observed for Trial 1 on the % of Total Number of Fixation on “New” and % of Total Fixation Time on “New.” Small effect sizes were also recorded on these measures at Retention Trial.

Table 2 shows the best cut-off scores to differentiate SI and NS participants, while setting the specificity at >90%. These cut-off scores were then used to identify the frequency of abnormal scores in the MS group and CS group.

Table 2
www.frontiersin.org

Table 2. Diagnostic statistics of TOMM-C behavioral and eye-fixation measures, in the comparison between healthy with standard Instruction (SI) and healthy with naïve simulation instruction (NS) groups.

Multiple logistic regression analyses revealed that, while taking into consideration demographic characteristics, MS patients had higher odds of abnormal score than SI participants on Total Response Time and Median Response Time for Correct Responses at Trial 1 (respectively adjusted odds = 6.441, p = 0.076 and adjusted odds = 25.027, p = 0.016) and Trial 2 (respectively adjusted odds = 11.001, p = 0.008 and adjusted odds = 4.476, p = 0.086). MS patients also had higher odds of abnormal score at Trial 1 on the following oculomotor measures: % of Total Number of Fixation on “New” (adjusted odds = 44.085, p = 0.005), % of Total Fixation Time on “New” (adjusted odds = 34.961, p = 0.003), and % of Fixation Time on “New” for Correct Responses (adjusted odds = 40.412, p = 0.007). No statistically significant difference (p > 0.05) on oculomotor measures was found between SI and MS participants on Trial 2 and Retention trial, when demographic characteristics were considered.

Simple logistic regressions were used to explore in MS patients the association between standard neuropsychological measures (i.e., SDMT and AVLT) and the following TOMM-C measures: Total Response Time, Median Response Time for Correct Responses, and % of Total Fixation Time on “New.” No significant association was found (p > 0.05).

Discussion

Study results revealed that both behavioral responses (i.e., response accuracy and response time) and eye-fixation data can distinguish simulators from non-simulators in a computerized version of TOMM. Healthy simulators were asked to imagine experiencing “real” memory problems and needing to exaggerate their cognitive difficulties to obtain disability benefits.

Eye-fixation recordings of the SI group showed a familiarity preference (i.e., shorter fixation time on “new” stimuli), especially on Trial 1, whereas both simulator groups showed a novelty preference (i.e., longer fixations on “new” stimuli than on previously presented stimuli) on the three TOMM-C trials. The eye-fixation measure with the best diagnostic statistics in differentiating SI from NS participants was % of Total Fixation Time on “New” (sensitivity of 81.0% and specificity of 93.5%). These findings are consistent with a recent study (Tomer et al., 2020), which revealed that simulators spent more time gazing at foils than target stimuli in another PVT - the Word Memory Test. In non-clinical samples, a novelty preference appears to be a marker of non-credible performance on PVTs that require forced-choice recognition. It has also been suggested that visual disengagement (i.e., gaze aversion) may be used by simulators to attenuate visual input and thereby decrease the cognitive load that they may be experiencing while performing the test (Tomer et al., 2020). Though, gaze aversion could not be documented in the present study, because only two areas of interest - AOI (i.e., the screen was divided in two - “old”/ “new” drawings) were considered, fixations in non-relevant spaces within each AOI were considered on target, and fixations outside the two AOI were discarded. Future studies ought to explore in greater detail the viewing pattern during the performance of TOMM.

Forced-choice memory recognition PVTs (e.g., TOMM and Word Memory Test) share some resemblance with visual-paired comparison (VPC) tasks, which were designed to measure infant recognition memory. Both typically involve a familiarization phase followed by a test phase. During the familiarization phase, the individual is presented with a set of visual stimuli. During the test phase, the familiarization stimulus is paired with a novel stimulus. On VPC tasks, the spontaneous eye-movements are recorded and the amount of time spent looking at each stimulus during the test phase is usually the primary dependent variable. A decreased attention to familiar patterns relative to novel ones (i.e., spending more time looking at novel images) has been observed in VPC applied to preverbal human infants (Fantz, 1964), human adults (Manns et al., 2000), and primates (Pascalis and Bachevalier, 1999). VPC may also elicit a preference for familiarity depending on the length of the retention interval (Bahrick and Pickens, 1995; Richmond et al., 2007). Unlike standard VPC tasks, forced-choice memory recognition tests require an explicit recognition instruction and the visual behaviour of healthy adults during the test phase has been shown to favour familiar stimuli (Richmond et al., 2007; Brooks et al., 2023). Both the preference for novelty and the preference for familiarity are usually interpreted as evidence of recognition memory, whereas null preferences can be interpreted as evidence of forgetting (Richmond et al., 2007).

MS patients exhibited a less evident familiarity preference on the eye-fixation data than the SI participants on Trial 1. It’s unclear why half of the patients with MS showed a preference for the “New” stimuli, as measured by the % of Total Fixation Time on “New.” Both the preference for novelty and the preference for familiarity are usually interpreted as evidence of recognition memory, whereas null preferences can be interpreted as evidence of forgetting (Richmond et al., 2007). It is reasonable to speculate that MS patients were more alert to the possibility that novel stimuli might be relevant, because of their prior experience with neuropsychological assessments (for clinical purposes) that require recall and recognition of previously presented stimuli without prior warning. In the MS group, the % of Total Fixation Time on “New” was not related to measures of visual working memory/psychomotor speed (i.e., SDMT) and verbal memory (i.e., AVLT Delayed Recall and Delayed Recognition), even though patients’ performance on these standard neuropsychological measures was as expected mildly below the norm (Martins Da Silva et al., 2015). Future studies ought to explore the preference for familiarity / novelty in bona fide MS patients and in other clinical populations and to investigate their associations with standard measures of memory (both visual and verbal).

The number of Correct Responses produced the most robust diagnostic statistics and the identified cut-off scores are consistent with most studies in the literature that explored simulation in healthy individuals (for a review see: Martin et al., 2020). Healthy individuals feigning memory impairment significantly produced fewer correct responses on TOMM-C than the SI group. NS and CS performance on TOMM-C approached chance level, especially on Trial 1 and Retention Trial. A ceiling effect was observed in healthy individuals with credible performance (Rees et al., 1998). The number of Correct Responses was similar between MS patients and SI healthy individuals on all trials. Furthermore, only the number of Correct Responses clearly differentiated MS patients from NS participants. These results provide support to its use in clinical practice, namely in patients with MS, cognitive complaints, and mild memory difficulties.

Response time differentiated SI participants from both simulator groups, confirming previous reports (Bolan et al., 2002; Kanser et al., 2019) that healthy simulators are slower to respond on TOMM. However, MS patients were also slower to respond than healthy individuals under SI condition and had similar latency to the NS group. These results highlight the need for caution when applying response time as a performance validity measure in clinical populations, namely in MS which is known to produce processing speed deficits in most patients (Ruano et al., 2017). Nonetheless, in the present study no clear association was found between response time on TOMM-C and a standard measure of visual working memory and psychomotor speed (i.e., SDMT).

No effect of coaching how to avoid detection of invalid performance was observed on any of the behavioural and eye-fixation measures of TOMM-C. In other words, the performance of NS and CS participants did not differ, which may reflect lack of statistical power or resistance of the test to coaching (Jelicic et al., 2011). Larger samples are necessary to confirm these negative findings.

The simultaneous recording of both behavioral and eye-fixation measures in one of the most widely used PVTs, the exploration of different experimental conditions, and the inclusion of a clinical sample are strengths of the study. Unfortunately, the inclusion of participants was cut short due to equipment failure. As ensuing, the small size of the studied groups and the demographic differences of the groups (namely regarding age and education) limit the informative value of group comparisons and the generalizability of the research findings. Though, the literature has recorded minimal or no effects of age or education on TOMM performance (Rees et al., 1998; Rai and Erdodi, 2021; Tchienga and Golden, 2022). Furthermore, the characteristics of the clinical group were not ideal, because none of the MS participants with cognitive complaints had a diagnosis of dementia and not all had memory impairment. Future studies ought to study other clinical aetiologies and suspected clinical malingers.

Recent studies with pupillometry have reported that pupil dilation can detect feigned cognitive impairment on TOMM (Heaver and Hutton, 2011; Patrick et al., 2021a,b). However, the present study focused only on eye-fixations. Future studies should explore the possibility of combining pupil reactivity with eye-fixation pattern in the detection of deception. The standardization of the viewing period (3 s) prior to the behavioral response facilitated the comparison between participants, though it also limited the informative value of the response time.

In sum, healthy individuals feigning memory impairment showed a distinct behavioral (i.e., fewer correct responses and longer response times) and oculomotor (i.e., longer fixation time on “new” stimuli) response pattern on a computerized version of TOMM, which may reflect an increased effort to inhibit a natural response. Further investigation is necessary to understand the potential application of response time and eye-fixation measures in real-life clinical situations.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by Comissão de Ética do Centro Hospitalar Universitário de Santo António. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

FG: Writing – original draft, Writing – review & editing, Investigation. IF: Investigation, Writing – review & editing, Writing – original draft. BR: Software, Writing – review & editing, Writing – original draft. AM: Methodology, Writing – review & editing, Writing – original draft. SC: Methodology, Funding acquisition, Supervision, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported by Bial Foundation Grant 430/14 UIDB/00215/2020; UIDP/00215/2020; LA/P/0064/2020.

Acknowledgments

We would like to express our gratitude for Ana Filipa Gerós and Paulo de Castro Aguiar contribution to data processing.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2024.1395434/full#supplementary-material

References

Bahrick, L. E., and Pickens, J. N. (1995). Infant memory for object motion across a period of three months: implications for a four-phase attention function. J. Exp. Child Psychol. 59, 343–371. doi: 10.1006/jecp.1995.1017

PubMed Abstract | Crossref Full Text | Google Scholar

Bolan, B., Foster, J. K., Schmand, B., and Bolan, S. (2002). A comparison of three tests to detect feigned amnesia: the effects of feedback and the measurement of response latency. J. Clin. Exp. Neuropsychol. 24, 154–167. doi: 10.1076/jcen.24.2.154.1000

Crossref Full Text | Google Scholar

Brooks, G., Whitehead, H., and Kӧhler, S. (2023). When familiarity not novelty motivates information-seeking behaviour. Sci. Rep. 13:5201. doi: 10.1038/s41598-023-31953-6

PubMed Abstract | Crossref Full Text | Google Scholar

Cavaco, S., Gonçalves, A., Pinto, C., Almeida, E., Gomes, F., Moreira, I., et al. (2015). Auditory verbal learning test in a large nonclinical Portuguese population. Appl. Neuropsychol. Adult 22, 321–331. doi: 10.1080/23279095.2014.927767

Crossref Full Text | Google Scholar

Chafetz, M. D., and Biondolillo, A. M. (2013). Feigning a severe impairment profile. Arch. Clin. Neuropsychol. 28, 205–212. doi: 10.1093/arclin/act015

PubMed Abstract | Crossref Full Text | Google Scholar

Fantz, R. L. (1964). Visual experience in infants: decreased attention to familiar patterns relative to novel ones. Science 146, 668–670. doi: 10.1126/science.146.3644.668

PubMed Abstract | Crossref Full Text | Google Scholar

Frederick, R. I., and Foster, H. G. (1991). Multiple measures of malingering on a forced-choice test of cognitive ability. Psychol. Assess. J. Consult. Clin. Psychol. 3, 596–602. doi: 10.1037/1040-3590.3.4.596

Crossref Full Text | Google Scholar

Heaver, B., and Hutton, S. B. (2011). Keeping an eye on the truth? Pupil size changes associated with recognition memory. Memory 19, 398–405. doi: 10.1080/09658211.2011.575788

PubMed Abstract | Crossref Full Text | Google Scholar

Heilbronner, R. L., Sweet, J. J., Morgan, J. E., Larrabee, G. J., and Millis, S. R.Conference Participants (2009). American academy of clinical neuropsychology consensus conference statement on the neuropsychological assessment of effort, response Bias, and malingering. Clin. Neuropsychol. 23, 1093–1129. doi: 10.1080/13854040903155063

PubMed Abstract | Crossref Full Text | Google Scholar

Horner, M. D., Denning, J. H., and Cool, D. L. (2023). Self-reported disability-seeking predicts PVT failure in veterans undergoing clinical neuropsychological evaluation. Clin. Neuropsychol. 37, 387–401. doi: 10.1080/13854046.2022.2056923

Crossref Full Text | Google Scholar

Jelicic, M., Ceunen, E., Peters, M. J. V., and Merckelbach, H. (2011). Detecting coached feigning using the test of memory malingering (TOMM) and the structured inventory of malingered symptomatology (SIMS). J. Clin. Psychol. 67, 850–855. doi: 10.1002/jclp.20805

PubMed Abstract | Crossref Full Text | Google Scholar

Jones, S. M. (2017). Dissimulation strategies on standard neuropsychological tests: a qualitative investigation. Brain Inj. 31, 1131–1141. doi: 10.1080/02699052.2017.1283444

Crossref Full Text | Google Scholar

Kanser, R. J., Bashem, J. R., Patrick, S. D., Hanks, R. A., and Rapport, L. J. (2020). Detecting feigned traumatic brain injury with eye tracking during a test of performance validity. Neuropsychology 34, 308–320. doi: 10.1037/neu0000613

PubMed Abstract | Crossref Full Text | Google Scholar

Kanser, R. J., Rapport, L. J., Bashem, J. R., and Hanks, R. A. (2019). Detecting malingering in traumatic brain injury: combining response time with performance validity test accuracy. Clin. Neuropsychol. 33, 90–107. doi: 10.1080/13854046.2018.1440006

PubMed Abstract | Crossref Full Text | Google Scholar

Manns, J. R., Stark, C. E. L., and Squire, L. R. (2000). The visual paired-comparison task as a measure of declarative memory. Proc. Natl. Acad. Sci. 97, 12375–12379. doi: 10.1073/pnas.220398097

PubMed Abstract | Crossref Full Text | Google Scholar

Manor, B. R., and Gordon, E. (2003). Defining the temporal threshold for ocular fixation in free-viewing visuocognitive tasks. J. Neurosci. Methods 128, 85–93. doi: 10.1016/S0165-0270(03)00151-1

Crossref Full Text | Google Scholar

Martin, P. K., Schroeder, R. W., Olsen, D. H., Maloy, H., Boettcher, A., Ernst, N., et al. (2020). A systematic review and meta-analysis of the test of memory malingering in adults: two decades of deception detection. Clin. Neuropsychol. 34, 88–119. doi: 10.1080/13854046.2019.1637027

Crossref Full Text | Google Scholar

Martins Da Silva, A., Cavaco, S., Moreira, I., Bettencourt, A., Santos, E., Pinto, C., et al. (2015). Cognitive reserve in multiple sclerosis: protective effects of education. Mult. Scler. J. 21, 1312–1321. doi: 10.1177/1352458515581874

PubMed Abstract | Crossref Full Text | Google Scholar

Pascalis, O., and Bachevalier, J. (1999). Neonatal aspiration lesions of the hippocampal formation impair visual recognition memory when assessed by paired-comparison task but not by delayed nonmatching-to-sample task. Hippocampus 9, 609–616. doi: 10.1002/(SICI)1098-1063(1999)9:6<609::AID-HIPO1>3.0.CO;2-A

PubMed Abstract | Crossref Full Text | Google Scholar

Patrick, S. D., Rapport, L. J., Kanser, R. J., Hanks, R. A., and Bashem, J. R. (2021a). Detecting simulated versus bona fide traumatic brain injury using pupillometry. Neuropsychology 35, 472–485. doi: 10.1037/neu0000747

PubMed Abstract | Crossref Full Text | Google Scholar

Patrick, S. D., Rapport, L. J., Kanser, R. J., Hanks, R. A., and Bashem, J. R. (2021b). Performance validity assessment using response time on the Warrington recognition memory test. Clin. Neuropsychol. 35, 1154–1173. doi: 10.1080/13854046.2020.1716997

PubMed Abstract | Crossref Full Text | Google Scholar

Polman, C. H., Reingold, S. C., Banwell, B., Clanet, M., Cohen, J. A., Filippi, M., et al. (2011). Diagnostic criteria for multiple sclerosis: 2010 revisions to the McDonald criteria. Ann. Neurol. 69, 292–302. doi: 10.1002/ana.22366

PubMed Abstract | Crossref Full Text | Google Scholar

Rai, J. K., and Erdodi, L. A. (2021). Impact of criterion measures on the classification accuracy of TOMM-1. Appl. Neuropsychol. Adult 28, 185–196. doi: 10.1080/23279095.2019.1613994

PubMed Abstract | Crossref Full Text | Google Scholar

Rees, L. M., Tombaugh, T. N., Gansler, D. A., and Moczynski, N. P. (1998). Five validation experiments of the test of memory malingering (TOMM). Psychol. Assess. 10, 10–20. doi: 10.1037/1040-3590.10.1.10

Crossref Full Text | Google Scholar

Richmond, J., Colombo, M., and Hayne, H. (2007). Interpreting visual preferences in the visual paired-comparison task. J. Exp. Psychol. Learn. Mem. Cogn. 33, 823–831. doi: 10.1037/0278-7393.33.5.823

Crossref Full Text | Google Scholar

Ruano, L., Portaccio, E., Goretti, B., Niccolai, C., Severo, M., Patti, F., et al. (2017). Age and disability drive cognitive impairment in multiple sclerosis across disease subtypes. Mult. Scler. J. 23, 1258–1267. doi: 10.1177/1352458516674367

PubMed Abstract | Crossref Full Text | Google Scholar

Rüsseler, J., Brett, A., Klaue, U., Sailer, M., and Münte, T. F. (2008). The effect of coaching on the simulated malingering of memory impairment. BMC Neurol. 8:37. doi: 10.1186/1471-2377-8-37

PubMed Abstract | Crossref Full Text | Google Scholar

Schroeder, R. W., Clark, H. A., and Martin, P. K. (2022). Base rates of invalidity when patients undergoing routine clinical evaluations have social security disability as an external incentive. Clin. Neuropsychol. 36, 1902–1914. doi: 10.1080/13854046.2021.1895322

Crossref Full Text | Google Scholar

Sherman, E. M. S., Slick, D. J., and Iverson, G. L. (2020). Multidimensional malingering criteria for neuropsychological assessment: a 20-year update of the malingered neuropsychological dysfunction criteria. Arch. Clin. Neuropsychol. 35, 735–764. doi: 10.1093/arclin/acaa019

PubMed Abstract | Crossref Full Text | Google Scholar

Slick, D. J., Sherman, E. M. S., and Iverson, G. L. (1999). Diagnostic criteria for malingered neurocognitive dysfunction: proposed standards for clinical practice and research. Clin. Neuropsychol. 13, 545–561. doi: 10.1076/1385-4046(199911)13:04;1-Y;FT545

PubMed Abstract | Crossref Full Text | Google Scholar

Sousa, C., Rigueiro-Neves, M., Passos, A. M., Ferreira, A., and Sá, M. J.Group for Validation of the BRBN-T in the Portuguese MS Population (2021). Assessment of cognitive functions in patients with multiple sclerosis applying the normative values of the Rao’s brief repeatable battery in the Portuguese population. BMC Neurol. 21:170. doi: 10.1186/s12883-021-02193-w

PubMed Abstract | Crossref Full Text | Google Scholar

Sweet, J. J., Heilbronner, R. L., Morgan, J. E., Larrabee, G. J., Rohling, M. L., Boone, K. B., et al. (2021). American Academy of clinical neuropsychology (AACN) 2021 consensus statement on validity assessment: update of the 2009 AACN consensus conference statement on neuropsychological assessment of effort, response bias, and malingering. Clin. Neuropsychol. 35, 1053–1106. doi: 10.1080/13854046.2021.1896036

PubMed Abstract | Crossref Full Text | Google Scholar

Tchienga, I., and Golden, C. (2022). A-235 the degree to which age, education and race predict TOMM performance in a retired NFL cohort. Arch. Clin. Neuropsychol. 37:1391. doi: 10.1093/arclin/acac060.235

Crossref Full Text | Google Scholar

Tombaugh, T. N. (1996). Test of memory malingering (TOMM). North Tonawanda, NY: Multi Health Systems.

Google Scholar

Tomer, E., Lupu, T., Golan, L., Wagner, M., and Braw, Y. (2020). Eye tracking as a mean to detect feigned cognitive impairment in the word memory test. Appl. Neuropsychol. Adult 27, 49–61. doi: 10.1080/23279095.2018.1480483

Crossref Full Text | Google Scholar

Keywords: malingering, novelty preference, familiarity preference, eye-tracking, performance validity tests

Citation: Gomes F, Ferreira I, Rosa B, Martins da Silva A and Cavaco S (2024) Using behavior and eye-fixations to detect feigned memory impairment. Front. Psychol. 15:1395434. doi: 10.3389/fpsyg.2024.1395434

Received: 03 March 2024; Accepted: 25 July 2024;
Published: 20 September 2024.

Edited by:

Elisa Cavicchiolo, University of Rome Tor Vergata, Italy

Reviewed by:

Mark Ettenhofer, University of California, San Diego, United States
Quan Wang, Chinese Academy of Sciences (CAS), China

Copyright © 2024 Gomes, Ferreira, Rosa, Martins da Silva and Cavaco. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Sara Cavaco, c2FyYS5jYXZhY29AY2hwb3J0by5taW4tc2F1ZGUucHQ=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.