- 1Medical Psychological Center, The Second Xiangya Hospital, Central South University, Changsha, Hunan, China
- 2Institute of Mental Health, Peking University Sixth Hospital, Peking University, Beijing, China
- 3NHC Key Laboratory for Mental Health, National Clinical Research Center for Mental Disorders, Peking University Sixth Hospital, Peking University, Beijing, China
- 4Department of Psychiatry, Sir Run Run Shaw Hospital, Zhejiang University School of Medicine, Hangzhou, Zhejiang, China
- 5Medical Psychological Institute, Central South University, Changsha, Hunan, China
- 6National Clinical Research Center for Mental Disorders, Changsha, Hunan, China
Background: The aging population is increasing, making it essential to have a standardized, convenient, and valid electronic memory test that can be accessed online for older people and caregivers. The electronic version of the Hopkins Verbal Learning Test-Revised (HVLT-R) as a test with these advantages and its reliability and validity has not yet been tested. Thus, this study examined the reliability and validity of the electronic version of the HVLT-R in middle-aged and elderly Chinese people to provide a scientific basis for its future dissemination and use.
Methods: We included 1,925 healthy participants aged over 40, among whom 38 were retested after 3–6 months. In addition, 65 participants completed both the pad and paper-and-pencil versions of the HVLT-R (PAP-HVLT-R). We also recruited 42 Alzheimer’s disease (AD) patients, and 45 amnestic mild cognitive impairment (aMCI) patients. All participants completed the Pad-HVLT-R, the Hong Kong Brief Cognitive Test (HKBC), the Brief Visual Memory Test-Revised (BVMT-R), and the Logical Memory Test (LM).
Results: (1) Reliability: the Cronbach’s α value was 0.94, the split-half reliability was 0.96. The test–retest correlation coefficients were moderate, ranging from 0.38 to 0.65 for direct variables and 0.16 to 0.52 for derived variables; (2) Concurrent validity: the Pad-HVLT-R showed a moderate correlation with the HKBC and BVMT-R, with correlation coefficients between total recall of 0.41 and 0.54, and between long-delayed recall of 0.42 and 0.59, respectively. It also showed a high correlation with the LM, with correlation coefficients of 0.72 for total recall and 0.62 for long-delayed recall; (3) Convergent validity: the Pad-HVLT-R was moderately correlated with the PAP version, with correlation coefficients ranging from 0.29 to 0.53 for direct variables and 0.15 to 0.43 for derived variables; (4) Discriminant capacity: the Pad-HVLT-R was effective in differentiating AD patients, as demonstrated by the ROC analysis with AUC values of 0.834 and 0.934 for total recall and long-delayed recall, respectively.
Conclusion: (1) The electronic version of HVLT-R has good reliability and validity in middle-aged and elderly Chinese people; (2) The electronic version of HVLT-R can be used as an effective tool to distinguish AD patients from healthy people.
1. Introduction
According to a recent report by the United Nations (World Population Prospects 2022: 10 Key; United Nations, 2022), China is currently undergoing an aging process due to rising life expectancy and declining fertility rates. As people age, various cognitive functions, especially memory, tend to decline. Studies have shown that memory exhibits a decreasing tendency with age, starting around age of 50 and becoming more marked after age 70 (Xu et al., 1985). Episodic memory, which is the most age-sensitive of the long-term memory systems (Nyberg, 1996), typically develops rapidly in childhood but tends to decline in adulthood (Kausler, 1994; Schneider and Pressley, 2013), with more rapid declines in very old age (Singer et al., 2003).
The Hopkins Verbal Learning Test-Revised (HVLT-R; Benedict et al., 1998) is a popular neuropsychological test for episodic memory (Rabin et al., 2016), comprising 12 items. It is brief, multi-dimensional, easily administered, and well accepted test with no ceiling effect. Therefore, it has the potential to be an ideal assessment instrument for short-term memory assessments or series of tests. A number of studies have supported the reliability and validity of the HVLT-R, including evidence of the test–retested reliability (Benedict et al., 1998; Barr, 2003; O’Neil-Pirozzi et al., 2012; Shi et al., 2015), concurrent validity (Shapiro et al., 1999), convergent validity (Lacritz et al., 2001; Kordes, 2004), and discriminant capacity (Frank and Byrne, 2000; Foster et al., 2009; Lonie et al., 2010; González-Palau et al., 2013). The HVLT-R has been demonstrated to be suitable for different ethnic groups and individuals with significant cognitive impairment (Shi et al., 2012; Scott et al., 2020; Díaz-Santos et al., 2021), with high classification accuracy and effectiveness in evaluating healthy elderly populations (Ryan et al., 2021), patients with mild cognitive impairment (MCI; González-Palau et al., 2013; Hammers et al., 2022), Alzheimer’s disease (AD) patients (Shapiro et al., 1999; Gómez-Gallego and Gómez-García, 2019; Hammers et al., 2022), and dementia patients (Frank and Byrne, 2000; Hogervorst et al., 2002; Foster et al., 2009; Liao et al., 2019).
As the global population ages, memory decline and impairment become more common. Therefore, it is increasingly important to assess and monitor memory function to detect and intervene early. However, traditional paper-and-pencil tests can be time-consuming and challenging to administer remotely. In recent years, the development of electronic and web-based memory tests has provided a convenient and accessible solution for assessing memory function. For example, Lewandowsky et al. (2010) developed an electronic version of a working memory test battery that has been shown to be reliable and valid for assessing working memory function. Additionally, the Cognitive Assessment for Dementia, iPad version (CADi), developed in Japan for large-scale dementia screening, can be run on an iPad and includes immediate and delayed recall of three words (Onoda et al., 2013). The electronic and online tests not only provide a standardized and reliable way to assess memory function, but also have the potential to reach a wider population and enable more frequent monitoring of memory function over time.
However, concerns have been raised about the reliability and validity of electronic tests compared to their traditional counterparts. While some studies have shown that electronic tests have good reliability and validity and can be compared to their traditional paper-and-pencil counterparts (Hoskins et al., 2010; Vanderslice-Barr et al., 2011), other researchers have found that participants’ computer anxiety, familiarity with electronic devices, and motor coordination may affect the results of the test, and these factors may reduce the reliability of the electronic test compared to the paper-and-pencil version (Whitener and Klein, 1995; Barrigón et al., 2017). Therefore, when an electronic test is administered, the reliability and validity of the electronic version should be examined, rather than relying solely on the previous paper-and-pencil results.
Given the many advantages of electronic tests, a pad version of the HVLT-R has recently been developed, which can automatically read out the standardized instructions and records the answers. However, its reliability and validity are unknown. As such, the present study sought to examine the reliability and validity of the electronic version of the HVLT-R in middle-aged and elderly Chinese individuals, providing a scientific basis for its future use in research or clinical settings.
2. Methods
2.1. Participants
A total of 2,077 participants were recruited by convenience sampling in communities or clinics from 6 regions in China between May 2019 and June 2021. All subjects completed the Chinese Neuropsychological Consensus Battery (CNCB) in its electronic version (Wang et al., 2019). This study is part of the larger Chinese Neuropsychological Normative (CN-NORM) project led by the Dementia Care & Research Center, Peking University Institute of Mental Health (Sixth Hospital). All subjects and their families signed informed consent.
2.1.1. Healthy sample
Subjects had to be over 40 years of age and were excluded if they scored <7 on the brief Community Screening Instrument for Dementia (CSI-D; Prince et al., 2011), had any medical condition that might interfere with cognition or had a related neurological disorder or major psychiatric disorder that could impact cognitive function (e.g., severe audiovisual impairment; history of drug or alcohol abuse; clear history of severe cerebrovascular disease (including cerebral hemorrhage and cerebral infarction), Parkinson’s syndrome, epilepsy, bipolar disorder, and other neuropsychiatric disorders; severe head trauma, carbon monoxide poisoning).
Depending on the administration procedure, we divided the healthy subjects into three groups as follows:
Sample 1: a total of 2,146 middle-aged and older adults completed the electronic version of the test battery. After excluding 221 subjects with a large number of missing values of the Pad-HVLT-R (e.g., incomplete short-delayed recall, long-delayed recall, or recognition trial), the final sample size was 1,925 people, resulting in a valid sample rate of 90%.
Sample 2: 44 subjects from Sample 1 were selected to complete the electronic version of the test battery again 3–6 months later to examine the test–retest reliability, of which 2 participants who did not complete the delayed recall were removed from analysis.
Sample 3: 66 subjects were administered both the electronic and PAP version of the test battery to examine the convergent validity. To counteract the sequential effect, a balanced design was used in this study. The 66 participants were divided into two groups, with the first group completing the PAP version of the test battery in the first phase and the electronic test later, while the second group did the opposite. One case was excluded for failure to complete the long-delayed recall of the Pad-HVLT-R. The interarrival time between two phases is shown in Table 1.
2.1.2. Clinic sample
A total of 100 subjects were enrolled, including 51 patients with amnestic mild cognitive impairment (aMCI) and 49 patients with AD. Among them, 6 aMCI patients and 7 AD patients failed to complete the immediate recall or delayed recall and were deleted. Finally, valid samples were collected: 45 aMCI patients and 42 AD patients. The details on the AD/MCI determination have been described in a previous study (Gu et al., 2023). Briefly, the participants completed a standardized neuropsychological assessment, underwent clinical interviews and brain imaging examinations, and received a clinical diagnosis by a memory specialist.
The inclusion criteria for patients with aMCI were: (1) Age ≥ 40; (2) Met the 2004 Petersen diagnostic criteria for aMCI (Petersen, 2004); (3) Can understand instructions; (4) Hachinski Ischemia Scale ≤4; (5) Geriatric Depression Scale ≤10.
The inclusion criteria for patients with AD were: (1) Age ≥ 40; (2) Met the diagnostic criteria of “probable AD dementia” of the National Institute on Aging-Alzheimer’s Association (NIA-AA, 2011; Jack et al., 2011); (3) Can understand instructions; (4) Hachinski Ischemia Scale ≤4; (5) Geriatric Depression Scale ≤10.
The exclusion criteria for both aMCI and AD patients were: (1) Unwilling to cooperate; (2) Severe audiovisual impairment; (3) History of drug or alcohol abuse; (4) A clear history of severe cerebrovascular disease (including cerebral hemorrhage and cerebral infarction), Parkinson’s syndrome, epilepsy, bipolar disorder, and other neuropsychiatric disorders; (5) Had severe head trauma, carbon monoxide poisoning, etc.
2.2. Instruments
2.2.1. The general information questionnaire
The questionnaire included: age, sex, years of education, marital status, prior psychiatric history, family history of dementia, history of physical diseases, etc.
2.2.2. Hopkins verbal learning test-revised
The CNCB included administration of the HVLT-R. The electronic version of the HVLT-R requires the subject to complete the test via a Pad. At the start of the test, the Pad automatically reads out the instructions in standard Mandarin, and when the subject clicks to confirm the test, the Pad starts to read out the list at a rate of one word every 2 s, and three immediate recall trials are performed. After 5 and 20 min, the Pad automatically performs the short and long-delayed recall trials. The long-delayed recall trial was followed by a recognition trial. During the learning and delayed recall trials, the screen displays 12 word buttons, with additional buttons such as “repeat,” “other word,” “not attempted,” and “undo” located below. The experimenter selects the appropriate button by clicking on it. During the recognition trial, the experimenter reads out the 24 words on the screen in sequence and selects the correct or incorrect button based on the participant’s response. The administration process of the PAP-HVLT-R can be found in Benedict et al.’s paper (Benedict et al., 1998). The same word lists were used for both PAP and Pad versions of the HVLT-R.
Variables that were directly derived from the number of correct words given in each trial were considered direct variables, while variables that were derived indirectly from ratios were considered derived variables. Table 2 shows a detailed description of each variable.
2.2.3. Other tests
The CNCB also included the Hong Kong Brief Cognitive Test (HKBC; Chiu et al., 2018), the Brief Visuospatial Memory Test-revised (BVMT-R; Benedict et al., 1996), the Logical Memory Test (LM) of the Wechsler Adult Memory Scale-Chinese Revision (WMS-RC; Wang et al., 2015). The LM is an assessment of narrative episodic memory. We used the total score of the HKBC, the scores of total recall (sum of correct responses on three immediate recall trials) and delayed recall of the BVMT-R, and the mean scores of immediate and delayed recall of the LM in our analyses.
Several other non-verbal tests were included in the test battery but were not included in the current study. Subjects completed the non-verbal tests in-between delayed recall sessions to avoid interference.
2.3. Statistical analyses
Statistical analyses were conducted using SPSS 22.0. First, the reliability of the Pad-HVLT-R was evaluated using internal consistency, split-half reliability, and test–retest reliability. The Cronbach’s alpha coefficient for trial 1 to trial 5 was used to measure internal consistency, and split-half reliability was assessed by correlating the two odd-even scores (test 1 + 3 vs. test 2 + 4, test 2 + 4 vs. test 3 + 5) of Sample 1. In addition, we separately calculated the Cronbach’s alpha coefficients for the PAD and PAP data of Sample 3. The test–retest reliability was measured by calculating Pearson’s product–moment correlations between two sessions of Sample 2, and r values of 0.3, 0.5, and 0.7 represent low, medium, and high correlations, respectively (Mukaka, 2012).
The validity of the Pad-HVLT-R was examined through concurrent validity, convergent validity, and discriminant capacity. Concurrent validity was assessed by calculating the correlation coefficients between the scores of Pad-HVLT-R and the corresponding scores of HKBC, BVMT, and LM in Sample 1. Convergent validity was examined by calculating the correlation coefficients for the corresponding variables between the Pad and the PAP version of the HVLT-R in Sample 3. Discriminant capacity was examined by conducting the group comparisons and the receiver operating characteristic curve (ROC) analysis for the healthy group and case with aMCI and AD. Since previous findings have shown a significant effect of age and education on the HVLT (Shi et al., 2012; Duff, 2016; Ryan et al., 2021), in this study we used the Propensity Score Matching (PSM) in SPSS 22.0 to select two healthy groups in Sample 1 that matched the age and education of the aMCI and AD groups, respectively, to reduce the effect of age and education. In the ROC analysis, the cut-off scores were determined by the Youden index, with the highest Youden index being the optimum value for sensitivity and specificity.
In this study, the normal probability plot was used to assess whether or not the data are approximately normally distributed (Ghosh, 1996). For correlation analysis, Pearson’s correlation analysis was used for normal data and Spearman’s correlation analysis was used for non-normality. In the group comparison, the independent samples t-test was used for normal data and the Mann–Whitney U-test was used for non-normality.
3. Results
3.1. Descriptive statistics
Demographic information of the healthy sample is presented in Table 1.
3.2. Reliability
3.2.1. Internal consistency reliability
It is calculated that the Cronbach’s α coefficient of the electronic version of HVLT-R is 0.94. The correlation coefficients of the two odd-even scores are 0.91 and 0.92, respectively, and the split-half reliability is 0.96 calculated by the Spearman-Brown formula. In addition, the electronic version had a Cronbach’s α coefficient of 0.849 and the paper-pencil version had a reliability coefficient of 0.899.
3.2.2. Test–retest reliability
Table 3 presents the raw descriptive data for the electronic version of HVLT-R, along with the r values for the twice examined. The test–retest correlation coefficients were higher for the direct variables overall compared to the derived variables. The highest correlation was observed for total recall (r = 0.62) and the lowest coefficient of retest was observed for the index of RDI (r = 0.38). Whereas among the derived variables, the highest correlation was observed for semantic clustering (r = 0.52), more modest test–retest stability coefficients were observed for serial clustering, perseverations, and intrusions errors, while the correlations for the other derived measures were lower.
3.3. Validity
3.3.1. Concurrent validity
Table 4 presents the correlation coefficients between the Pad-HVLT-R and the HKBC, BVMT-R, and LM. There was a moderate correlation between the HVLT-R, HKBC, and BVMT-R. The total and long-delayed recall of the HVLT-R correlated most strongly with the corresponding variables of the LM.
3.3.2. Convergent validity
Correlation analysis between the PAP and electronic versions of the HVLT-R showed that the correlation coefficients for all direct variables were significant, with only the correlation coefficient for the RDI being slightly below 0.30, while among the derived variables only the LDR percent retained and semantic clustering were above 0.30 (Table 5).
3.3.3. Discriminant capacity
We rigorously compared the performance of the Pad-HVLT-R between HC and aMCI and AD groups. The results showed no significant differences in age, education, and sex ratio between the HC group and the aMCI and AD groups after automatic matching by PSM.
The results of group comparisons showed that the HC group had significantly higher scores than the aMCI group on all direct variables as well as on the long-delayed recall percent retained, semantic cluster ratio, and perseverations of the derived variable, while there were no significant differences between the two groups on other variables. The results of the comparison are shown in Table 6.
The results for the AD and the HC group showed that the HC group scored significantly higher than the AD group on all variables except intrusions errors, the results of which are shown in Table 7.
To further determine the discriminant ability of the Pad-HVLT-R to detect aMCI and AD patients from the healthy population, we calculated the AUC which is typically rated as acceptable (0.70–0.79), good (0.80–0.89), excellent (0.90–0.99) or perfect (1.0; Carter et al., 2016). In distinguishing aMCI and HC, the AUC of all variables was higher than 0.70, among which the AUC of the total recall was the highest, and the optimal sensitivity and specificity were achieved when the cut-off score was 21.50 (out of a maximum score of 36). When distinguishing between AD and HC groups, the AUC was higher than 0.90 for all variables except the total recall. The best sensitivity and specificity were achieved when the cut-off score for the total recall was 16.50. The optimal sensitivity and specificity were achieved when the cut-off score of the long-delayed recall was 1.50 (out of a maximum score of 12). Table 8 shows the AUC, cut-off score, sensitivity, and specificity of the direct variables of the Pad-HVLT-R for the aMCI and AD patients, respectively.
4. Discussion
In this study, we examined the reliability and validity of the electronic version of the Hopkins Verbal Learning Test-Revised in middle-aged and elderly Chinese people, and the results showed that the Pad-HVLT-R had good reliability and validity. In clinical samples, the results showed that the Pad-HVLT-R was effective in assessing verbal memory function and its impairment, and sensitive to changes in memory ability in AD patients, but not as effective in differentiating between aMCI patients and the healthy population.
The Cronbach’s alpha was 0.94 and the split-half reliability was 0.96, indicating good internal consistency reliability of the electronic version of HVLT-R. Although the Cronbach’s α coefficient of the electronic version was slightly lower than that of the paper-pencil version calculated in Sample 3, both versions still exceeded the recommended threshold of 0.70 for internal consistency reliability. Therefore, the findings suggest that the electronic HVLT-R is a reliable tool for episodic memory assessment. The test–retest reliability coefficients of the direct variables were relatively high, with total recall and short-delayed recall being the highest. The results are broadly similar to those described by Barr (2003) and the work of Woods et al. (2005) but lower than those reported by Benedict et al. (1998), which is presumably related to the sample size or the interval between two examine (e.g., the three studies mentioned above had retest intervals of 60, 370, and 47 days). However, the correlation coefficients for some derived variables were relatively low or even negative, except for semantic clustering. Therefore, it is possible to be cautious in interpreting changes in the derived variables of the HVLT-R over time (Woods et al., 2006). Overall, the test–retest reliability of the electronic version of the HVLT-R direct variables is moderate, but of some derived variables is low.
The concurrent validity was assessed by calculating the correlation coefficients of the Pad-HVLT-R with the HKBC, the BVMT-R and the LM, respectively. Scores on the HVLT-R correlated most strongly with the LM scores. The HVLT-R also correlated strongly with the BVMT-R but more modestly correlated with the HKBC. Notably, the LM and the BVMT-R used in this study were administered in electronic format, and their reliability and validity have not yet been established, which could potentially impact the correlations between the tests. Overall, the findings provide sufficient evidence that the electronic version of the HVLT-R has concurrent validity to justify its use in clinical neuropsychological assessments.
Regarding convergent validity, we used a crossover design to control for possible order effects. The results showed moderately significant correlations between the direct variables, with the exception of the recognition discrimination index. In contrast, only moderate correlation coefficients were found between long-delayed recall percent retained and semantic clustering among the derived variables, while all others were relatively low. Due to practical constraints, this part of the psychometric analysis only included healthy participants, which may have led to a restricted range of scores on the variables being measured and could have influenced the results of the convergent validity analysis. Additionally, measurement error may have been introduced due to differences between the two test formats, as well as differences in the context of the two administrations and the time interval between the two administrations, which may result in lower correlation coefficients. In addition, we compared the correlation coefficients for the PAP and Pad with the test–retest coefficients for the Pad-HVLT-R, found that the two were similar and that the correlation coefficients for the PAP and Pad format were lower than the test–retest coefficients except for the perseverations, which maybe since the variations in scores between PAP and Pad came not only from random variation but also from variation in the construct between assessments. The lower number of perseverations on the PAP compared to the Pad version may be because of the fact that the electronic version is more convenient to record, allowing for a quick recording of all of the subject’s responses without omission.
We evaluated the capacity of the electronic version of the HVLT-R to distinguishing aMCI and AD cases from HC group. The results showed significant differences between participants in the impaired groups (aMCI and AD) and healthy controls in all direct variables of the Pad-HVLT-R. As for the derived variables, there were significant differences between the impaired and HC groups except for intrusions errors. All direct variables and most derived variables of the Pad-HVLT-R were able to separate older adults with AD or aMCI from healthy individuals.
When detecting AD from HC, the AUC for all direct variables except total recall were higher than 0.90. The optimal balance between sensitivity and specificity was achieved when the cut-off score for long-delayed recall was 1.50, at which point the specificity was good (0.930) and the sensitivity was moderate (0.791). Compared with other variables, the AUC of total recall (0.811) was slightly lower, with high sensitivity (0.953) and lower specificity (0.628) when the optimal demarcation was 16.50. Overall, the electronic version of the HVLT-R is good at distinguishing between AD and HC.
However, the results showed a relatively low discrimination capacity of the Pad-HVLT-R to detecting aMCI cases from HC group. The AUC for total recall was highest, and the optimal balance between sensitivity (75.6%) and specificity (60.0%) with the total recall score was at a cut-off point of 21.50. The most optimal cut-off scores of the total recall for detecting aMCI and AD from healthy subjects in this study are similar to the results of Shi et al. (2012). They examined the discriminant capacity of the Chinese version of HVLT for MCI and dementia and showed that the optimal cut-off score for discriminating aMCI from the HC group was 21.50, when the sensitivity and specificity were lower (69.1% and 70.7%), whereas the sensitivity and specificity for distinguishing AD from the HC group was high when the cut-off score was 15.50. The results of the present study are broadly similar to those of previous studies on the discriminant capacity of the PAP-HVLT-R for AD and aMCI, which showed high discrimination for AD and relatively lower discrimination for aMCI (de Jager et al., 2003; Schrijnemaekers et al., 2006; Shi et al., 2012).
There are some limitations in this study: (1) There was a design flaw in the computerized test system upfront, leading to the loss of some data due to the failure of a few subjects to complete delayed recall, which has now been corrected; (2) The sample size of AD and aMCI cases included in this study was small and their education level was high (73% of aMCI patients had more than 9 years of education), which limited the follow-up analysis of different ages and education levels; (3) Due to practical constraints, we did not collect data on patients’ performance on the PAP version. This may have affected the results of the convergent validity analysis. In future studies examining the convergent validity of the electronic test, both healthy and clinical participants could be included to complete the PAP and Pad versions of the test, in order to improve the generalizability of the results.
In conclusion, the electronic version of the Hopkins Verbal Learning Test-Revised has good reliability and validity, and can be used in clinical and basic research to assess the verbal memory function of middle-aged and older Chinese people.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
The studies involving human participants were reviewed and approved by the Second Xiangya Hospital of Central South University Ethics Committee. The patients/participants provided their written informed consent to participate in this study.
Author contributions
DW and CS designed this study. LJ and DW analyzed the data and prepared the manuscript. All authors contributed to the article and approved the submitted version.
Funding
This work was supported by the National Key R&D Program of China (2018YFC1314202).
Acknowledgments
We are grateful to the older people in China who participated in this study and to the community workers who helped us contact with these participants.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Barr, W. B. (2003). Neuropsychological testing of high school athletes: preliminary norms and test–retest indices. Arch. Clin. Neuropsychol. 18, 91–101. doi: 10.1016/S0887-6177(01)00185-8
Barrigón, M. L., Rico-Romano, A. M., Ruiz-Gomez, M., Delgado-Gomez, D., Barahona, I., Aroca, F., et al. (2017). Comparative study of pencil-and-paper and electronic formats of GHQ-12, WHO-5 and PHQ-9 questionnaires. Revista de Psiquiatría y Salud Mental. 10, 160–167. doi: 10.1016/j.rpsmen.2017.05.009
Benedict, R. H., Schretlen, D., Groninger, L., and Brandt, J. (1998). Hopkins verbal learning test–revised: normative data and analysis of inter-form and test-retest reliability. Clin. Neuropsychol. 12, 43–55. doi: 10.1076/clin.12.1.43.1726
Benedict, R. H., Schretlen, D., Groninger, L., Dobraski, M., and Shpritz, B. (1996). Revision of the brief Visuospatial memory test: studies of normal performance, reliability, and validity. Psychol. Assess. 8, 145–153. doi: 10.1037/1040-3590.8.2.145
Carter, J. V., Pan, J., Rai, S. N., and Galandiuk, S. (2016). ROC-ing along: evaluation and interpretation of receiver operating characteristic curves. Surgery 159, 1638–1645. doi: 10.1016/j.surg.2015.12.029
Chiu, H. F. K., Zhong, B. L., Leung, T., Li, S. W., Chow, P., Tsoh, J., et al. (2018). Development and validation of a new cognitive screening test: the Hong Kong brief cognitive test (HKBC). Int. J. Geriatr. Psychiatry 33, 994–999. doi: 10.1002/gps.4883
de Jager, C. A., Hogervorst, E., Combrinck, M., and Budge, M. M. (2003). Sensitivity and specificity of neuropsychological tests for mild cognitive impairment, vascular cognitive impairment and Alzheimer’s disease. Psychol. Med. 33, 1039–1050. doi: 10.1017/s0033291703008031
Díaz-Santos, M., Suárez, P. A., Marquine, M. J., Umlauf, A., Rivera Mindt, M., Artiola I Fortuny, L., et al. (2021). Updated demographically adjusted norms for the brief Visuospatial memory test-revised and Hopkins verbal learning test-revised in Spanish-speakers from the U.S.-Mexico border region: the NP-NUMBRS project. Clin. Neuropsychol. 35, 374–395. doi: 10.1080/13854046.2020.1861329
Duff, K. (2016). Demographically corrected normative data for the Hopkins verbal learning test-revised and brief Visuospatial memory test-revised in an elderly sample. Appl. Neuropsychol. Adult 23, 179–185. doi: 10.1080/23279095.2015.1030019
Foster, P. S., Drago, V., Crucian, G. P., Rhodes, R. D., Shenal, B. V., and Heilman, K. M. (2009). Verbal learning in Alzheimer’s disease: cumulative word knowledge gains across learning trials. J. Int. Neuropsychol. Soc. 15, 730–739. doi: 10.1017/S1355617709990336
Frank, R. M., and Byrne, G. J. (2000). The clinical utility of the Hopkins verbal learning test as a screening test for mild dementia. Int. J. Geriatr. Psychiatry 15, 317–324. doi: 10.1002/(SICI)1099-1166(200004)15:4<317::AID-GPS116>3.0.CO;2-7
Ghosh, S. (1996). A new graphical tool to detect non-normality. J R Stat Soc Series B 58, 691–702. doi: 10.1111/j.2517-6161.1996.tb02108.x
Gómez-Gallego, M., and Gómez-García, J. (2019). Stress and verbal memory in patients with Alzheimer’s disease: different role of cortisol and anxiety. Aging Ment. Health 23, 1496–1502. doi: 10.1080/13607863.2018.1506741
González-Palau, F., Franco, M., Jiménez, F., Parra, E., Bernate, M., and Solis, A. (2013). Clinical utility of the Hopkins verbal test-revised for detecting Alzheimer’s disease and mild cognitive impairment in Spanish population. Arch. Clin. Neuropsychol. 28, 245–253. doi: 10.1093/arclin/act004
Gu, D., Wang, L., Zhang, N., Wang, H., and Yu, X. (2023). Decrease in naturally occurring antibodies against epitopes of Alzheimer’s disease (AD) risk gene products is associated with cognitive decline in AD. J. Neuroinflammation. Arch. Clin. Neuropsychol. 20:74. doi: 10.1186/s12974-023-02750-9
Hammers, D. B., Suhrie, K., Dixon, A., Gradwohl, B. D., Duff, K., and Spencer, R. J. (2022). Validation of HVLT-R, BVMT-R, and RBANS learning slope scores along the Alzheimer’s continuum. Arch. Clin. Neuropsychol. 37, 78–90. doi: 10.1093/arclin/acab023
Hogervorst, E., Combrinck, M., Lapuerta, P., Rue, J., Swales, K., and Budge, M. (2002). The Hopkins verbal learning test and screening for dementia. Dement. Geriatr. Cogn. Disord. 13, 13–20. doi: 10.1159/000048628
Hoskins, L. L., Binder, L. M., Chaytor, N. S., Williamson, D. J., and Drane, D. L. (2010). Comparison of oral and computerized versions of the word memory test. Arch. Clin. Neuropsychol. 25, 591–600. doi: 10.1093/arclin/acq060
Jack, C. R. Jr., Albert, M. S., Knopman, D. S., McKhann, G. M., Sperling, R. A., Carrillo, M. C., et al. (2011). Introduction to the recommendations from the National Institute on Aging-Alzheimer’s Association workgroups on diagnostic guidelines for Alzheimer’s disease. Alzheimers Dement. 7, 257–262. doi: 10.1016/j.jalz.2011.03.004
Kordes, T. L. (2004). The use of the Hopkins verbal learning test for the evaluation of mild dementia in medical and rehabilitation settings: An analysis of construct and concurrent validity. Doctoral dissertation. Gannon University.
Lacritz, L. H., Cullum, C. M., Weiner, M. F., and Rosenberg, R. N. (2001). Comparison of the Hopkins verbal learning test-revised to the California verbal learning test in Alzheimer’s disease. Appl. Neuropsychol. 8, 180–184. doi: 10.1207/S15324826AN0803_8
Lewandowsky, S., Oberauer, K., Yang, L. X., and Ecker, U. K. (2010). A working memory test battery for MATLAB. Behav. Res. Methods 42, 571–585. doi: 10.3758/BRM.42.2.571
Liao, Z., Bu, Y., Li, M., Han, R., Zhang, N., Hao, J., et al. (2019). Remote ischemic conditioning improves cognition in patients with subcortical ischemic vascular dementia. BMC Neurol. 19:206. doi: 10.1186/s12883-019-1435-y
Lonie, J. A., Parra-Rodriguez, M. A., Tierney, K. M., Herrmann, L. L., Donaghey, C., O’Carroll, R. E., et al. (2010). Predicting outcome in mild cognitive impairment: 4-year follow-up study. Br. J. Psychiatry 197, 135–140. doi: 10.1192/bjp.bp.110.077958
Mukaka, M. M. (2012). Statistics corner: a guide to appropriate use of correlation coefficient in medical research. Malawi Med. J. 24, 69–71.
Nyberg, L. (1996). Classifying human long-term memory: evidence from converging dissociations. Eur. J. Cogn. Psychol. 8, 163–184. doi: 10.1080/095414496383130
O’Neil-Pirozzi, T. M., Goldstein, R., Strangman, G. E., and Glenn, M. B. (2012). Test–re-test reliability of the Hopkins verbal learning test-revised in individuals with traumatic brain injury. Brain Inj. 26, 1425–1430. doi: 10.3109/02699052.2012.694561
Onoda, K., Hamano, T., Nabika, Y., Aoyama, A., Takayoshi, H., Nakagawa, T., et al. (2013). Validation of a new mass screening tool for cognitive impairment: cognitive assessment for dementia, iPad version. Clin. Interv. Aging 8, 353–360. doi: 10.2147/CIA.S42342
Petersen, R. C. (2004). Mild cognitive impairment as a diagnostic entity. J. Intern. Med. 256, 183–194. doi: 10.1111/j.1365-2796.2004.01388.x
Prince, M., Acosta, D., Ferri, C. P., Guerra, M., Huang, Y., Jacob, K., et al. (2011). A brief dementia screener suitable for use by non-specialists in resource poor settings—the cross-cultural derivation and validation of the brief community screening instrument for dementia. Int. J. Geriatr. Psychiatry 26, 899–907. doi: 10.1002/gps.2622
Rabin, L. A., Paolillo, E., and Barr, W. B. (2016). Stability in test-usage practices of clinical neuropsychologists in the United States and Canada over a 10-year period: a follow-up survey of INS and NAN members. Arch. Clin. Neuropsychol. 31, 206–230. doi: 10.1093/arclin/acw007
Ryan, J., Woods, R. L., Murray, A. M., Shah, R. C., Britt, C. J., Reid, C. M., et al. (2021). Normative performance of older individuals on the Hopkins verbal learning test-revised (HVLT-R) according to ethno-racial group, gender, age and education level. Clin. Neuropsychol. 35, 1174–1190. doi: 10.1080/13854046.2020.1730444
Schneider, W., and Pressley, M. (2013). Memory development between two and twenty. New York, NY: Psychology Press.
Schrijnemaekers, A., de Jager, C. A., Hogervorst, E., and Budge, M. (2006). Cases with mild cognitive impairment and Alzheimer’s disease fail to benefit from repeated exposure to episodic memory tests as compared with controls. J. Clin. Exp. Neuropsychol. 28, 438–455. doi: 10.1080/13803390590935462
Scott, T. M., Gouse, H., Joska, J., Thomas, K. G. F., Henry, M., Dreyer, A., et al. (2020). Home-versus acquired-language test performance on the Hopkins verbal learning test-revised among multilingual south Africans. Appl. Neuropsychol. Adult 27, 173–180. doi: 10.1080/23279095.2018.1510403
Shapiro, A. M., Benedict, R. H., Schretlen, D., and Brandt, J. (1999). Construct and concurrent validity of the Hopkins verbal learning test–revised. Clin. Neuropsychol. 13, 348–358. doi: 10.1076/clin.13.3.348.1749
Shi, C., Kang, L., Yao, S., Ma, Y., Li, T., Liang, Y., et al. (2015). The MATRICS consensus cognitive battery (MCCB): co-norming and standardization in China. Schizophr. Res. 169, 109–115. doi: 10.1016/j.schres.2015.09.003
Shi, J., Tian, J., Wei, M., Miao, Y., and Wang, Y. (2012). The utility of the Hopkins verbal learning test (Chinese version) for screening dementia and mild cognitive impairment in a Chinese population. BMC Neurol. 12:136. doi: 10.1186/1471-2377-12-136
Singer, T., Verhaeghen, P., Ghisletta, P., Lindenberger, U., and Baltes, P. B. (2003). The fate of cognition in very old age: six-year longitudinal findings in the Berlin aging study (BASE). Psychol. Aging 18, 318–331. doi: 10.1037/0882-7974.18.2.318
Stricker, J. L., Brown, G. G., Wixted, J., Baldo, J. V., and Delis, D. C. (2002). New semantic and serial clustering indices for the California Verbal Learning Test–Second Edition: Background, rationale, and formulae. J. Int. Neuropsychol. Soc. 8, 425–435. doi: 10.1017/S1355617702813224
United Nations. (2022). World population prospects 2022: Ten key. Department of Economic and Social Affairs, Population Division. Available at: https://www.un.org/development/desa/pd/sites/www.un.org.development.desa.pd/files/undesa_pd_2022_wpp_key-messages.pdf
Vanderslice-Barr, J. L., Miele, A. S., Jardin, B., and McCaffrey, R. J. (2011). Comparison of computerized versus booklet versions of the TOMM™. Appl. Neuropsychol. 18, 34–36. doi: 10.1080/09084282.2010.523377
Wang, H., Fan, Z., Shi, C., Xiong, L., Zhang, H., Li, T., et al. (2019). Consensus statement on the neurocognitive outcomes for early detection of mild cognitive impairment and Alzheimer dementia from the Chinese Neuropsychological Normative (CN-NORM) Project. J. Glob. Health 9. doi: 10.7189/jogh.09.020320
Wang, J., Zou, Y.-Z., Cui, J.-F., Fan, H.-Z., Chen, R., Chen, N., et al. (2015). Revision of the Wechsler memory scale-of Chinese version (adult battery). Chin. Ment. Health J. 29, 53–59. doi: 10.3969/j.issn.1000-6729.2015.01.010
Whitener, E. M., and Klein, H. J. (1995). Equivalence of computerized and traditional research methods: the roles of scanning, social environment, and social desirability. Comput. Hum. Behav. 11, 65–75. doi: 10.1016/0747-5632(94)00023-B
Woods, S. P., Delis, D. C., Scott, J. C., Kramer, J. H., and Holdnack, J. A. (2006). The California verbal learning test–second edition: test-retest reliability, practice effects, and reliable change indices for the standard and alternate forms. Arch. Clin. Neuropsychol. 21, 413–420. doi: 10.1016/j.acn.2006.06.002
Woods, S. P., Scott, J. C., Conover, E., Marcotte, T. D., Heaton, R. K., Grant, I., et al. (2005). Test-retest reliability of component process variables within the Hopkins verbal learning test-revised. Assessment 12, 96–100. doi: 10.1177/1073191104270342
Keywords: reliability, validity, Alzheimer’s disease, amnestic mild cognitive impairment (aMCI), computerized applications
Citation: Jiang L, Xu M, Xia S, Zhu J, Zhou Q, Xu L, Shi C and Wu D (2023) Reliability and validity of the electronic version of the Hopkins verbal learning test-revised in middle-aged and elderly Chinese people. Front. Aging Neurosci. 15:1124731. doi: 10.3389/fnagi.2023.1124731
Edited by:
Pariya Wheeler, University of Alabama at Birmingham, United StatesReviewed by:
Paloma Roa, Instituto Nacional de Geriatría, MexicoLeticia Vivas, National Scientific and Technical Research Council (CONICET), Argentina
Copyright © 2023 Jiang, Xu, Xia, Zhu, Zhou, Xu, Shi and Wu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Chuan Shi, shichuan@bjmu.edu.cn; Daxing Wu, wudaxing2017@csu.edu.cn; wudaxing2012@126.com