- 1Rehabilitation Sciences Institute, University of Toronto, Toronto, ON, Canada
- 2Toronto Rehabilitation Institute-University Health Network, Toronto, ON, Canada
- 3Acquired Brain Injury Research Lab, University of Toronto, Toronto, ON, Canada
- 4Faculty of Life Sciences, McMaster University, Hamilton, ON, Canada
- 5Department of Biology, University of Toronto Mississauga, Mississauga, ON, Canada
Objectives: The purpose of evaluative instruments is to measure the magnitude of change in a construct of interest over time. The measurement properties of these instruments, as they relate to the instrument's ability to fulfill its purpose, determine the degree of certainty with which the results yielded can be viewed. This work systematically reviews all instruments that have been used to evaluate cognitive functioning in persons with traumatic brain injury (TBI), and critically assesses their evaluative measurement properties: construct validity, test-retest reliability, and responsiveness.
Data Sources: MEDLINE, Central, EMBASE, Scopus, PsycINFO were searched from inception to December 2016 to identify longitudinal studies focused on cognitive evaluation of persons with TBI, from which instruments used for measuring cognitive functioning were abstracted. MEDLINE, instrument manuals, and citations of articles identified in the primary search were then screened for studies on measurement properties of instruments utilized at least twice within the longitudinal studies.
Study Selection: All English-language, peer-reviewed studies of longitudinal design that measured cognition in adults with a TBI diagnosis over any period of time, identified in the primary search, were used to identify instruments. A secondary search was carried out to identify all studies that assessed the evaluative measurement properties of the instruments abstracted in the primary search.
Data Extraction: Data on psychometric properties, cognitive domains covered and clinical utility were extracted for all instruments.
Results: In total, 38 longitudinal studies from the primary search, utilizing 15 instruments, met inclusion and quality criteria. Following review of studies identified in the secondary search, it was determined that none of the instruments utilized had been assessed for all the relevant measurement properties in the TBI population. The most frequently assessed property was construct validity.
Conclusions: There is insufficient evidence for the validity and reliability of instruments measuring cognitive functioning, longitudinally, in persons with TBI. Several instruments with well-defined construct validity in TBI samples warrant further assessment for test-retest reliability and responsiveness.
Registration Number: www.crd.york.ac.uk/PROSPERO/, identifier CRD42017055309.
Introduction
Cognitive impairments are among the most important concerns for persons with traumatic brain injury (TBI). These impairments include a wide range of deficits in attention, memory, executive function, and behavioral and emotional difficulties, such as limited flexibility, impulsivity, reduced behavioral control, and inhibition, as well as other affective changes (1). Cognitive impairments directly impact their ability to maintain employment (2), personal and community independence (3), to participate in social activities (4), and their response to rehabilitation interventions (5). These are also among the main concerns of clinicians developing systems of care for TBI patients (6), and patients' family members and/or caregivers, who interact with and aid the injured persons on a daily basis (7). Cognition is a multi-dimensional construct, encompassing learning and memory, language, complex attention, executive functioning, perceptual-motor ability, and social cognition (8). Research to date has used numerous measures of cognitive functioning in persons with TBI longitudinally, to investigate its natural history (i.e., course over time) and the effectiveness of interventions aimed at improving cognition in clinical trials (9). The results were inconsistent, even when accounting for differences in time since injury and injury severity, with reports of improvement, decline, and no change over time (9). To elucidate the source of these inconsistencies, an important consideration involves investigation of the measures of cognitive functioning that have been utilized in the TBI population to date, to assess their suitability to perform the function for which they are intended. This route is one that has received relatively little attention in the discussion of generalization and interpretation of results of studies and this is a tremendous limitation, as selection of a measure affects the validity of the results reported (10). It has been argued that the usefulness of an outcome study or a clinical trial, in terms of the contribution made to the understanding of an issue and the potential to inform how the issue is viewed and treated in a clinical setting, hinges on the appropriateness of the measure used, and cannot be made up for even with otherwise superior design and execution (10). Measures used to study change in a construct over time are termed “evaluative,” and their most relevant psychometric properties, according to criteria developed by Feinstein (11) and Kirshner and Guyatt (12), are (i) construct validity, (ii) test-retest reliability, and (iii) responsiveness (11–13).
Construct validity refers to an instrument's ability to measure the construct it is intended to measure in the population of interest (e.g., cognitive functioning in the TBI population) (11). Developing a tool for measuring cognitive functioning that has construct validity is challenging because there is no generally accepted reference or gold standard instrument that is known to accurately define and measure the multidimensional construct, against which all new instruments could be compared (convergent validity). Divergent validity is another subcategory of construct validity, and it involves assessment of the relatedness of constructs thought to be unrelated and thus expected to yield scores on their respective measures that are not positively correlated (11). Finally, within construct validity there is also known-groups validity, which refers to the application of an instrument to two groups known or hypothesized to differ in the construct measured (11, 12). For an instrument to have construct validity, at least two of the construct validity subcategories must be assessed—convergent or divergent validity, and known-groups validity (11, 12).
Test-retest reliability concerns the extent to which application of the same instrument yields the same results in repeated trials under the same conditions (11, 13). This psychometric property is important for quantifying the degree of variance attributed to true differences in the construct under study over time, rather than systematic changes that occur when a procedure is learned (13).
Responsiveness refers to an instrument's ability to detect small, clinically significant differences in a construct of interest over time (12). This property is emphasized for instruments used in clinical trials, where the responsiveness of an instrument is directly related to the observed magnitude of the change in person's score, which may or may not constitute a clinically important difference (12). Responsiveness is inversely proportional to between-person variability in individual changes in score over time (12). In the TBI population, as the baseline variability increases, a larger treatment effect is needed to demonstrate intervention efficacy.
Finally, it is important to consider the specifics of the TBI population in the development and use of an instrument for cognitive functioning. Traumatic brain injury can impact not only cognition, but also behavioral and emotional functioning, and concentration, and this is expected to reflect in the ease of comprehension, extent of completion and overall burden on both the test taker and the administrator (in explaining the procedure and assisting with comprehension and completion) associated with administration of an instrument.
To identify the most appropriate instrument(s) for measuring cognitive functioning in the TBI population, we undertook a systematic review of all instruments used for this purpose. The objectives were to: (i) describe each evaluative instrument's key measurement properties (i.e., construct validity, test-retest reliability, and responsiveness); (ii) classify instruments according to the cognitive domains they assess; and (iii) summarize information relevant to their clinical and research applications. The present work intends to inform researchers and clinicians on each instrument's utility as an evaluative measure of cognitive functioning in the TBI population, while identifying pitfalls and future directions for their utility.
Methods
This systematic review is part of a larger study that focuses on central nervous system (CNS) trauma [TBI and spinal cord injury (SCI)] as a risk factor of cognitive decline over time. For more information, the reader is referred to the published protocol (14) and registry with the International Prospective Register of Systematic Reviews (PROSPERO) (registration number CRD42017055309) (15).
Primary Search: Studies of Cognitive Functioning in TBI
A comprehensive search strategy was developed in collaboration with a medical information specialist (JB) at a large rehabilitation teaching hospital. All English language peer-reviewed studies published from onset to December 2016 with prospective or retrospective data collection and a longitudinal design, identified in six electronic databases (i.e., MEDLINE, Central, EMBASE, Scopus, PsycINFO, and supplemental PubMed), were considered eligible. The following medical subject headings in MEDLINE were used to identify publications of interest (i) TBI terms: exp “brain injuries” or “craniocerebral trauma” or exp “head Injuries, closed” or exp “skull fractures” or “mTBI*2.tw.” or “tbi*2.tw” or “concuss*.tw.” AND (ii) cognition terms: exp “cognition” or exp “cognition disorders” or “neurocognit*.tw,kw.” Or “executive function” or exp “arousal” or “attention*.tw,kw.” or “vigilan*.tw,kw.” or exp “dementia” AND (iii) evaluation terms: exp “cohort studies” or “longitudinal studies” or “follow-up studies” or “prospective studies” or “retrospective studies” or “controlled before-after studies/or interrupted time series analysis” or exp “clinical trials” or exp “clinical trials as topic.” The search terms were adapted for use in other bibliographic databases. The reader is referred to the published protocol (14) and the PROSPERO registry (15) for the full search strategy. Additional studies were identified through review of reference lists of included articles.
Inclusion and Exclusion Criteria
Studies were included if they met the following criteria: (i) focused on longitudinal change in cognitive functioning in adults (i.e., ≥16 years) with an established clinical diagnosis of TBI based on accepted definitions [e.g., Glasgow Coma Scale (GCS) score, duration of loss of consciousness, and post traumatic amnesia, etc.], excluding self-report; (ii) reported cognitive functioning outcome data at baseline assessment and follow-up as a score on a standardized measurement instrument; and (iii) the work was published in English in a peer-reviewed journal. Studies were excluded if they: (i) evaluated cognitive functioning in children/adolescents; (ii) studied persons with minor head injury (cases before 1993) without providing assessment criteria; or (iii) reported results in letters to the editor, reviews without data, case/public reports, conference abstracts, articles with no primary data, or theses.
Selection and Quality Assessment of Studies
In the first stage of screening, two reviewers (NP and AD, or SM and AD) assessed study titles and abstracts for potential agreement with the inclusion criteria. In the second stage, each reviewer individually assessed the full texts of studies selected in the first stage to determine whether they met the inclusion criteria. Discrepancies in article inclusion/exclusion were resolved by discussion with TM.
Previously developed standardized forms were used to assess study quality (16) and to synthesize results (17). Study quality was assessed using the Quality in Prognosis Studies (QUIPS) guidelines (18). Assessments were based on the presence of six potential sources of bias (i.e., participation, attrition, prognostic factors, outcome measures, consideration of and accounting for confounders, and data analyses). Each study was assigned an overall “risk of bias,” and those with the greatest risk were excluded. Studies of a retrospective nature were automatically excluded from a “low risk” rating, as recommended by the Scottish Intercollegiate Guidelines Network (SIGN) (19). Any discrepancies between the two reviewers in quality assessment were resolved in discussion among the research team followed by independent review by the research supervisor (TM).
Secondary Search: Studies of Measurement Properties of Abstracted Instruments
Instruments used to evaluate cognitive functioning in studies that met inclusion and quality criteria were abstracted. In collaboration with a medical information specialist (JB), proposed MEDLINE search filters were used to identify studies reviewing the abstracted instruments' measurement properties in TBI samples. Supplementary File 1 provides the terms and outputs from searches for each measure. The reference lists of eligible articles, instrument manuals and Google Scholar were reviewed for other relevant publications. Studies where the primary objective was not the evaluation of measurement properties were excluded.
Evidence-Based Assessment of Instruments Evaluating Cognitive Functioning
Criteria for evidence-based assessment proposed by Holmbeck et al. (20) were utilized, previously applied in a systematic review of measurement properties of sleep-related instruments in the TBI population (17). Instruments used in at least two of the studies identified in the primary search were given ratings of “well-established,” “approaching well-established,” or “promising,” based on the following criteria: (i) use in peer-reviewed studies by different research teams; (ii) availability of sufficient information for critical appraisal and replication; and (iii) demonstration of validity and reliability in the TBI population (20).
Descriptive Aspects of Instruments of Cognitive Functioning
To assess research and clinical feasibility, in-depth descriptions were completed following a previously developed format for instruments in medical research (17). The following descriptors were abstracted from data sources and reported: (i) general: purpose, content, response options, recall period (ii) application: how to obtain, method of administration, scoring and interpretation, administrator and respondent burden, currently available translations; and (iii) critical appraisal as reported by the researchers who utilized the instrument in TBI and other samples: strengths, considerations, clinical, and research applicability (17).
Categorization of Instruments of Cognitive Functioning by Content
Content validity refers to the degree to which an instrument's items adequately reflect the construct of interest. For a measure to be sensitive to certain or all aspect(s) of cognitive functioning in a person with TBI, it needs to feature representative items or tasks that are part of the construct of cognition, as understood by the instrument's developer. As such, each instrument was categorized according to the cognitive domain(s) it assesses, focusing on those listed in the Diagnostic and Statistical Manual of Mental Disorders-Fifth Edition (DSM-5) (8): (i) complex attention, (ii) executive functioning, (iii) learning and memory, (iv) language, (v) perceptual-motor ability, and (vi) social cognition (8). An additional domain, information processing and reaction time, was included, given its relevance to the TBI population (21, 22). Instruments were then classified, based on the number of domains they assess, as either “global” (all domains), “multi-domain” (two or more domains), or “domain-specific” (one domain).
Results
Literature Search and Quality Assessment
Of 29,566 studies identified in the primary search of articles assessing cognitive functioning longitudinally in the TBI population, 39 met inclusion and quality criteria (Figure 1): 14 studies involved patients from acute care (23–36), seven from the rehabilitation setting (37–43), ten involved college-age athletes (44–53), four were clinical trials (54–57), two involved samples from community care settings (58, 59), and another two involved military and veteran participants (60, 61). Sample sizes ranged from 10 (35, 55) to 509 (57), and consisted of mostly males (mean 76.6%, range 38-100%) with participant age ranging from 18 (57) to >60 years (57) (Table 1).
Table 1. Summary of study characteristics, including details on study sample, purpose, and instruments used to evaluate cognition.
All 39 studies were assessed as having “Partly” or “No” on all bias criteria. Twenty-nine studies were of fair quality (23, 26, 29–32, 34, 37–45, 48–54, 54–61), ten were of good quality (24, 25, 27, 28, 33, 35, 36, 46, 47, 53) and none were of high quality. Studies were most frequently penalized by the SIGN criteria for unknown reliability and validity of the utilized instruments, incomplete statistical analysis, and selection bias due to study attrition (Supplementary File 2).
Instruments Measuring Cognitive Functioning
Within the 39 studies, 15 instruments were used more than once. The Mini Mental State Examination (MMSE) (54, 55), Hopkins Verbal Learning Test (HVLT) (47, 49), Paced Auditory Serial Addition Test (PASAT) (27, 48), and Rey-Osterrieth Complex Figure Test (ROCF) (36, 58) were each used twice; the California Verbal Learning Test (CVLT) (57, 58, 61) and Wechsler Memory Scale (WMS) (37, 38, 42) were used three times; the Automated Neuropsychological Assessment Metrics (ANAM) (33, 50, 51, 53) and FIM-Cog (Functional Independence Measure-Cognitive Subscale) (34, 40, 43, 56, 58) were used four and five times, respectively; the Rey Auditory Verbal Learning Test (RAVLT) (28, 37, 39, 41, 42, 59), Stroop Color Word Test (SCWT) (27, 28, 37, 49, 57, 58), and Symbol Digit Modalities Test (SDMT) (26, 39, 41, 42, 48, 49) were each used six times; the Immediate Post-concussion Assessment and Cognitive Test (ImPACT) (31, 35, 44–46, 52, 60) and Controlled Oral Word Association Test (COWAT) (24, 25, 39, 41, 42, 49, 57, 58) were used seven and eight times, respectively. The most frequently used instruments were the Trail Making Test (TMT) (23, 24, 28, 32, 37–39, 41, 42, 48, 49, 57, 59) and the Wechsler Adult Intelligence Scale (WAIS) (23–26, 28–30, 32, 36–39, 41, 42, 55, 57, 58), used 13 and 17 times, respectively.
Assessment of TBI
Diagnostic criteria and definitions of TBI varied considerably between studies included in this review (Table 1). Nineteen studies used a combinatorial approach to confirm and assess TBI. This included the use of such tools as the Glasgow coma scale (GCS), duration of posttraumatic amnesia (PTA) and/or loss of consciousness, neuroimaging results [i.e., magnetic resonance imaging (MRI), computed tomography (CT)], and clinical evaluations and tests (30, 31, 33, 34, 36, 40, 42–44, 46–49, 53, 61–65). Five studies used the American Academy of Neurology (AAN) graded concussion assessment test (44, 53, 55, 59, 66), six used GCS scores (38, 39, 50, 56, 58, 67), one study assessed CT scans (60), one—MRI scans (57), and one assessed PTA (37) alone to confirm TBI. Two studies used other methods, including description of damage and/or lesions based on medical records, and diagnoses of referring professionals (29, 68).
Injury Severity in Samples Assessed
Three measures were used to assess cognitive functioning in mild TBI (mTBI) samples only (ANAM, HVLT, ImPACT), while the rest were applied to samples of varying injury severities. Among the most commonly used measures were the TMT, used in 11 studies, of which six (23, 28, 32, 42, 48, 49) comprised mTBI samples, three—mixed injury severity samples (24, 39, 41), and two—severe TBI samples (58, 59). The COWAT was used in eight studies, of which two (42, 49) featured mTBI samples, four (24, 39, 41, 57)—mixed injury severity samples, and two (25, 58)—severe TBI samples. Several versions of the WAIS were used seven times: once (23) in a study of mTBI participants, twice (36, 41) in samples of mixed injury severities, and four times (25, 30, 38, 58) in severe TBI samples (Table 3).
Evaluation of Measurement Properties
Construct Validity
Convergent, divergent, and/or known-groups validity (62) were reported for all instruments in TBI samples of all severities and mixed-severity samples (39, 53, 63–81). Where construct validity was evaluated, at baseline or follow-up assessments, correlation strength between scores of instruments measuring the same construct, or scores of groups of people with known differences in cognitive functioning, were not always in line with clinical expectations. There is evidence of moderate to strong convergent validity of the original version of CVLT in mixed severity TBI samples (64), FIM-Cog in severe and mixed TBI samples (67); ImPACT in mTBI samples (65, 66); MMSE in mixed severity TBI (80); PASAT in mild and mixed severity samples (69); ROCF in severe and mixed samples (70, 71); SCWT, and WAIS and WMS in all TBI severity samples (63, 72–75, 77, 78).
Divergent validity hypotheses were tested by analyzing the correlation of the PASAT with measures of intellectual, mathematical, and verbal abilities, academic achievement and complex motor skills (i.e., r = 0.29–0.59, p < 0.05), all of which were significant positive correlations (69). Correlations of the ImPACT with difficulty concentrating and remembering were negative (i.e., r = −0.48– (−0.41), p < 0.01) (Table 2) (65).
Table 2. Quality assessment of the 15 selected instruments based on criteria proposed by Holmbeck et al. (20).
Known-groups validity was reported for several domains of the ANAM and HVLT in mTBI samples (53, 68, 79); COWAT in mixed and severe TBI samples (72); list learning/delayed recall from the RAVLT in all injury severity samples (39); oral and written SDMT in mild and mixed severity samples (72); MMSE in a TBI sample of unknown severity (81); TMT, WAIS-R and WAIS-III in mixed severity samples (39, 72, 76, 77), and WAIS-IV in mTBI sample (82). Significant differences were observed between the scores of persons with TBI and healthy controls, and TBI and other neurological populations (Table 2).
Test-Retest Reliability
Test-retest reliability refers to the consistency of scores attained by the same patient over the course of several attempts at different times (62). It concerns the stability of the instrument's performance over a period of time, when a real change in the measured construct (i.e., cognition) is unlikely (62). The assumption is that while between-person differences in scores on a given measure are expected, the score for any individual will remain constant across successive administrations of the instrument. In our population of interest, persons with TBI, there is evidence for the test-retest reliability of the HVLT-R and the ANAM4. One study (83) reported the Pearson correlation coefficient (PCC) and the other (68) reported the intra-class correlation coefficient (ICC), the preferred statistic. The HVLT-R was administered twice to 75 adults with TBI of unknown injury severity (71% men, 46.5 ± 10.5 years of age, at 11.8 ± 9.6 years post injury), the two sessions occurring 6–8 weeks apart from one another (83). Correlation coefficients for two of the eight HVLT-R scores (total recall and delayed recall) reflected high test-retest reliability (r = 0.82) and the remaining six scores (T-score, delayed recall T-score, retention, retention T-score, recognition discrimination index and recognition discrimination index T-score) reflected moderate (r = 0.64) test-retest reliability. The ANAM4 was administered twice to 1,324 members of the Marine Corps unit (all men, 22.5 ± 3.4 years of age) with a known high rate of concussion from combat and blast exposure. The average interval between the two test sessions was 357 ± 88 days (range 99–637 days) (68). After injury classification, 238 members were designated to the concussed group and 264 to the non-concussed group. While there were no significant differences between the mean scores of the two groups at the first session, differences emerged at second session, with the concussed group having lower mean scores than the non-concussed group on the cognitive tasks assessing attention, memory, spatial processing, reaction time, and cognitive fatigue [i.e., code substitution delayed (CDD), matching to sample (M2S), procedural reaction time (PRT), and simple reaction time (repeat) (SRT, SRT2) subscales]. The test of simple effects revealed that the mean score for the concussed group decreased significantly from T1 to T2 on the SRT, SRT2, PRT, Code Substitution Learning, M2S, Mathematical Processing (MTH), and CDD subscales. The ICC between the scores from the first and second sessions was reported only for the non-concussed group: of the seven domains, the CDD and MTH domains met the cut-off for the mean score correlation between the two time points for the entire group (i.e., >0.70) but not in the comparison of scores of individual patients at the two time points (i.e., >0.90). Practice effects were reported for the ANAM, where it was noted that individuals with TBI displayed inconsistent performance in 30 administrations over four days, while controls showed consistent improvement (85) (Table 3).
Table 3. Summary of domains assessed in instruments measuring cognitive functioning, and interpretation of scores.
Responsiveness
Responsiveness, defined as the ability of an instrument to detect change over time in the construct being measured (62), was reported for the ANAM4 in a sample of concussed males (68) and for the WAIS in a sample of severe TBI (84). The former study of young men from the Marine Corps unit (68) tested the rate of performance decline on ANAM4 subscales from the first test session to the second, applying the reliable change (RC) methodology (86). Researchers reported that 48% of the concussed group demonstrated a decrease in performance on two or more subscales, compared to 28% of the non-concussed group.
When the WAIS was administered to 40 adults who sustained TBI and experienced posttraumatic amnesia (PTA) lasting at least 4 days (95% men, 28.3 ± 13.36 years of age) in the latter stages of their PTA and to a matched group of 40 non-injured persons, TBI group scores on the verbal subscales indicated less initial impairment and were restored to levels exhibited by the comparison group at a faster rate than were the scores on non-verbal subscales (84). The mean verbal intelligence quotient (IQ) of the TBI group approached that of the comparison group within the first year after injury, while performance IQ continued to improve over the course of 3 years (84, 87).
Classification of Instruments of Cognitive Functioning by Cognitive Domain(s) Assessed
The WAIS assesses all seven cognitive domains, qualifying as a “global” measure of cognitive functioning. The remaining instruments were “multi-domain” measures. The most represented domain was learning and memory, assessed in 13 instruments, and the least represented was social cognition, included in two instruments (Table 3; Supplementary File 3).
Information Relevant to Clinical and Research Applications
Instruments' manuals, assessment forms, and scoring instructions are available from their publishers; some are available online for free (Table 2). Information about the instruments (e.g., purpose, content, measurement properties, etc.) can be obtained online (Supplementary File 3).
All instruments require participants to complete one or more tasks, typically through written or spoken responses. The ANAM and ImPACT are computerized, the WAIS and WMS can be computer- or paper-based, and the FIM-Cog is administered via interview or participant observation. The number of items in each instrument varies: for instance, the COWAT presents three letters and relies on free word recall, while the PASAT contains a 61-item list of digits that the test-taker must sum up (i.e., adding each digit presented to the one that came just before it). The FIM-Cog, and larger batteries like the ANAM, ImPACT, WAIS and WMS, contain multiple tasks assessing different cognitive domains, ranging from five (e.g., FIM-Cog) up to 15 (e.g., WAIS), of which ten are core and five are supplementary subtests. Completion times range from 3 min (e.g., COWAT) to more than 90 min (e.g., ANAM, WAIS). Scoring procedures and score interpretation varies across instruments. Scoring sheets/software and instructions are available with purchase of the instruments.
Evidence-Based Assessment of Cognitive Functioning Measures
All 15 instruments were utilized in at least two peer-reviewed studies by two different research teams. The WAIS and TMT were the most frequently used instruments: different versions of the WAIS were used 17 times by 14 different teams, and the TMT was used 14 times by 11 teams (Table 2). None of the instruments met the criteria for a “well-established,” “approaching well-established,” or “promising” rating in the TBI population (Table 3). Known-groups validity and responsiveness were reported for two versions of the ANAM in two concussion samples (53, 68). The hypotheses regarding known-groups validity were accepted for the HVLT-R delayed recall task but rejected for the total score in a mTBI sample (79), and test-retest reliability in a TBI sample of unknown severity met the correlation cut-off for the group, but not for individual patients for total recall and delayed recall only (83). Convergent and divergent construct validity hypothesis testing for other instruments were not supported for all subscales/tasks, and were not always in line with clinical expectations (69, 78).
Table 4 provides a summary of the measurement properties of measures of cognitive functioning in TBI samples.
Table 4. Summary of measurement properties of evaluative instruments of cognitive functioning in TBI samples.
Discussion
This review provides a comprehensive overview of existing instruments used to evaluate cognitive functioning in clinical and non-clinical settings in persons with TBI. An extensive search strategy led to identification of 15 instruments; each was reviewed and comprehensively described (Supplementary File 3), providing information on content and level of evidence existing regarding measurement properties as they concern the use of these tools for evaluative purposes, i.e., their ability to measure change in a construct over time. Our results highlight that most scientific evidence pertains to construct validity, with limited evidence on test-retest reliability and responsiveness in TBI samples. This poses a risk to TBI researchers and clinicians when it comes to interpreting the results produced in longitudinal studies, with no certainty of the instruments' ability to measure change in cognitive functioning in persons with TBI longitudinally. The results are informative nevertheless, having implications for the understanding of and future research related to the measurement properties of evaluative instruments, and their subsequent utilization for studying the natural history and clinical course of cognitive functioning and treatment effects in clinical trials in persons with TBI.
Construct Validity
Construct validity refers to the development of a mini theory to describe how well an instrument measuring a construct of interest would agree with another instrument measuring a related construct (11, 62). In the measurement of cognitive functioning in persons with TBI, there does not exist a gold standard or criterion measure (11), and therefore understanding the domains each instrument is trying to measure (content validity) and whether the different instruments relate to one another in the way one would expect, is an important property to consider. It is important to highlight the basic principle of construct validation, which is that hypotheses about the relationship of scores of any instrument with scores on other instruments measuring a similar or different construct (convergent and divergent construct validity, respectively) should be formulated in advance; the specific expectations with regards to certain relationships can be based either on an underlying conceptual model or on the data in the literature (11). This review has found that only a few researchers tested hypotheses related to the relationship between the measures of cognitive functioning studies and measures of other constructs, stating ahead of analysis the expected direction and magnitude of associations, based on what was known about the constructs under study. In future studies, to assess similarity or dissimilarity between instruments' scores, when formulating hypotheses, one should first have a solid grasp of the contents of comparable instruments, which we provide in this review, as domains within any given instrument are expected to be correlated strongly with domains of conceptually similar instruments. There should also be a clear description of what is known about the TBI population under study, including but not limited to the circumstances surrounding injury, injury severity and mechanism, brain maturity and brain health at the time of injury, comorbid mental and physical disorders, coping ability, and psychotropic medication use, as each of these have the ability to influence cognitive functioning at assessment (88–92).
Known-groups validity (of construct validity) refers to an instrument's ability to discriminate between groups of individuals known to have a particular trait and those who do not have that trait (62). This property is most relevant for discriminative (i.e., diagnostic) instruments (11, 12), however is significant for evaluative instruments, where it is imperative that an instrument is responsive to all clinically important differences between constructs under investigation or different courses or outcomes of the construct (13). This includes identification and deletion of unresponsive items within a construct from the instrument over time. One way to identify such items within the cognitive functioning constructs assessed by instruments in the TBI population is to administer the instrument to a group of people with TBI of varying severities and associated cognitive impairments and to healthy people without impairments, and compare the scores yielded at baseline testing (known groups validity) and at follow-up after an intervention with known efficacy in improving cognition (e.g., cognitive training). Presence and absence of differences between the two groups in items will indicate items that are responsive and those that are not, respectively (longitudinal known-group construct validity). Only one study assessing measurement properties provided parameters of longitudinal known-group validity for the ANAM4 in young men with concussion (68) and those without.
Test-Retest Reliability
Not every change on a measurement instrument can be considered a real or true change in the construct the instrument is believed to measure (13). Observed changes in scores over time may be due to measurement error, natural variability in a person's ability to concentrate throughout the day, i.e., peak in performance, mood at the time of investigation, which may determine positive or negative responses in case of doubt, evaluators' variability in applying criteria more or less strictly, or the natural course of the construct under study (i.e., recovery or deterioration) (11–13). Therefore, interpretation of change in score over time requires assessment of measurement error by test-retest of a stable population of interest (13). But what is a stable TBI population? By choosing a timeframe with 6–8 weeks between the two test sessions of the HVLT-R, researchers reported moderate to high test-retest reliability on all eight parameters tested (83). Unfortunately, the researchers did not ask their participants how their cognitive activity changed over the 6–8 weeks, and therefore, the stability of the group of TBI patients that took part in this study is unknown and thus the interpretability of the score and corresponding attribution of functional status is uncertain. When the researchers assessed test-retest reliability of the ANAM4 in young men in the chronic stage post-concussion and in their non-injured counterparts, the groups' scores were comparable at baseline, but differences in some but not all subscales (CDD, M2S, PRT, SRT, and SRT2) emerged at 357 ± 88 days after baseline assessment, with the concussed group exhibiting lower scores than the non-concussed group; the reason for this observation is not clear (68). Consequently, issues related to the meaning and interpretation of results of longitudinal studies utilizing instruments with unknown test-retest reliability in the population of interest remain pressing.
Responsiveness
Responsiveness, defined as an instrument's ability to accurately detect change when it has occurred, is key for instruments applied for the purpose of evaluation (13). Responsiveness is not an attribute of the instrument itself, but reflects the application of the instrument in a given context (e.g., for quantifying the benefit of an intervention in a clinical trial), or to a certain type of change (i.e., natural history, recovery, etc.). Responsiveness is the ability of an instrument to detect change when it has taken place, and can be expressed either as absolute difference within a person or group, or the effect size, referred to as the standardized response mean (91). Responsiveness is the least studied attribute of measures of cognitive functioning. The results of one study that investigated responsiveness of the WAIS in a sample with severe head injury supported the hypothesis of differential speed of recovery of different domains of cognition in TBI (i.e., verbal IQ of the head injured group approached that of the comparison group within about one year of injury, while recovery of performance IQ continued to improve over about 3 years) (84). The results of another study assessing the ANAM4 in a concussed sample (68) reported that 48% of the concussed group demonstrated decreases in performance on two or more subscales compared with 28% of the non-concussed group, driven by the CDS, CDD, PRO, SRT, and ST2 subscales (assessing abilities related to information processing speed, reaction time, attention, memory, and learning). Neither study however, formulated specific hypotheses with respect to expected mean differences in scores in the studied groups a priori. Without the a priori hypotheses the risk of bias is high, because retrospectively, it is tempting to think of explanations for the observed results instead of concluding that an instrument might not be responsive. Also worth noting is that when measurement properties were assessed, participant samples consisted of mostly or strictly men. Without gender-related specifications on the applications of the ANAM and the HVLT, it impossible to draw any firm conclusions on the instruments' test-retest reliability and responsiveness, and therefore their ability to detect change over time.
Feasibility, and Clinical and Research Utility
Feasibility concerns practicality of administering an instrument to a person in the setting it which it needs to be administered (11). In order to accurately measure what it intends to measure, and ensure valid responses, self-administered instruments must be completely self-explanatory, and those administered by research or clinical personnel require personnel to be trained to collect the required information. None of the studies included in this review reported on the training of respondents and administrators, or lack thereof. This is important for ensuring the procedure is standardized and to confirm the ability of persons with TBI to complete the procedure and respond with insight. Instruments that require more than 30 min to complete can be tiring for persons with TBI, who commonly experience decreased stamina, especially during tasks involving novelty. Multi-domain instruments demand continuous, goal-directed activity, which would be affected by diminished motivation, impairments, psychological state, and age. In such cases, researcher or clinicians may administer combinations of relevant sub-tests over some time, which can present a challenge for calculation of an accurate score and interpretation of scores. A composite score is the most practical approach for researchers looking to quantify a population's global change over time: if sample sizes are sufficiently large, the between-person variation in certain subscales or domains will be balanced in the calculation of a mean global score for the group. In the clinical setting, however, where clinicians are working one-on-one with individuals to study the natural history of or intervention-induced changes in a construct over time, focus on the individual subscales is key. The implication is that the scores of each of the subscales of a global or multi-domain measure have to be validated separately in a population of interest. This is particularly relevant for certain domains of cognition such as crystalized intellectual abilities, which have been hypothesized to be resistant to the effects of TBI. Finally, study of the evaluative properties of measures of cognitive functioning that are reflective of everyday cognitive skills is needed.
Study Limitations
There are a number of limitations to this review. All of the studies included were published in English, and therefore instruments used in non-English-speaking populations and studies have not been captured. The review team did not contact authors of reviewed papers for additional methodological details that were not available in their publications. Further, the team appraised methodological quality utilizing Holmbeck et al.'s criteria (20), developed for measures of psychosocial adjustment and psychopathology, and therefore not specific for measures of cognitive functioning. Nevertheless, justifications for quality ratings given to each of the reviewed instruments were reported to clarify the resultant assessment grades. Several instruments [i.e., standardized assessment of concussion (SAS), short orientation-memory-concentration test (SOMT), and the neuropsychological assessment battery (NAB)] assessing orientation (i.e., time and place), were not reviewed in this work, as they were utilized just once within the articles identified in the primary search, and therefore did not meet the frequency of use criteria set for this review.
Another potential issue is that while the instruments themselves are standardized, there are a number of scores that could be derived from each one. For instance, CVLT scores can be based on the number of words recalled from multiple lists, recall after short/long delay, number of errors in recall, etc. The use of these scores was not consistent from study to study. This lack of consistency not only makes it challenging, if impossible, to compare scores across studies, but may impact evaluation of certain measurement properties, such as concurrent validity, if only certain scores are associated with instruments meant to measure similar/different constructs (Table 3; Supplementary File 3). There are limitations related to the identification of data on construct validity. For the purpose of this review, data on measurement properties was gathered from studies of longitudinal design only. Despite attempts to include all articles relevant to construct validity in the TBI population, it is possible some cross-sectional studies evaluating construct validity were missed.
Finally, while the potential application of the described instruments measuring cognitive functioning in TBI can be for diagnostic/descriptive and prognostic/predictive purposes, the focus of our work was to examine the evaluative properties of such instruments (i.e., their ability to measure the magnitude of change longitudinally), when no external criterion is available for validating the construct. Thus, the assessment of properties of instruments included in this review, as descriptive or predictive measures of cognitive functioning, requires further study.
Conclusions
In research utilizing evaluative, or other, psychometrics, the suitability of the instruments, or lack thereof, in terms of their psychometric properties, is rarely discussed, or acknowledged as a limitation in the case of measures whose scores have not been validated. This review highlights the problematic use of certain measures that lack the properties necessary for their use as evaluative measures. The evidence on measurement properties of instruments used to assess cognitive functioning in TBI samples longitudinally is limited, and thus, the way forward appears to be consensus on a set of measures with the greatest potential for evaluative purposes in TBI, and assessment of these select measures to build the evidence on measurement properties and establish or refute rationale for their application in TBI research. Refinement and testing of this group of instruments in TBI samples of varying severities in terms of longitudinal construct validity, test-retest reliability, and responsiveness, not only to reliably study the course of cognitive functioning after brain injury, but also for quantifying treatment benefits in clinical trials, is timely. Assessment of psychometric properties should not be an afterthought, but rather should preface the application of a measure in any new population or context, serving as the deciding factor on whether to proceed with its use. It is important that future research on psychometric properties of evaluative psychometrics, take into account the heterogeneity in cognitive functioning of persons with TBI, and report stratified results for subgroups of people based on injury severity, mechanism, and baseline cognitive abilities, to mitigate some of the heterogeneity. Further, to present these results in the context of change in everyday functional capabilities as time since injury progresses or in response to intervention, which will provide valuable insights. Furthermore, emphasis on within-individual variability in the TBI population, where each person serves as their own control, is likely to be the best technique to analyse change, and to answer the question—to which extent is the inherent regenerative capacity affected by injury-related variables, as opposed to internally (age, sex, genetic profile, etc.) or environmentally driven variables, as well as brain-behavior relations. The trade-off with the latter, however, is limited standardization of individual outcome measures, and lack of current theories of psychometric properties that speak to single case experimental design studies only (relevant to personalized medicine theory) and limited external validity (i.e., generalizability).
Ethics Statement
This article does not contain any studies with human or animal subjects performed by any of the authors.
Author Contributions
TM and AC contributed to the conception of the study. TM developed the idea, registered the review on PROSPERO, designed and published the protocol, and developed the study screening criteria and quality assessment criteria. TM, SM, NP, FJ and AD executed the study in accordance with the protocol protocol. AD and NP screened all abstracts. AD extracted the data. SM doubled checked all extracted data. AD, SM, and NP performed study quality assessment and abstracted the data. TM guided the process and checked all data. AD and SM performed data analyses. AD wrote the first draft of the review, which was then edited by SM and TM. TM and AC provided mentorship to AD, SM, NP, and FJ throughout the course of the study.
Funding
This research program is supported by the postdoctoral research grant to TM from the Alzheimer's Association (AARF-16-442937). AC is supported by the Canadian Institutes for Health Research Grant–Institute for Gender and Health (#CGW-126580). The funders had no role in study design, data collection, decision to publish, or preparation of the manuscript.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The handling editor declares a shared association, though no other collaboration, with one of the authors AC in the Canadian concussion consortium.
Acknowledgments
We gratefully acknowledge the involvement of Ms. Jessica Babineau, information specialist at the Toronto Rehabilitation Institute, for her help with the literature search.
Supplementary Material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fneur.2019.00353/full#supplementary-material
Abbreviations
ANAM, Automated Neuropsychological Assessment Metrics; CNS, Central Nervous System; COWAT, Controlled Oral Word Association Test; CVLT, California Verbal Learning Test; DSM-5, Diagnostic and Statistical Manual of Mental Disorders-Fifth Edition; FIM-Cog, Functional Independence Measure-Cognitive Subscale; GCS, Glasgow Coma Scale; HVLT, Hopkins Verbal Learning Test; ImPACT, Immediate Post-concussion Assessment and Cognitive Test; MMSE, Mini-mental State Examination; PASAT, Paced Auditory Serial Addition Test; PROSPERO, International Prospective Register of Systematic Reviews; QUIPS, Quality in Prognosis Studies; RAVLT, Rey Auditory Verbal Learning Test; ROCF, Rey-Osterrieth Complex Figure Test; SDMT, Symbol Digit Modalities Test; SIGN, Scottish Intercollegiate Guidelines Network; TBI, Traumatic Brain Injury; TMT, Trail Making Test; WAIS, Wechsler Adult Intelligence Scale; WMS, Wechsler Memory Scale.
References
1. Jenkins PO, Mehta MA, Sharp DJ. Catecholamines and cognition after traumatic brain injury. Brain. (2016) 139(Pt. 9):2345–71. doi: 10.1093/brain/aww128
2. Mani K, Cater B, Hudlikar A. Cognition and return to work after mild/moderate traumatic brain injury: a systematic review. Work. (2017) 58:51–62. doi: 10.3233/WOR-172597
3. Takada K, Sashika H, Wakabayashi H, Hirayasu Y. Social participation and quality-of-life of patients with traumatic brain injury living in the community: a mixed methods study. Brain Inj. (2016) 30:1590–98. doi: 10.1080/02699052.2016.1199901
4. Goverover Y, Genova H, Smith A, Chiaravalloti N, Lengenfelder J. Changes in activity participation following traumatic brain injury. Neuropsychol Rehabil. (2017) 27:472–85. doi: 10.1080/09602011.2016.1168746
5. McLafferty FS, Barmparas G, Ortega A, Roberts P, Ko A, Harada M, et al. Predictors of improved functional outcome following inpatient rehabilitation for patients with traumatic brain injury. Neurorehabilitation. (2016) 39:423–30. doi: 10.3233/NRE-161373
6. Cicerone VG, Dahlberg C, Kalmar K, Langenbahn DM, Malec JF, Bergquist TF, et al. Evidence-based cognitive rehabilitation: recommendations for clinical practice. Arch Phys Med Rehabil. (2000) 20:215–28. doi: 10.1053/apmr.2000.19240
7. Bivona U, Formisano R, De Laurentiis S, Accetta N, Di Cosimo MR, Massicci R, et al. Theory of mind impairment after severe traumatic brain injury and its relationship with caregivers' quality of life. Restor Neurol Neurosci. (2015) 33:335–45. doi: 10.3233/RNN-140484
8. Eramudugolla R, Mortby ME, Sachdev P, Meslin C, Kumar R, Anstey KJ. Evaluation of a research diagnostic algorithm for DSM-5 neurocognitive disorders in a population-based cohort of older adults. Alzheimers Res Ther. (2017) 9:15. doi: 10.1186/s13195-017-0246-x
9. Mollayeva T, Mollayeva S, Pacheco N, D'Souza A, Colantonio A. The course and prognostic factors of cognitive outcomes after traumatic brain injury: a systematic review and meta-analysis. Neurosci Biobehav Rev. (2018) 99:198–250. doi: 10.1016/j.neubiorev.2019.01.011
10. Coster WJ. Making the best match: selecting outcome measures for clinical trials and outcome studies. Am J Occup Ther. (2013) 67:162–70. doi: 10.5014/ajot.2013.006015
11. Feinstein AR. The theory and evaluation of sensibility: In: AR Feinstein editor. Clinimetrics. New Haven: Yale University Press (1987). p. 141–66.
12. Kirshner B, Guyatt G. A methodological framework for assessing health indices. J Chron Dis. (1985) 38:27–36. doi: 10.1016/0021-9681(85)90005-0
13. Guyatt G, Walter S, Norman G. Measuring change over time: assessing the usefulness of evaluative instruments. J Chron Dis. (1987) 40:171–8. doi: 10.1016/0021-9681(87)90069-5
14. Mollayeva T, Pacheco N, D'Souza A, Colantonio A. The course and prognostic factors of cognitive status after central nervous system trauma: a systematic review protocol. BMJ Open. (2017) 7:e017165. doi: 10.1136/bmjopen-2017-017165
15. Mollayeva T, Colantonio A. Alzheimer's Disease in Men and Women With Trauma to the Central Nervous System. PROSPERO 2017 CRD42017055309 (2017). Available online at: http://www.crd.york.ac.uk/PROSPERO/display_record.php?ID=CRD42017055309 (accessed August 20, 2018).
16. Mollayeva T, Kendzerska T, Mollayeva S, Shapiro CM, Colantonio A, Cassidy JD. A systematic review of fatigue in patients with traumatic brain injury: the course, predictors and consequences. Neurosci Biobehav Rev. (2014) 47:684–716. doi: 10.1016/j.neubiorev.2014.10.024
17. Mollayeva T, Kendzerska T, Colantonio A. Self-report instruments for assessing sleep dysfunction in an adult traumatic brain injury population: a systematic review. Sleep Med Rev. (2013) 17:411–23. doi: 10.1016/j.smrv.2013.02.001
18. Hayden JA, Cote P, Bombardier C. Evaluation of the quality of prognosis studies in systematic reviews. Ann Intern Med. (2006) 144:427–37. doi: 10.7326/0003-4819-144-6-200603210-00010
19. Scottish Intercollegiate Guidelines Network. (2013). Published Guidelines. Retrieved from: https://www.sign.ac.uk/checklists-and-notes.html (accessed August 20, 2018).
20. Holmbeck GN, Thill AW, Bachanas P, Garber J, Miller KB, Abad M, et al. Evidence-based assessment in pediatric psychology: measures of psychosocial adjustment and psychopathology. J Pediatr Psychol. (2008) 33:958–80. doi: 10.1093/jpepsy/jsm059
21. Johansson B, Berglund P, Rnnbck L. Mental fatigue and impaired information processing after mild and moderate traumatic brain injury. Brain Inj. (2009) 23:1027–40. doi: 10.3109/02699050903421099
22. Dautricourt S, Violante I, Mallas E.-J, Daws R, Ross E, Jolly A, et al. Reduced information processing speed and event-related EEG synchronization in traumatic brain injury (P6.149). Neurology. (2017) 88.
23. Dikmen S, Machamer J, Temkin N. Mild traumatic brain injury: longitudinal study of cognition, functional status, and post-traumatic symptoms. J Neurotrauma. (2017) 34:1524–30. doi: 10.1089/neu.2016.4618
24. Farbota KDM, Sodhi A, Bendlin B, McLaren DG, Xu G, Rowley HA, et al. Longitudinal volumetric changes following traumatic brain injury: a tensor based morphometry study. J Int Neuropsychol Soc. (2012) 18:1006–18. doi: 10.1017/S1355617712000835
25. Kersel DA, Marsh NV, Havill JH, Sleigh JW. Neuropsychological functioning during the year following severe traumatic brain injury. Brain Inj. (2001) 15:283–96. doi: 10.1080/02699050010005887
26. Kwok FY, Lee TM, Leung CH, Poon WS. Changes of cognitive functioning following mild traumatic brain injury over a 3-month period. Brain Inj. (2008) 22:740–51. doi: 10.1080/02699050802336989
27. Liberman JN, Stewart WF, Wesnes K, Troncoso J. Apolipoprotein E 4 and short-term recovery from predominantly mild brain injury. Neurology. (2002) 58:1038–44. doi: 10.1212/WNL.58.7.1038
28. Losoi H, Silverberg ND, Wäljas M, Turunen S, Rosti-Otajärvi E, Helminen M, et al. Recovery from mild traumatic brain injury in previously healthy adults. J Neurotrauma. (2016) 33:766–76. doi: 10.1089/neu.2015.4070
29. Maksymiuk G, Szlakowska M, Stȩpień A, Modrzewski A. Changes in regional cerebral blood flow, cognitive functions and emotional status in patients after mild traumatic brain injury–retrospective evaluation. Biol Sport. (2005) 22:281–91. doi: 10.1089/neu.2017.5237
30. Mandleberg IA. Cognitive recovery after severe head injury. 3. WAIS Verbal and Performance IQs as a function of post-traumatic amnesia duration and time from injury. J Neurol Neurosurg Psychiatry. (1976) 39:1001–07. doi: 10.1136/jnnp.39.10.1001
31. Ponsford J, Cameron P, Fitzgerald M, Grant M, Mikocka-Walus A. Long-term outcomes after uncomplicated mild traumatic brain injury: a comparison with trauma controls. J Neurotrauma. (2011) 28:937–46. doi: 10.1089/neu.2010.1516
32. Powell TJ, Collin C, Sutton K. A follow-up study of patients hospitalized after minor head injury. Disabil Rehabil. (1996) 18:231–7. doi: 10.3109/09638289609166306
33. Sours C, Rosenberg J, Kane R, Roys S, Zhuo J, Gullapalli RP. Associations between interhemispheric functional connectivity and the Automated Neuropsychological Assessment Metrics (ANAM) in civilian mild TBI. Brain Imaging Behav. (2015) 9:190–203. doi: 10.1007/s11682-014-9295-y
34. Tofil S, Clinchot DM. Recovery of automatic and cognitive functions in traumatic brain injury using the functional independence measure. Brain Inj. (1996) 10:901–10. doi: 10.1080/026990596123873
35. Wylie GR, Freeman K, Thomas A, Shpaner M, OKeefe M, Watts R, et al. Cognitive improvement after mild traumatic brain injury measured with functional neuroimaging during the acute period. PLoS ONE. (2015) 10:e0126110. doi: 10.1371/journal.pone.0126110
36. Zaninotto AL, Vicentini JE, Solla DJ, Silva TT, Guirado VM, Feltrin F, et al. Visuospatial memory improvement in patients with diffuse axonal injury (DAI): a 1-year follow-up study. Acta Neuropsychiatr. (2017) 29:35–42. doi: 10.1017/neu.2016.29
37. Christensen BK, Colella B, Inness E, Hebert D, Monette G, Bayley M, et al. Recovery of cognitive function after traumatic brain injury: a multilevel modeling analysis of Canadian outcomes. Arch Phys Med Rehabil. (2008) 89(Suppl. 12):S3–15. doi: 10.1016/j.apmr.2008.10.002
38. Prigatano GP, Fordyce DJ, Zeiner HK, Roueche JR, Pepping M, Wood BC. Neuropsychological rehabilitation after closed head injury in young adults. J Neurol Neurosurg Psychiatry. (1984) 47:505–13. doi: 10.1136/jnnp.47.5.505
39. Robertson K, Schmitter-Edgecombe M. Self-awareness and traumatic brain injury outcome. Brain Inj. (2015) 29:848-58. doi: 10.3109/02699052.2015.1005135
40. Sandhaug M, Andelic N, Langhammer B, Mygland A. Functional level during the first 2 years after moderate and severe traumatic brain injury. Brain Inj. (2015) 29:1431–38. doi: 10.3109/02699052.2015.1063692
41. Schmitter-Edgecombe M, Robertson K. Recovery of visual search following moderate to severe traumatic brain injury. J Clin Exp Neuropsychol. (2015) 37:162–77. doi: 10.1080/13803395.2014.998170
42. Till C, Colella B, Verwegen J, Green RE. Postrecovery cognitive decline in adults with traumatic brain injury. Arch Phys Med Rehabil. (2008) 89(Suppl. 12):S25–34. doi: 10.1016/j.apmr.2008.07.004
43. Whyte J, Nakase-Richardson R, Hammond FM, McNamee S, Giacino JT, Kalmar K, et al. Functional outcomes in traumatic disorders of consciousness: 5-year outcomes from the national institute on disability and rehabilitation research traumatic brain injury model systems. Arch Phys Med Rehabil. (2013) 94:1855–60. doi: 10.1016/j.apmr.2012.10.041
44. Covassin T, Schatz P, Swanik CB. Sex differences in neuropsychological function and post-concussion symptoms of concussed collegiate athletes. Neurosurgery. (2007) 61:345–50. doi: 10.1227/01.NEU.0000279972.95060.CB
45. Covassin T, Stearne D, Elbin R. Concussion history and postconcussion neurocognitive performance and symptoms in collegiate athletes. J Athl Train. (2008) 43:119–24. doi: 10.4085/1062-6050-43.2.119
46. Covassin T, Elbin RJ, Harris W, Parker T, Kontos A. The role of age and sex in symptoms, neurocognitive performance, and postural stability in athletes after concussion. Am J Sports Med. (2012) 40:1303–12. doi: 10.1177/0363546512444554
47. Field M, Collins MW, Lovell MR, Maroon J. Does age play a role in recovery from sports-related concussion? A comparison of high school and collegiate athletes. J Pediatr. (2003) 142:546–53. doi: 10.1067/mpd.2003.190
48. Macciocchi SN, Barth JT, Littlefield L, Cantu RC. Multiple concussions and neuropsychological functioning in collegiate football players. J Athl Train. (2001) 36:303–6.
49. McCrea M., Barr WB, Guskiewicz K, Randolph C, Marshall SW, Cantu R, et al. Standard regression-based methods for measuring recovery after sport-related concussion. J Int Neuropsychol Soc. (2005) 11:58–69. doi: 10.1017/S1355617705050083
50. Meier TB, Bellgowan PS, Singh R, Kuplicki R, Polanski DW, Mayer AR. Recovery of cerebral blood flow following sports-related concussion. JAMA Neurol. (2015) 72:530–8. doi: 10.1001/jamaneurol.2014.4778
51. Register-Mihalik JK, Guskiewicz KM, Mihalik JP, Schmidt JD, Kerr ZY, McCrea MA. Reliable change, sensitivity, and specificity of a multidimensional concussion assessment battery: implications for caution in clinical practice. J Head Trauma Rehabil. (2013) 28:274–83. doi: 10.1097/HTR.0b013e3182585d37
52. Sosnoff JJ, Broglio SP, Ferrara MS. Cognitive and motor functions are associated following mild traumatic brain injury. Exp Brain Res. (2008) 187:563–71. doi: 10.1007/s00221-008-1324-x
53. Bleiberg J, Cernich AN, Cameron K, Sun W, Peck K, Ecklund PJ, et al. Duration of cognitive impairment after sports concussion. Neurosurgery. (2004) 54:1073–78. doi: 10.1227/01.NEU.0000118820.33396.6A
54. Chen C.-C, Wei S.-T, Tsaia S.-C, Chen X.-X, Cho D.-Y. Cerebrolysin enhances cognitive recovery of mild traumatic brain injury patients: double-blind, placebo-controlled, randomized study. Br J Neurosurg. (2013) 27:803–7. doi: 10.3109/02688697.2013.793287
55. Lee H, Kim SW, Kim JM, Shin IS, Yang SJ, Yoon JS. Comparing effects of methylphenidate, sertraline and placebo on neuropsychiatric sequelae in patients with traumatic brain injury. Hum Psychopharmacol. (2005) 20:97–104. doi: 10.1002/hup.668
56. Wang S, Cheng H, Dai G, Wang X, Hua R, Liu X, et al. Umbilical cord mesenchymal stem cell transplantation significantly improves neurological function in patients with sequelae of traumatic brain injury. Brain Res. (2013) 1532:76–84. doi: 10.1016/j.brainres.2013.08.001
57. Zafonte RD, Bagiella E, Ansel BM, Novack TA, Friedewald WT, Hesdorffer DC, et al. Effect of citicoline on functional and cognitive status among patients with traumatic brain injury: Citicoline Brain Injury Treatment Trial (COBRIT). JAMA. (2012) 308:1993–2000. doi: 10.1001/jama.2012.13256
58. Failla MD, Myrga JM, Ricker JH, Dixon CE, Conley YP, Wagner AK. Posttraumatic brain injury cognitive performance is moderated by variation within ANKK1 and DRD2 genes. J Head Trauma Rehabil. (2015) 30:E54–66. doi: 10.1097/HTR.0000000000000118
59. Snow P, Douglas J, Ponsford J. Conversational discourse abilities following severe traumatic brain injury: a follow up study. Brain Inj. (1998) 12:911–35. doi: 10.1080/026990598121981
60. Kontos AP, Elbin RJ, Kotwal RS, Lutz RH, Kane S, Benson PJ, et al. The effects of combat-related mild traumatic brain injury (mTBI): Does blast mTBI history matter? J Trauma Acute Care Surg. (2015) 79(Suppl. 2):S146–51. doi: 10.1097/TA.0000000000000667
61. Vanderploeg RD, Donnell AJ, Belanger HG, Curtiss G. Consolidation deficits in traumatic brain injury: the core and residual verbal memory defect. J Clin Exp Neuropsychol. (2014) 36:58–73. doi: 10.1080/13803395.2013.864600
62. Mokkink LB, Terwee CB, Patrick DL, Alonso J, Stratford PW, Knol DL, et al. The COSMIN study reached international consensus on taxonomy, terminology and definition of measurement properties for health-related patient-reported outcomes. J Clin Epidemiol. (2010) 6:737–45. doi: 10.1016/j.jclinepi.2010.02.006
63. Wilde MC, Boake C, Sherer M. Wechsler adult intelligence scale-revised block design broken configuration errors in nonpenetrating traumatic brain injury. Appl Neuropsychol. (2000) 7:208–14. doi: 10.1207/S15324826AN0704_2
64. Stallings G, Boake C, Sherer M. Comparison of the California verbal learning test and the rey auditory verbal learning test in head-injured patients. J Clin Exp Neuropsychol. (1995) 17:706–12. doi: 10.1080/01688639508405160
65. Broglio SP, Sosnoff JJ, Ferrara MS. The relationship of athlete-reported concussion symptoms and objective measures of neurocognitive function and postural control. Clin J Sport Med. (2009) 19:377–82. doi: 10.1097/JSM.0b013e3181b625fe
66. Iverson GL, Lovell MR, Collins MW. Validity of ImPACT for measuring processing speed following sports-related concussion. J Clin Exp Neuropsychol. (2005) 27:683–9. doi: 10.1081/13803390490918435
67. Hall KM, Hamilton BB, Gordon WA, Zasler ND. Characteristics and comparisons of functional assessment indices: disability rating scale, functional independence measure, and functional assessment measure. J Head Trauma Rehabil. (1993) 8:60–74. doi: 10.1097/00001199-199308020-00008
68. Haran FJ, Alphonso AL, Creason A, Campbell JS, Johnson D, Young E, et al. Reliable change estimates for assessing recovery from concussion using the ANAM4 TBI-MIL. J Head Trauma Rehabil. (2016) 31:329–38. doi: 10.1097/HTR.0000000000000172
69. Sherman EMS, Strauss E, Spellacy F. Validity of the paced auditory serial addition test (PASAT) in adults referred for neuropsychological assessment after head injury. Clin Neuropsychol. (1997) 11:34–45. doi: 10.1080/13854049708407027
70. Schwarz L, Penna S, Novack T. Factors contributing to performance on the Rey Complex Figure Test in individuals with traumatic brain injury. Clin Neuropsychol. (2009) 23:255–67. doi: 10.1080/13854040802220034
71. Tang L, Ge Y, Sodickson DK, Miles L, Zhou Y, Reaume J, et al. Thalamic resting-state functional networks: disruption in patients with mild traumatic brain injury. Radiology. (2011) 260:831–40. doi: 10.1148/radiol.11110014
72. Robertson IH, Manly T, Andrade J, Baddeley BT, Yiend J. ‘Oops!': performance correlates of everyday attentional failures in traumatic brain injured and normal subjects. Neuropsychologia. (1997) 35:747–58. doi: 10.1016/S0028-3932(97)00015-8
73. Bogod NM, Mateer CA, Macdonald SW. Self-awareness after traumatic brain injury: a comparison of measures and their relationship to executive functions. J Int Neuropsychol Soc. (2003) 9:450–8. doi: 10.1017/S1355617703930104
74. Adjorlolo S. Ecological validity of executive function tests in moderate traumatic brain injury in Ghana. Clin Neuropsychol. (2016) 30(Suppl. 1):1517–37. doi: 10.1080/13854046.2016.1172667
75. Wilson JT, Wiedmann KD, Hadley DM, Condon B, Teasdale G, Brooks DN. Early and late magnetic resonance imaging and neuropsychological outcome after head injury. J Neurol Neurosurg Psychiatry. (1988) 51:391–6. doi: 10.1136/jnnp.51.3.391
76. Paré N, Rabin LA, Fogel J, Pépin M. Mild traumatic brain injury and its sequelae: Characterisation of divided attention deficits. Neuropsychol Rehabil. (2009) 19:110–37. doi: 10.1080/09602010802106486
77. Pritagano GP. Wechsler memory scale: a selective review of the literature. J Clin Psychol. (1978) 34:816–32.
78. Reid DB, Kelly MP. Wechsler memory scale-RRevised in closed head injury. J Clin Psychol. (1993) 49:245–54.
79. Preece MH, Geffen GM. The contribution of pre-existing depression to the acute cognitive sequelae of mild traumatic brain injury. Brain Inj. (2007) 21:951–61. doi: 10.1080/02699050701481647
80. de Guise E, LeBlanc J, Champoux M, Couturier C, Alturki AY, Lamoureux J, et al. The mini-mental state examination and the Montreal cognitive assessment after traumatic brain injury: an early predictive study. Brain Injury. (2013) 27:1428–34. doi: 10.3109/02699052.2013.835867
81. Zhang H, Zhang XN, Zhang HL, Huang L, Chi Q, Zhang X, et al. Differences in cognitive profiles between traumatic brain injury and stroke: a comparison of the Montreal Cognitive Assessment and Mini-Mental State Examination. Chin J Traumatol. (2016) 19:271–4. doi: 10.1016/j.cjtee.2015.03.007
82. Walker WC, Hirsch S, Carne W, Nolen T, Cifu DX, Wilde EA, et al. Chronic Effects of Neurotrauma Consortium (CENC) multicentre study interim analysis: differences between participants with positive versus negative mild TBI histories. Brain Inj. (2018) 32:1079–89. doi: 10.1080/02699052.2018.1479041
83. O'Neil-Pirozzi TM, Goldstein R, Strangman GE, Glenn MB. Test-re-test reliability of the Hopkins Verbal Learning Test-Revised in individuals with traumatic brain injury. Brain Inj. (2012) 26:1425–30. doi: 10.3109/02699052.2012.694561
84. Mandleberg IA, Brooks DN. Cognitive recovery after severe head injury. 1. Serial testing on the Wechsler Adult Intelligence Scale. J Neurol Neurosurg Psychiatry. (1975) 38:1121–6. doi: 10.1136/jnnp.38.11.1121
85. Bleiberg J, Garmoe W, Halpern E, Reeves D, Nadler J. Consistency of within-day and across-day performance after mild brain injury. Neuropsychiatry Neuropsychol Behav Neurol. (1997) 10:247–53.
86. Roebuck-Spencer TM, Vincent AS, Twillie DA, et al. Cognitive change associated with self-reported mild traumatic brain injury sustained during the OEF/OIF conflicts. Clin Neuropsychol. (2012) 26:473–89. doi: 10.1080/13854046.2011.650214
87. Mandleberg IA. Cognitive recovery after severe head injury. 2. Wechsler Adult Intelligence Scale during post-traumatic amnesia. J Neurol Neurosurg Psychiatry. (1975) 38:1127–32. doi: 10.1136/jnnp.38.11.1127
88. Sun D, Daniels TE, Rolfe A, Waters M, Hamm R. Inhibition of injury-induced cell proliferation in the dentate gyrus of the hippocampus impairs spontaneous cognitive recovery after traumatic brain injury. J Neurotrauma. (2015) 32:495–505. doi: 10.1089/neu.2014.3545
89. Amen DG, Taylor DV, Ojala K, Kaur J, Willeumier K. Effects of brain-directed nutrients on cerebral blood flow and neuropsychological testing: a randomized, double-blind, placebo-controlled, crossover trial. Adv Mind Body Med. (2013) 27:24–33.
90. Cristofori I, Zhong W, Chau A, Solomon J, Krueger F, Grafman J. White and gray matter contributions to executive function recovery after traumatic brain injury. Neurology. (2015) 84:1394–401. doi: 10.1212/WNL.0000000000001446
91. Meyer V, Jones HG. Patterns of cognitive test performance as functions of the lateral localization of cerebral abnormalities in the temporal lobe. J Ment Sci. (1957) 103:758–72. doi: 10.1192/bjp.103.433.758
92. Ryan NP, Catroppa C, Cooper JM, Beare R, Ditchfield M, Coleman L, et al. The emergence of age-dependent social cognitive deficits after generalized insult to the developing brain: a longitudinal prospective analysis using susceptibility-weighted imaging. Hum Brain Mapp. (2015) 36:1677–91. doi: 10.1002/hbm.22729
Glossary
Keywords: measurements, neuropsychological tests, psychometrics, clinimetrics, systematic review
Citation: D'Souza A, Mollayeva S, Pacheco N, Javed F, Colantonio A and Mollayeva T (2019) Measuring Change Over Time: A Systematic Review of Evaluative Measures of Cognitive Functioning in Traumatic Brain Injury. Front. Neurol. 10:353. doi: 10.3389/fneur.2019.00353
Received: 22 October 2018; Accepted: 22 March 2019;
Published: 08 May 2019.
Edited by:
Karen M. Barlow, University of Queensland, AustraliaCopyright © 2019 D'Souza, Mollayeva, Pacheco, Javed, Colantonio and Mollayeva. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Tatyana Mollayeva, tatyana.mollayeva@mail.utoronto.ca
†These authors have contributed equally to this work and share first authorship