- 1Department of Experimental and Neurocognitive Psychology, Freie Universität Berlin, Berlin, Germany
- 2Center for Cognitive Neuroscience Berlin, Freie Universität Berlin, Berlin, Germany
- 3Centre for Cognitive Neuroscience, Department of Psychology, Paris Lodron University of Salzburg, Salzburg, Austria
Reading proficiency, i.e., successfully integrating early word-based information and utilizing this information in later processes of sentence and text comprehension, and its assessment is subject to extensive research. However, screening tests for German adults across the life span are basically non-existent. Therefore, the present article introduces a standardized computerized sentence-based screening measure for German adult readers to assess reading proficiency including norm data from 2,148 participants covering an age range from 16 to 88 years. The test was developed in accordance with the children’s version of the Salzburger LeseScreening (SLS, Wimmer and Mayringer, 2014). The SLS-Berlin has a high reliability and can easily be implemented in any research setting using German language. We present a detailed description of the test and report the distribution of SLS-Berlin scores for the norm sample as well as for two subsamples of younger (below 60 years) and older adults (60 and older). For all three samples, we conducted regression analyses to investigate the relationship between sentence characteristics and SLS-Berlin scores. In a second validation study, SLS-Berlin scores were compared with two (pseudo)word reading tests, a test measuring attention and processing speed and eye-movements recorded during expository text reading. Our results confirm the SLS-Berlin’s sensitivity to capture early word decoding and later text related comprehension processes. The test distinguished very well between skilled and less skilled readers and also within less skilled readers and is therefore a powerful and efficient screening test for German adults to assess interindividual levels of reading proficiency.
Introduction
Reading and comprehending written words and texts is an essential skill for an active and self-determined participation in everyday life. Successful reading includes basic decoding skills, often investigated by means of single word recognition (e.g., Jacobs and Grainger, 1994; Ziegler et al., 2001). It also relies on higher cognitive processes such as integration of syntax and general comprehension, which is mainly studied using longer passages or texts (e.g., Graesser et al., 1994; Singer et al., 1994; Myers and O’Brien, 1998). Reading also includes affective and aesthetic processes that have been neglected by mainstream reading research but now get more into focus in the emerging perspective of neurocognitive poetics (Lüdtke et al., 2014; Jacobs, 2015a, b, 2017; Lüdtke and Jacobs, 2015; Willems and Jacobs, 2016; Jacobs and Willems, 2018). Not surprisingly, the most common standardized screening procedures applied in reading research and educational settings focus mainly at the first two levels, i.e., reading abilities of participants are assessed via tests, which are based on recognizing words/naming letter strings [e.g., SLRT-II (Moll and Landerl, 2010), TOWRE (Torgesen et al., 2012)] or on answering comprehension questions related to text passages [e.g., Nelson-Denny Reading Test (Brown et al., 1993), ELFE-II (Lenhard et al., 2018)]. Only very few tests use single sentences to evaluate reading skills, such as the Sentence Reading Fluency test from the Woodcock-Johnson IV: Tests of Cognitive Abilities (Schrank et al., 2014) or the SLS 2-9 (Wimmer and Mayringer, 2014). Given that words, on the one hand, and texts, on the other hand represent a continuum of potential reading material, single sentences represent an intermediate form of written input. Basic decoding skills as well as aspects of integration and comprehension converge at this intermediate level. Thus, single sentences are well-suited for measuring reading proficiency including reading fluency, i.e., the ability to read accurately and at a rate that enables comprehension (Cunningham, 2001; Wolf and Katzir-Cohen, 2001). The importance for investigating reading abilities with words, sentences and texts is reflected in the number of well-established and broadly used tests that exist for English-speaking research participants. These existing tests cover all three levels and offer norm data for the whole life span (e.g., Woodcock-Johnson IV: Tests of Cognitive Abilities, Nelson-Denny Reading Test). However, when it comes to German-speaking participants, there is an apparent lack of comparable screening measures, especially when considering younger and older adults. The absence of such tests, particularly for German (older) adult readers, is even more astonishing, given the importance of skilled reading for its (potential) role for the development and preservation of verbal memory processes (Demoulin and Kolinsky, 2016) as well as for maintaining functional independence (in old age; e.g., Meyer and Pollard, 2006). Moreover, failing to provide adequate screening tests addressed to a German-speaking population may potentially endanger research within that area. Therefore, the aim of the present paper is to introduce a standardized computerized sentence-based screening measure for German adult readers (ages: 16–88 years), which can easily be implemented in any research setting in German orthography and which provides norms for reading proficiency based on 2,148 participants.
Reading Words, Sentences, and Texts
Reading proficiency is a complex, highly automatized multicomponent skill which combines a person’s ability to successfully process early word-based information and utilize this information in later processes of sentence and text comprehension. Early reading processes involve the effective interplay of recognizing single words and efficient eye movement control (Jacobs and Ziegler, 2015). More precisely, they include sublexical and lexical processing of orthographic, phonological and semantic information (Ziegler et al., 2008; Froehlich et al., 2016). Later processes related to the comprehension of sentences and texts contain component processes like deeper lexical, syntactic, and thematic analysis (Carpenter et al., 1995). All of them are necessary to construct representations of the information described in the reading material and enable understanding, enjoying and remembering (Rayner and Reichle, 2010; Jacobs, 2015a). According to the construction-integration model (Kintsch, 1988, 1998), proficient readers form three types of text representations, the surface form, the textbase and the situation model. While the surface level addresses word meaning and the syntactic form of the text, the textbase refers to the semantics of the text in form of propositions, that describe explicitly stated interrelationships of the given facts or ideas. The construction of a situation model requires a reader to integrate prior knowledge with the new information to build a coherent representation of the situation outlined by the text (e.g., Zwaan and Radvansky, 1998). Research has shown that situation models are created when reading single sentences (e.g., Radvansky and Dijkstra, 2007), i.e., already at the sentence level readers make use of preexisting pragmatic representations to facilitate comprehension (Carpenter et al., 1995). Processing sentences requires the reader to not only decode and assign meaning to words but also to make use of syntax, grammar, context (McClelland et al., 1989) and affective semantic information (Lüdtke and Jacobs, 2015). Furthermore, extra time is additionally allocated for conceptual integration at the end of sentences (Stine-Morrow et al., 2008). Besides, reading depends on accurate eye movement control and draws on a number of other fundamental cognitive and affective functions, such as attention, processing speed and working memory (Radvansky and Copeland, 2001; Reichle et al., 2003; Stine-Morrow et al., 2008), or mental simulation, empathy, immersion, affective evaluation and aesthetic appreciation (Willems and Jacobs, 2016; Jacobs and Willems, 2018).
To grasp and explain a proficient reading process, a number of computational models have been put forward at word (e.g., Grainger and Jacobs, 1996; Jacobs et al., 1998; Hofmann and Jacobs, 2014), sentence (McClelland et al., 1989), and text levels (e.g., Kintsch, 1988; Tzeng et al., 2005). Crucially, word identification models focus on early, basic processes of reading, whereas text comprehension models center largely on later processes establishing local and global coherence (Rayner and Reichle, 2010). The interaction of those early and late processes marks successful reading. Thus, the intermediate character of sentences provides an ideal way to assess reading proficiency, i.e., the quality with which the integration of early and late processes takes place.
Assessing Individual Differences in Reading Across the Life Span
Besides being easily implemented and evaluated, using sentences to assess reading proficiency has the advantage that early basic reading processes are not necessarily masked by later comprehension processes. Not surprisingly, tests based on sentence reading have been developed both in English (e.g., Sentence Reading Fluency Test from Woodcock-Johnson IV) and German to assess reading development in children. One widely used screening test for German speaking children is the Salzburger Lese-Screening (SLS; SLS 1-4: Mayringer and Wimmer, 2003; SLS 5-8: Auer et al., 2005; SLS 2-9: Wimmer and Mayringer, 2014). The SLS measures basic reading ability in a natural environment with special emphasis on reading speed by simultaneously assessing reading comprehension. In this paper and pencil test, children are asked to silently read simple sentences and to mark them as true or false depending on the plausibility of the sentence (e.g., “Tomatoes are blue.” from SLS 2-9, Wimmer and Mayringer, 2014). The screening consists of 100 sentences (SLS 2-9; 70 for the SLS 1-4 and 5-8) and is administered within a 3 min time span. The SLS has a good parallel test reliability (SLS 2-9: r = 0.95 and r = 0.87 for second and eighth graders, respectively). SLS scores correlate well with the speed of reading aloud (SLS 2-9: r = 0.80–0.89 and r = 0.49–0.55 for second and eighth graders, respectively) and with more complex measurements of reading competence of the OECD Programme for International Student Assessment study (PISA; preliminary version of SLS 5-8: r = 0.64; Wimmer and Mayringer, 2014). Moreover, adolescents with low test scores in sentences tasks similar to those used in the children’s SLS showed a higher number of fixations and also a longer gaze duration (Hawelka et al., 2010).
While English reading research and diagnostic assessment is rather privileged to further evaluate reading performance over the life span with tests based on sentence and text reading, no such thing has been established for German (older) adults. This is rather surprising. Even though there is evidence that core reading skills are preserved, the process of reading, the plasticity of its associated neural networks as well as its supporting cognitive functions undergo at least some developmental changes across the entire life span (e.g., Allen et al., 1991; Spieler and Balota, 2000; Hedden and Gabrieli, 2004; Rönnlund et al., 2005; Madden et al., 2007; Stine-Morrow et al., 2008; Froehlich et al., 2016; Froehlich et al., 2018). Moreover, as German and English differ with respect to the depth of their orthographies (e.g., Ziegler and Goswami, 2005), one cannot not simply transfer results of English readers to German readers as has been shown for both children and adults (e.g., Frith et al., 1998; Landerl, 2000; Ziegler et al., 2001; Rau et al., 2015; Oganian et al., 2016). Nonetheless, at the single word level, the age of an adult proficient reader seems to not affect orthographic/lexical and lexico-semantic processing in both orthographies (e.g., Daselaar et al., 2003; Ratcliff et al., 2004; Spaniol et al., 2006; Cohen-Shikora and Balota, 2016; Froehlich and Jacobs, 2016; Froehlich et al., 2016) while sublexical and phonological processing show age-related alterations (e.g., Allen et al., 1991; Madden et al., 2007; Froehlich and Jacobs, 2016; Froehlich et al., 2016). At the sentence and text level, age-related decrements have primarily been reported for surface and discourse processing whereas efficient processing of the situation model seems to be preserved (Radvansky and Dijkstra, 2007). Especially older adults seem to profit from contextual cues in longer texts, however, compared to younger adults, they need to allocate more resources during reading to ensure understanding (e.g., Stine-Morrow et al., 2008). As the vast majority of sentence and text related studies have been conducted with English younger and older participants, it still remains unknown whether these results are applicable to German participants. To our knowledge, the only studies investigating age-related differences in sentence reading in German adults focused primarily on eye tracking measures (Kliegl et al., 2004) or on syntactic complexity (Junker, 2005) underscoring the need for further research within this specific area.
The Present Study
The overall aim of this study is to introduce a standardized sentence-based screening procedure, the SLS-Berlin, to reliably capture reading proficiency across the life span and thus foster research in that area. For this purpose, we adapted 77 sentences from the speeded sentence reading test (cf. Bergmann and Wimmer, 2008; Hawelka et al., 2010; Wimmer et al., 2010) for which plausibility judgments had to be made. The resulting SLS-Berlin score, i.e., the sum of correctly answered sentences within a 3-min time frame, then indicates the level of reading proficiency which can be compared to an age-matched norming sample. The 3-min cutoff was chosen in accordance with the children’s version of the SLS (Mayringer and Wimmer, 2003; Auer et al., 2005; Wimmer and Mayringer, 2014).
Word-based measures of reading assessment often focus on reading speed and/or accuracy. These measures are somewhat narrow in their scope and underrepresent the construct of reading (Cunningham, 2001; Kuhn and Stahl, 2003). Given the central intermediate character of sentences (cf. see section “Reading Words, Sentences, Texts”), sentences based measures appear to be an ecologically more valid indicator for reading proficiency. Moreover, by providing norms based on a large age diverse sample (ages: 16–88 years, N = 2,148), it is not only possible to compare and/or select participants according to their rank in reading proficiency within one cohort but also across cohorts. The latter is especially important as speeded measures usually penalize increasing age (cf. Salthouse, 1996; Stine-Morrow et al., 2008). The SLS-Berlin can be administered to both individuals as well as groups, is easy to implement and to analyze.
Additionally, we aimed at establishing convergent validity for the SLS-Berlin. To assess the test’s ability to capture basic early processes of reading proficiency, participants had to first read words or pseudowords. Secondly, to validate the SLS-Berlin’s potential of assessing more complex, i.e., later, processes of reading proficiency, participants had to read texts with their eye-movements being recorded. Eye-tracking, being rather elaborate and time consuming for a simple screening procedure, is an otherwise excellent method to investigate individual differences in reading performance, including developing and aging readers (Schroeder et al., 2015). Whereas measures such as first fixation duration (FFD) are thought to reflect early processes related to visual word recognition and orthographic/lexical access, measures such as total reading time (TRT) are considered to capture later processes associated with comprehension (e.g., Hyönä et al., 2003; Radach and Kennedy, 2004).
Finally, as we collected data from 1,493 persons being at least 60 years or older, we also expect to shed more light on the development of written language processing during later and late adulthood in German given the sparsity of available data, especially when it comes to sentence and text reading, i.e., reading in an ecologically valid setting.
To meet our objectives, in Study 1, we describe the sentence characteristics for all 77 sentences included in the SLS-Berlin as well as the characteristics of the norm sample. We report the distribution of the SLS-Berlin scores for the entire sample and the dependencies of the score from the sample characteristics. Additionally, we present how sentence characteristics relate to the response times of all sentences. Finally, Study 2 was conducted to further validate the SLS-Berlin. To do so, we compared the SLS-Berlin test scores of a new sample with two single (pseudo)word reading tests, one processing speed and attention test and eye-movements during text-reading.
Study 1: the SLS-Berlin
Methods
SLS-Berlin
Contrary to the children’s version of the SLS, the SLS-Berlin is a computer-based screening test. After a short adaptation phase, participants must read all 77 sentences which increase in their complexity. Participants are instructed to verify the content of each of the sentences by deciding whether it violates basic world knowledge. Participants are instructed to make the plausibility judgments for all sentences as quickly and as accurately as possible. In accordance with the children’s version of the SLS, the final SLS-Berlin score is derived from the number of correctly judged sentences within a 3-min time frame. To deemphasize the focus on processing speed, the 3-min cut-off is not mentioned, and participants are not interrupted while reading and judging the 77 test sentences. On average, it takes participants approximately 7 min to complete the test.
Stimuli
The 77 sentences presented within the SLS-Berlin were taken from a pool of 154 sentences for an adult paper-pencil version (Bergmann and Wimmer, 2008; Hawelka et al., 2010; Wimmer et al., 2010) of the SLS 1-4 (Wimmer and Mayringer, 2014). To minimize shallow processing and skim reading all sentences included at least one long word (with more than six letters). Sentences contained either confirmations or violations of basic world knowledge (e.g., “For safety reasons, smoking is prohibited at petrol stations” and “With a scale, the height of a person is measured,” respectively). Correct (confirmations) and incorrect (violations) sentences did not differ with respect to sentence length, number of punctuation marks and number of long words. In total, we selected 36 correct and 41 incorrect easy to understand sentences of varying length (ranging from 4 to 18 words; see Supplementary Table 1). Table 1 reports sentence characteristics associated with sentence length, such as Number of all Characters, Number of Syllables, Number of Words, and Number of Long Words (with more than six letters) as well as sentence characteristics typically associated with syntax, such as Number of Nouns, Number of Punctuation Marks and the German adaptation of the Flesch’s Reading Ease Index (Amstad, 1978). Also reported are the results of the t-tests comparing sentence characteristics between correct and incorrect sentences.
Table 1. Descriptive statistics of different sentences characteristics for all 77 SLS-Berlin sentences and differences between correct and incorrect sentences.
For the final version of the SLS-Berlin, sentences were ordered in a fixed sequence of correct and incorrect sentences so that no more than three sentences of one type were presented in a row (see Figure 1, upper section). Sentence order followed the principle of increasing complexity with sentences containing fewer and shorter words being shown more to the beginning and sentences with more and longer words closer to the end of the test (see Table 2 for examples of simple and more complex correct and incorrect sentences).
Figure 1. Distribution of correct and incorrect sentences (upper part) and split of all 77 sentences into first and second test part to calculate split-half reliability (lower part).
The principle of increasing complexity was tested in a prestudy. Here, 19 native German readers (M = 22.2 years, SD = 1.92, range: 19–26, 14 females) made plausibility judgments on each of the 77 sentences, marked the Point of Decision (word at which plausibility judgment was possible), and rated the Sentence Complexity on a scale ranging from 1 (simple) to 5 (complex). Mean ratings for Sentence Complexity ranged from 1 (sentence 1) to 3.68 (sentence 69, see also Supplementary Table 1). The Mean Point of Decision (reported in percent) occurs rather late in the sentence, i.e., after reading more than 90% of the sentence (M = 92.8%, SD = 9.43%).
The intended increasing complexity was verified by a regression analysis1 predicting Sentence Complexity ratings by sentence position within the test (Sentence ID), which identified a positive cubic relationship: adjusted (adj.) R2 = 0.51, F(1,74) = 27.3, p < 0.001.
To further describe the sentences used in the SLS-Berlin, we conducted regression analyses (cf., see section “Study 1-Data Analysis”) with Sentence ID as a predictor for all sentence characteristics. The results are reported in Table 3. Significant positive cubic relationships were identified for Number of all Characters, Number of Syllables and Number of Words [all adj. R2s > 0.27, all Fs(1,74) > 10.5, all ps < 0.001]. For Number of Nouns, Number of Long Words and the Flesch Index significant linear relationships were observed [all adj. R2s > 0.15, all Fs(1,74) > 13.8, all ps < 0.001]. These significant results all confirm the (non-linearly) increasing complexity of the sentences. Only for one sentence characteristic, Mean Point of Decision (in percent), no significant relationship with Sentence ID was observed (F < 1). This was also the only variable for which the comparison for correct and incorrect sentences indicates a significant difference [t(76) = 3.60, p < 0.001, d = 0.82]. The Mean Point of Decision (in percent) for incorrect sentences (M = 89.6%, SD = 11.5%) occurs earlier than for correct sentences (M = 96.7%, SD = 3.63%). However, most importantly, the response times of incorrect vs. correct sentences did not differ from each other within the norm and the subsamples (see the following section; all ts < |0.34|, all ps > 0.73).
Table 3. Model parameters and summaries of linear and polynomial regression analyses of Sentence ID predicting SLS-Berlin sentence characteristics.
Norm Sample
We collected data from 2,174 native German speaking adults with age ranging from 16 to 88 years. Twenty-three participants had to be excluded as they obtained negative SLS-Berlin scores (i.e., they classified more sentences incorrectly than correctly within 3 min reading time). Three additional participants were removed from all further analyses due to technical difficulties while recording their data. In all, the norm sample consists of data from 2,148 adult participants (see Table 4 for sample characteristics).
Table 4. Sample sizes and characteristics of norm sample and the two subsamples of younger adults (<60) and older adults (> = 60).
As most research in adult reading is conducted with either younger (ages: 20 to 35) and/or older readers (60 years and older), we focused on these age groups while collecting norm data. However, to provide a norm sample which covers a wide range of the adult life span, we also present data for ages 36 to 59 (see Supplementary Table 2 for norm table). More specifically, we report analyses with respect to the entire norm sample [N = 2,148, Mage = 58.9 years, SDage = 19.0, 1,130 females (52.6%), Meducation = 14.6 years, SDeducation = 2.97] as well as for two subsamples, i.e., for all participants being younger than 60 years of age [subsample younger adults; N = 655, Mage = 30.4 years, SDage = 7.01, range: 16–59, 368 females (56.2%)] and those being 60 or above [subsample older adults; N = 1,493, Mage = 70.1 years, SDage = 3.93, range: 60–88, 762 females (51.0%)]. The subsamples differed with respect to years of formal education with the younger subsample yielding more years of formal education (Meducation = 15.4, SDeducation = 2.98) than did the older subsample (Meducation = 14.2, SDeducation = 2.89), t(1079.1) = 8.06, p < 0.001, d = 0.41.
All participants had normal or to corrected-to-normal vision, reported no history of language impairment, neurological disease, psychiatric disorders or a history of head injuries. Prior to testing, participants gave written informed consent. Most participants received financial compensation, a minority within the younger subsample obtained course credit for their participation. The study was approved by the Ethics Committee of the Max Planck Institute for Human Development, Berlin (MPIB) and by that of the Freie Universität Berlin (FU Berlin).
Procedure
Test sessions took place at the FU Berlin and at the Charité Campus, Berlin between the years 2013 and 2017. The 1,884 subjects tested at the Charité Campus were part of the Berlin Aging Study II cohort (BASE-II: Bertram et al., 2014; Gerstorf et al., 2016) and participated in groups of up to six individuals. At the FU Berlin, data from 264 subjects were collected in individual test sessions. At both locations, additional tests were conducted, however, the SLS-Berlin was always the first part of the experimental session and identical in its procedure.
Contrary to the children’s version of the SLS, the SLS-Berlin was run on a computer while using Objective-C (MPIB) and Python 2.7 (FU). Ten training sentences and the 77 test sentences were shown individually one after another at the center of a computer screen (left aligned and single spaced) in 32 pt Helvetica using black font on white background. Prior the presentation of each sentence a fixation cross appeared at the middle of the screen for 1500 ms, followed by the sentence. Short sentences were presented in one line, longer sentences in two or three lines. Participants were asked to make plausibility judgments as quickly and as accurately as possible after reading each of the sentences. Responses were recorded with the help of a button box or a standard keyboard using the index and middle finger of the left hand. By demanding only little motor movements, the SLS-Berlin is well suited to be used in MRT and EEG settings. Feedback on responses was only given during training trials. The instruction did not stress the time dependency of the test. Contrary to the paper pencil versions for children and adults, the 3-min cut-off for calculating the final test score was not mentioned and participants were not interrupted while reading and judging the 77 test sentences. Overall, the completion of the SLS-Berlin took approximately 7 min. The final SLS-Berlin score of each participant was calculated by adding the number of correctly answered sentences within a 3-min time window of pure reading time (i.e., omitting the times of fixation crosses).
Data Analysis
Analyses of the SLS-Berlin scores and the response times were conducted with R version 3.4.3 (R Core Team, 2017). All analyses were performed for the whole norm sample as well as for the two subsamples. In a first step, we assessed the consistency of the test by calculating the split-half reliability. For this purpose, the 77 sentences of the SLS-Berlin were split into two parts with the number of correct and incorrect sentences being roughly the same (part 1 18 correct and 21 incorrect, part 2: 18 correct and 20 incorrect). We also controlled for mean Sentence IDs (part 1: 39.5; part 2: 38.4; cf. Figure 1, lower section) given their predictive power for the other item characteristics. We then obtained the number of correctly answered sentences within 1.5 min per subject to calculate the SLS-Berlin scores for both parts of the test and correlated the results using the Spearman-Brown prophecy formula (Brown, 1910; Spearman, 1910).
In a second step, we analyzed the relationship between the SLS-Berlin score and sample characteristics (i.e., Gender, Age, and Education; see Table 5) as well as assessed the influence of the above described sentence characteristics on the response times for the sentences (see Tables 6–8). To do so, we used only correctly classified sentences (data loss of 4.73%). Outliers based on response times were excluded by estimating a linear mixed-effects model with random intercepts for participants and items using the lme4 package (version 1.1-15; Bates et al., 2015). Response times with standardized residuals larger than ±3 were removed from the data set (data loss of 1.48%).
Table 5. Model parameters and summaries of linear and polynomial regression analyses of SLS-Berlin scores predicted by Age and Education for all participants and the two subsamples of younger (<60 years) and older adults (> = 60 years).
Table 6. Model parameters and summaries of linear and polynomial regression analyses of SLS-Berlin sentence characteristics predicting response times for the entire norm sample.
Table 7. Model parameters and summaries of linear and polynomial regression analyses of SLS-Berlin sentence characteristics predicting response times for the subsample of younger adults (<60 years).
Table 8. Model parameters and summaries of linear and polynomial regression analyses of SLS-Berlin sentence characteristics predicting response times for the subsample of older adults (> = 60 years).
To assess the model fit of the regression analyses and to test the accuracy of the predictions, we randomly assigned 70% of the sample data to a training group. The remaining 30% were assigned to a test group. Neither the training nor the test group differed in terms of mean response times, age, gender distribution, years of education and mean SLS-Berlin score from the complete sample as well as from each other (all p’s > 0.20). We conducted linear, quadratic and cubic regression analyses for each of the sentence characteristics. All regression models were checked for influential cases by inspecting Cook’s distance (Cook, 1977), the models’ standardized residuals as well as the data’s leverage. Observations were removed if Cook’s distance was larger than 1 (Pardoe, 2012), if standardized residuals were larger than ±3 or if they had a leverage above 3*(k+1)/n with k being the number of predictors in the model and n the sample size. Finally, we identified the best fitting model for each sentence characteristic and fitted the winning model on the remaining 30% of the population to compare the predicted with the observed values. If these values turned out to be comparable, this regression model was applied to the entire population. Model parameters and summaries as well as results from the cross-validation procedure are displayed in Tables 6–8.
Results
Split-Half Reliability
The internal consistency of the SLS-Berlin is very high with a split-half reliability for the test scores of rAll = 0.91, t(2146) = 101.1, p < 0.001 for the entire norm sample. Separate analyses showed a slightly higher reliability for the younger compared to the older subsample, rY = 0.93, t(653) = 66.0, p < 0.001 and rO = 0.89, t(1,491) = 73.5, p < 0.001. The comparison of both reliability coefficients with Fischer’s z by using the cocor module (Diedenhofen and Musch, 2015) indicated statistical significance (z = 5.03, p < 0.001, CI [0.02, 0.05]). Correcting for original test length using the Spearman-Brown formula, reliabilities increased to rAll = 0.95 for the entire norm sample and to rY = 0.97 and rO = 0.94 for the younger and older subsamples, respectively. All coefficients indicate a high test reliability (cf. Flom et al., 2012).
Distribution of SLS-Berlin Scores and Predictions by Sample Characteristics
Figures 2A,B shows the distribution of SLS-Berlin scores for the younger and older subsample. 3.21% of the younger and 0.54% of the older adults scored the highest possible value (77) leading to slightly left-skewed distributions. We observed no gender differences, neither in the entire norm sample (t < 1) nor in one of the two subsamples (both ts ≤ 1), however, both Age and Education significantly predicted SLS-Berlin scores within the norm sample (both adj. R2s < 0.07, ps < 0.001; see Table 5). For Age, we observed a quadratic, for Education a cubic relationship. Separate analyses for the two subsamples indicated no effect of Age on SLS-Berlin scores for younger adults (F < 1), yet a significant influence was observed for the older subsample (adj. R2 = 0.03, p < 0.001; see Figures 2C–F and Table 5). The quadratic nature of this relationship resulted from a steeper decrease starting around age of 60, which flattened around the age of 70. For Education, we observed small, albeit significant, quadratic effects for both subsamples. It was slightly stronger for older (adj. R2 = 0.05, p < 0.001) compared to younger adults (adj. R2 = 0.03, p = 0.001). Similar low adjusted R2s from the cross-validation procedure underpinned the generally small effects of Age and Education on SLS-Berlin scores.
Figure 2. Distribution of the SLS-Berlin scores in the two subsamples of younger (A) and older adults (B) as well as the relationship of Age and Education on the SLS-Berlin score for the subsamples of younger (C,E) and older adults (D,F).
Regression Analysis of Response Times
Model parameters and summaries of linear and polynomial regressions for the norm sample as well as the two subsamples of younger and older adults are displayed in Tables 6–8, respectively. The cross-validation produced very similar adj. R2 compared to final models calculated for all participants (cf. Table 6), for the younger subsample (cf. Table 7) and the older subsample (cf. Table 8) indicating a very good generalizability of our data. Figure 3 depicts the relationships between six selected sentence characteristics and the average response times for the 77 sentences from the SLS-Berlin.
Figure 3. Relationships between sentence characteristics Sentence ID (A), Complexity ratings (B), Number of characters (C), Number of syllables (D), Number of words (E), and Number of long words (F) and response times for the test sentences from the SLS-Berlin for the entire norm sample.
Norm sample
Except for the German adaptation of the Flesch Reading Ease Scores (Amstad, 1978) and the Mean Point of Decision (in percent), all SLS-Berlin sentence characteristics predicted significantly the response times in the norm sample with adj. R2 ranging from 0.26 to 0.78, all Fs > 28.3, all ps < 0.001. Best model fit (adj. R2) was obtained for Sentence ID (0.78), followed closely by Complexity Ratings (0.77) and further by Number of Characters (0.73). Number of Syllables showed the fourth best fit (0.65). Word related measures, i.e., Number of Words, Number of Long Words, and Number of Nouns followed suit (0.51, 0.37, and 0.26, respectively). However, the nature of the relationship between sentence characteristics and response times varied with respect to the predictor (see Figure 3). While a linear relationship was observed for Number of Characters, Number of Syllables and Number of Words, a quadratic one was registered for Complexity Ratings, Number of Words and Number of Long Words. Sentence ID produced a cubic model fit.
Younger subsample
Results of the younger subsample were very similar to those of the norm sample. Again, the German adaptation of the Flesch Reading Ease Scores and the Mean Point of Decision (in percent) showed no predictive power on response times. All other seven sentence characteristics produced slightly higher adj. R2 than the norm sample ranging from 0.29 to 0.82, all Fs > 29.9, all ps < 0.001. The order with which significant predictors explained data variance (adj. R2) was identical to that of the norm sample: Sentence ID (0.82), Complexity Ratings (0.71), Number of Characters (0.70), Number of Syllables (0.61), Number of Words (0.43), Number of Long Words (0.35), and Number of Nouns (0.29).
Comparable to the norm sample, we found linear relationships for Complexity Ratings, Number of Long Words and Number of Nouns. However, with respect to Number of Syllables the linear regression also showed the best model fit (vs. a quadratic one in the norm sample). The only quadratic relationship between predictor and outcome variable was identified for Number of Words (vs. Number of Words, Number of Characters and Numbers of Syllables in the norm sample). Cubic model fits were observed for Sentence ID (as was in the norm sample) and additionally for Number of Characters (vs. a quadratic fit in the norm sample).
Older subsample
Like the younger subsample, results of the older subsample resembled those of the norm sample. As before, the German adaptation of the Flesch Reading Ease Scores and the Mean Point of Decision (in percent) did not yield any significant effects while the other seven sentence characteristics produced reliable model fits, with adjusted R2s ranging from 0.24 to 0.78, all Fs > 23.5, all ps < 0.001. However, different from the norm sample and the younger subsample, the order of the strongest two predictors (adj. R2) was reversed with Complexity Ratings explaining more variance in response times (0.78) than did Sentence ID (0.74). The order of the other predictors, e.g., Number of Characters (0.73), Number of Syllables (0.65), Number of Words (0.56), Number of Long Words (0.38), and Number of Nouns (0.24) was yet again comparable to that of the norm sample and the younger subsample.
The only linear relationship was identified for Number of Nouns (as was in the norm sample and the younger subsample). Complexity Ratings, Number of Characters and Number of Syllables all showed quadratic fits (the latter two in accordance with the norm sample, while the first being linear in the norm sample). Sentence ID and Number of Words yielded a cubic relationship between predictor and outcome variables in the older subsample (the first in accordance with the norm sample while the latter being quadratic in the norm sample).
Discussion
We collected data from > 2,000 participants including a group of younger (below 60 years of age) and a group of older adults (above 60). The split-half reliabilities for the entire norm sample as well as for the two subsamples indicated a high consistency of the SLS-Berlin for all investigated age groups.
Contrary to the children’s version of the SLS, which reports slightly higher reading scores for girls than for boys (cf. Wimmer and Mayringer, 2014), the SLS-Berlin is independent of gender. This is well in line with current results about gender differences in cognitive components of adult reading comprehension. As pointed out by Hannon (2014), quantitative gender differences are mostly marginal, especially in comparison to the more important differences in the predictive powers of specific cognitive components of reading comprehension.
In both the younger and the older subsample, years of formal education had only minor influences on SLS-Berlin scores, i.e., test scores slightly increased with increasing years of education. Age related effects were observed for older adults only. The quadratic relationship indicated a steeper decline at the age of 60, which almost leveled out around 70 years of age. These rather late effects correspond to reported age effects for lexical access during single word recognition (e.g., Cohen-Shikora and Balota, 2016; Froehlich et al., 2016) and building successful situational representations in text comprehension (e.g., Radvansky, 1999; Radvansky and Dijkstra, 2007). A relatively stable performance up to the age of 60 has likewise been described in longitudinal studies for cognitive functions such as verbal ability, verbal (working) memory as well as episodic memory (Hedden and Gabrieli, 2004; Rönnlund et al., 2005) known to contribute to reading performance (DeDe et al., 2004; Stine-Morrow et al., 2008; Frick et al., 2011).
Taken together, the SLS-Berlin is a robust screening test to assess reading proficiency over the entire life span. The absence of gender effects, the minor effects of Education and Age all imply versatile application possibilities in both, research and educational/diagnostic settings. Particularly, the absence of significant age effects in the younger age group below 60, marks the SLS-Berlin as an excellent screening test for research projects focusing on changes in cognitive functioning across the life span.
Our detailed analyses of response times of the test sentences constituting the SLS-Berlin clearly underpin the increasing complexity, i.e., the further the test progresses the more complex the sentences become (cf., see section “Study 1-Stimuli”). This distinguishes the SLS-Berlin from the Woodcock-Johnson Sentence Reading Fluency Task (Schrank et al., 2014). The increasing complexity relates directly to increasing reading times as the regression analyses of Complexity Ratings on response times showed for the entire norm sample as well as for both subsamples. Sentence ID as well as Complexity Ratings were the strongest predictors for reading times. That is not surprising given that Sentence ID is positively confounded with all other item characteristics. The same can be assumed for Complexity Ratings. The cubic relationship between Sentence ID and response times indicated a strong increase in response times, especially within the first 20 and the last 20 sentences. Therefore, we conclude that the SLS-Berlin is not only sensitive to individual differences within poor readers, between poor and good readers but also within good readers. Nevertheless, the sensitivity for the upper performance spectrum of reading proficiency appears to be not as high as for the lower spectrum as indicated by the slightly right-skewed distribution of SLS-Berlin scores.
The reported relationships between sentence characteristics and response times also indicated that neither the Flesch’s Reading Ease Index nor the Mean Point of Decision (in percent) reliably explained variance in response times. As the Flesch’s Reading Ease Index has been developed to particularly assess the difficulty of texts, the former result is not surprising and suggests the evaluation of the complexity of single sentences with the help of rating studies rather than the Flesch’s Reading Ease Index. The finding regarding the Mean Point of Decision (in percent) suggests that participants had read the complete sentences of the SLS-Berlin rather than terminating the reading process early, for instance, before or as soon as a pragmatic violation within the sentence occurred. This underscores particularly well the fact that the SLS-Berlin is sensitive to both early visual processing associated with decoding and later comprehension processes. In order to further explore this conclusion and to fully validate the SLS-Berlin, we compared it to other, already established, measures of reading (ability).
Study 2: Comparing the Sls-Berlin to Other Reading Measures
Study 2 was conducted to describe the relationship between SLS-Berlin and other reading measures. More importantly, to show that our screening test at the sentence level captures both basic decoding processes such as word recognition and later more complex processes such as meaning comprehension and integration. We therefore compared the SLS-Berlin scores with results from single word and pseudoword reading as well as from tasks focusing on text comprehension. Relating performance in single word processing as well as text reading to the SLS-Berlin captures the two essential foci of empirical research about text reading (Stine-Morrow et al., 2006; De Beni et al., 2007; Radvansky and Dijkstra, 2007; Borella et al., 2011; Froehlich et al., 2018) and the development and changes in reading and language comprehension over the life span (Hannon, 2012; Hannon and Frias, 2012; Froehlich and Jacobs, 2016; Froehlich et al., 2018).
To evaluate the relationship between the SLS-Berlin and measures for word identification, we correlated SLS-Berlin scores with a well-established test for single word and pseudoword reading. Considering that timed reading of unconnected words and pseudowords underrepresent the construct of reading fluency (Cunningham, 2001; Kuhn and Stahl, 2003), we expected strong but not perfect correlations of SLS-Berlin scores with these very basal measures of reading proficiency.
Similar to other standardized tests of reading fluency and proficiency for children and adults (e.g., SLRT-II, SLS 2-9, Sentence Reading Fluency Test from Woodcock-Johnson IV: Tests of Cognitive Abilities), scores of the SLS-Berlin are time dependent. Even though participants are not made aware of the 3-min cutoff, they are instructed to read as fast and as accurately as possible. Given the time dependency of the SLS-Berlin participants with higher processing speed should have an advantage compared to participants with lower processing speed regardless of their reading proficiency. Therefore, we additionally administered a test for visual attention and processing speed. Here, we expected a positive correlation between general processing speed and SLS-Berlin scores. However, this correlation should be noticeably weaker than the correlations between SLS-Berlin scores and the indicators for word/pseudoword reading and text comprehension as the SLS-Berlin was designed to specifically assess reading proficiency without emphasizing the speed component.
To further evaluate the SLS-Berlin’s ability and to capture more complex processes related to comprehension, we also obtained eye-tracking data during text reading as an external criterion (cf. Moore and Gordon, 2015). Eye-tracking is an excellent method to investigate visual and cognitive processing during text reading and comprehension (Adler-Grinberg and Stark, 1978; Duffy, 1992; Rayner et al., 2006). Several studies conducted with both, children and adults, demonstrated that individual differences in reading fluency and proficiency are associated with differences in eye movement patterns (Kemper et al., 2004; Ashby et al., 2005; Hawelka et al., 2010; Kuperman and Van Dyke, 2011). Fast readers compared to slow readers, for example, were found to produce fewer fixations, shorter gaze durations, larger saccades and more skipping when reading single sentences and short texts (Ashby et al., 2005; Thaler et al., 2009; Hawelka et al., 2015; Krieber et al., 2016). Corresponding patterns were found for the differences between dyslexic and normal readers (Hyönä and Olson, 1995; Hutzler and Wimmer, 2004; Hawelka et al., 2010). If, as assumed, the SLS-Berlin score reliably measures individual differences in reading proficiency, it should therefore also reliably predict differences in eye movement patterns observed during text reading. Moreover, eye-tracking research in text comprehension indicates that eye movement measures reflect also higher order comprehension processes to some extent (Duffy, 1992; Rayner et al., 2006). Rayner et al. (2006) assumed, that “higher order comprehension processes can override the normal default situation in which lexical processing is driving the eyes and result in longer fixations or regressions back to earlier parts of the text” (p. 244). To underline the SLS-Berlin’s sensitivity for both early and automatic decoding processes as well as higher order comprehension processes, we analyzed two different eye-tracking parameters, i.e., FFD and TRT. While the former is usually associated with initial visual recognition and lexical access, the latter is related to higher order comprehension processes (see Table 16.4 in Hyönä et al., 2003, and Tables 3, 4 in Boston et al., 2008, for a detailed description of eye movement measures and relevant references). Eye tracking data was obtained while participants read two texts with different levels of complexity. For the less complex text, we assumed the SLS-Berlin to be highly predictive, especially with regards to TRT. However, as higher order comprehension processes associated with reading difficult texts may overwrite basic processes related to reading proficiency (cf. Rayner et al., 2006), we expected less predictive power of SLS-Berlin scores on FFD and TRT in the more complex text.
Methods
Test Materials
Word and pseudoword recognition
To measure word and pseudoword recognition, we conducted two subtests of the Salzburger Lese- und Rechtschreib-Test [Salzburg Reading and Spelling Test (SLRT II); Moll and Landerl, 2010]. Each subtest contains a word and a pseudoword list with increasing item length and complexity. Participants are asked to read out loud as much items as possible without making pronunciation errors. The final score is the number of correctly read items within 1 min and is separately obtained for words and pseudowords.
Visual attention and processing speed
To measure serial allocation of visual attention and discrimination as well as processing speed, we used a language free version (i.e., the ‘Smiley-Test’; cf. Landerl, 2001; Marx et al., 2016) of the d2-R test (Brickenkamp et al., 2010). In the original version of the d2-R, participants receive a sheet of paper containing nine lines with the letters d and p. They mark all cases where the letter d is flanked by two quotation marks, all other cases must be ignored. In the language free version, the letters are replaced by line-drawings of happy and unhappy faces. Participants mark the smiley faces flanked by two quotation marks while ignoring the distractors. Each of the nine lines comprises 47 faces (30 smiley faces; 20 correct answers on average) and participants are given 20 s per line to tag their choices. Once the 20 s are over, they stop and start with the next line. The number of correctly marked smiley faces is used as the critical attention score.
Text reading
Within the text reading task two expository texts with different levels of complexity were presented. The less complex text was about the planet Venus and consisted of a heading and 10 sentences (150 words in total). The second, more scientific text covered the topic of using plants as energy storage. It also had a heading but was only seven sentences long (134 words in total). The German adaptation of the Flesch Reading Ease Scores (Amstad, 1978) for both texts confirmed the differences in readability (easy text: 56, hard text: 43). Additionally, 20 native German students from the FU Berlin (Mage = 22.3, SDage = 3.9; 13 females) rated both text types on how readable and how interesting they were using five-point scales (1 = low and 5 = high). As assumed, Readability ratings for the text about the planet Venus were significantly higher than Readability ratings for the text about energy storage [Measytext = 4.8, SDeasytext = 0.6; Mdifficulttext = 4.0, SDdifficulttext = 0.7, t(19) = 3.92, p < 0.001, d = 1.16], whereas both text types were rated as equally interesting [Measytext = 3.9, SDeasytext = 0.9; Mdifficulttext = 3.7, SDdifficulttext = 1.2, t(19) < 1].
Participants were asked to read the texts silently while their eye movements were recorded. After each text they answered four comprehension questions. The two variables of interest were FFD as a measure of the difficulty of the word’s initial visual recognition as well as lexical access and TRT as a global measure of basic reading ability as well as general comprehension (Hyönä et al., 2003; Boston et al., 2008). Results are reported based at text level analyses, for additional word level analyses see Supplementary Table 3.
Sample
The sample consisted of 34 native German speaking participants (24 females) with a mean age of 28.6 years (SD = 8.92, range: 19–50) and a mean of 17.3 years (SD = 3.43) of formal education, who did not take part in the first study. Participants were recruited within the FU Berlin and outside the FU Berlin via announcements in local newspapers. Prior to testing, participants gave written informed consent. The mean SLS-Berlin score was 59.2 (SD = 10.7, range: 39–76), which corresponds to the mean from the age equivalent norm sample, i.e., that of younger adults (see Supplementary Table 3). All participants had normal or corrected-to-normal vision and none had a diagnosed reading or learning disability. They were paid 10 EUR for about 1 h of testing. The study was approved by the ethics committee of the FU Berlin.
Procedure
The SLS-Berlin was always administered first in each test session. In contrast to the norm sample, we recorded eye movements during the completion of the SLS-Berlin as well as during silent text reading. For this purpose, participants were seated in front of a 19-inch monitor (120-Hz refresh rate, 1024 × 768 resolution) at a viewing distance of approximately 65 cm. An EyeLink 1000 tower mounted eye tracker build by SR Research (Mississauga, ON, Canada2) was used to collect the eye movement data. Sampling frequency was 1,000 Hz. Only the right eye was tracked. Participants’ responses were collected with a standard keyboard, a forehead rest was used to minimize head movements. A nine-point calibration cycle at the beginning of each task (SLS-Berlin, easy text, hard text) and session was used to ensure a spatial resolution error of less than 0.5° of visual angle. All tasks were controlled with Experiment Builder software (RS Research, version 1.10.1630). Font type and size was the same in all tasks. The sentences within the SLS-Berlin and the two expository texts were written in Arial, font size 20, double spaced and left aligned. The procedure for the SLS-Berlin followed the routine described for the norm sample. Due to the font size, long sentences, especially from the second part of the test, were presented on three lines.
After finishing the SLS-Berlin, participants took a rest in which the experimenter described the text reading task. Participants were instructed to silently read each text at their own speed. They were also informed about a short memory test administered directly after reading. Each text presentation started with the calibration described above followed by a fixation cross presented for 500 ms to the left of the first word of the text. Both texts were presented on three pages, the end of a page always coincided with the end of a sentence. Participants proceeded to the next page by pressing the spacebar. At the end of the text reading task, participants were asked four comprehension questions. The easy text was always read first, followed by the more complex text.
The SLRT-II was conducted next with word reading preceding pseudoword reading. For both tasks, participants received a training list of 8 items to be read out loud and corrective feedback if necessary. During the test session, the experimenter noted the number of correctly read words and pseudowords within 1 min.
Finally, at the very end of the entire test session, participants conducted the Smiley-Test. The experimenter explained the task with the help of a short example, followed by a short training session. Then the 3-min test session started. Overall, the whole experimental session, including eye-tracking, took approximately 40 min.
Data Analysis
To obtain additional information about the validity of the SLS-Berlin score, we calculated Pearson’s-product-moment correlations between SLS-Berlin scores, the two measures for word and pseudoword identification and the results from the Smiley-Test.
Eye-tracking data was analyzed by first removing participants that had missing data for more than 50% of the words. This led to the exclusion of one subject. For the remaining 33 participants, we then removed fixations shorter than 50 ms or longer than 1000 ms as well as blinks and instances of track-loss. We also excluded data recorded for text headings, initial and final words. The latter was done since words in the beginning and the end of a page and a sentence are often processed differently (Rayner et al., 2000; Kuperman and Van Dyke, 2011). All words with a TRT of zero were handled as missing. This procedure left 4,388 data points for the easy text and 4,038 data points for the more complex text for which each participant’s mean FFD and TRT were finally calculated. Extreme values, i.e., values deviating more than 2.5 standard deviations from an individual’s mean FFD and TRT of each text were removed beforehand (data loss of less than 3.5% for both texts). To assess the predictive power of SLS-Berlin scores for both texts, we conducted linear, quadratic and cubic regression analyses following the same procedure described in Study 1. Due to the small sample size of 33 participants, we did not perform cross-validation.
Results
Comparing the SLS-Berlin With Measures of Word and Pseudoword Recognition as Well as Visual Attention and Processing Speed
On average, the participants read 122.0 words (SD = 14.0, range: 84–156) and 80.3 pseudowords (SD = 17.9, range: 46–134). Existing norms of the SLRT-II only cover slightly younger adults (high school and university students) and according to these norms, the average word reading score of the present sample corresponds to a percentile rank of 63 (range: 6–99), the average pseudoword reading score to one of 61 (range: 9–99). The Smiley-Test does not have any normative data. Our participants marked on average 249.3 smileys correctly (SD = 36.1, range: 175–319). After checking for normal distribution, we calculated Pearson’s product-moment correlations between SLS-Berlin scores and the two measurements for word and pseudoword identification as well as Smiley-Test scores. The highest correlation was observed between SLS-Berlin scores and pseudoword reading: rSLS–Berlin– SLRT–II pseudoword = 0.59, t(32) = 4.18, p < 0.001, 95% CI [0.32, 0.78], followed by the correlation with word reading: rSLS–Berlin– SLRT–II word = 0.51, t(32) = 3.38, p = 0.002, 95% CI [0.21, 0.73]. The correlation between the SLS-Berlin and the Smiley-Test was weaker and not significant: rSLS–Berlin –Smiley–Test = 0.23, t(32) = 1.32, p = 0.18, 95% CI [−0.11, 0.53]. We used the cocor module (Diedenhofen and Musch, 2015) to test whether the correlations between the SLS-Berlin and the two SLRT-II reading tasks were higher than the correlations between the SLS-Berlin and the Smiley-Test. The one-tailed z-test from Hittner et al. (2003) indicated that the correlation coefficient of the SLS-Berlin and pseudoword reading was significantly higher than that of the SLS-Berlin and the Smiley-Test (zSLS–Berlin– SLRT–II pseudowordvs.SLS–Berlin –Smiley–Test = 1.99, p = 0.02). The comparison of the correlation coefficients of the SLS-Berlin and word reading with the SLS-Berlin and the Smiley-Test failed to reach statistical significance, zSLS–Berlin– SLRT–II wordvs.SLS–Berlin –Smiley–Test = 1.46, p = 0.07.
Assessing the Predictive Power of the SLS-Berlin on Eye-Tracking Measures
To investigate the predictive power of the SLS-Berlin with respect to early and late processes related to reading proficiency, we analyzed the error rates for the comprehension questions and calculated FFD and TRT, respectively for both the easy and the more complex text. Mean error rates for the easy text (M = 15.4, SD = 15.1) were significantly lower [t(33) = −6.12, p < 0.0001, d = 1.15] compared to mean error rates for the difficult text (M = 40.4, SD = 26.8). Mean FFD (Measytext = 196.2, SDeasytext = 20.4; Mcomplextext = 204.8, SDcomplextext = 20.2) as well as mean TRT (Measytext = 293.4, SDeasytext = 54.5; Mcomplextext = 383.3, SDcomplextext = 78.1) were significantly longer while reading the difficult text compared to the easy text, tMeanFFD (32) = 5.56, p < 0.001, d = 0.40, tMeanTRT (32) = 8.81, p < 0.001, d = 1.33. The correlations between mean error rates and SLS-Berlin scores were not significant (both rs < −0.18, ts(32) < −1.01, ps > 0.06; 95% CIeasy text [−0.56, 0.06], 95% CIdifficulttext [−0.48, −0.17]). The following regression analyses (model parameters and summaries are reported in Table 9) indicated decreasing FFD and TRT with increasing SLS-Berlin scores for both text types. With the exception of TRT of the easy text, quadratic models showed best model fits (easy text: adj. R2 MeanFFD = 0.24; more complex text: adj. R2MeanFFD = 0.31; adj. R2MeanTRT = 0.24). These quadratic relationships indicated a strong relationship between the eye-tracking measures and the SLS-Berlin score for less skilled readers, i.e., for readers who score below the expected mean norm value of the younger age group. For skilled readers with SLS-Berlin scores above the age equivalent mean norm, the relationship is quite low or absent. For mean TRT of the easy text, we observed a negative linear relationship (adj. R2MeanTRT = 0.37) indicating a general decrease of TRT with increasing SLS-Berlin scores.
Table 9. Model parameters and summaries of linear and polynomial regression analyses of SLS-Berlin score predicting eyetracking measures collected for reading an easy and a difficult scientific text.
Discussion
In Study 2, we first compared the SLS-Berlin with two measures of basal reading fluency (single word and pseudoword recognition) and one measure of serial allocation of visual attention and discrimination as well as general processing speed. Regarding the multidimensional nature of reading comprehension (Fletcher, 2006), we assumed strong but not perfect positive relationships with the two SLRT-II measures and a significantly weaker correlation with the Smiley-Test. Our results are in line with these assumptions. Number of words and pseudowords correlated significantly with SLS-Berlin scores (r = 0.51 and r = 0.59). Moreover, both correlations coefficients were descriptively higher than the non-significant correlation between the SLS-Berlin scores and the results of the Smiley-Test. Taken together, these results indicate a mild, if any, influence of general processing speed on SLS-Berlin scores. The absence of a strong correlation between the SLS-Berlin and the Smiley-Test indicates that the problem of the well-documented phenomenon of age-related slowing (Salthouse, 1996, 2000) is not central for the SLS-Berlin. Age-related slowing would imply longer response times and therefore lower SLS-Berlin scores for older compared to younger adults independent of reading proficiency. Yet in line with the non-significant correlation between the SLS-Berlin and the Smiley-Test, observed age effects within the norm sample were relatively small and only significant within the older subsample.
That the SLS-Berlin measures foremost reading proficiency is further confirmed by the text based analyses of the eye-tracking data measured while reading an easy and a difficult text. All four regression analyses using the SLS-Berlin score as a predictor for mean FFD and mean TRT are highly significant and demonstrated that readers with lower SLS-Berlin scores exhibited longer mean FFDs and mean TRTs per word than readers with high SLS-Berlin scores. These results are in line with previous research, i.e., more and longer fixations and TRTs are reported for less skilled and dyslexic readers compared to skilled readers (Ashby et al., 2005; Thaler et al., 2009; Hawelka et al., 2015; Krieber et al., 2016). The additional word level analyses reported in the Supplementary Table 3 also emphasizes the predictive power of the SLS-Berlin. Here, we also observed longer reading times and less skipping for readers with lower SLS-Berlin scores compared to readers with higher scores. In line with Kuperman and Van Dyke (2011), error rates of responses to comprehension questions did not correlate significantly with measures of individual differences in reading proficiency.
The analyses of mean FFD and mean TRT also replicated the well-established effect of significantly longer FFDs and TRTs per word when reading a difficult text as when reading an easy one (Rayner et al., 2006). We therefore investigated separately for each text, how well SLS-Berlin scores relate to FFD and TRT. If the SLS-Berlin, as a sentence-based screening tool, is sensitive to both, basic subprocesses of reading and general comprehension processes necessary for meaning integration and situation model building, the SLS-Berlin score should best predict TRT in the easy text condition. Compared to FFD, which is usually interpreted as being sensitive to early and stronger automatized processes around lexical access (Boston et al., 2008), TRT also entails rereading and backward regressions which are commonly associated with difficulties in linguistic processing and text comprehension processes (Reichle et al., 1998; Engbert et al., 2005). While in the more complex text condition, these later processes may override effects of early and more basic processes (Rayner et al., 2006), TRT in the easy text condition should reflect both early and later processes. The results of our regression analyses using SLS-Berlin scores to predict FFD and TRT in text reading are fully in line with this assumption. We obtained the best model fit (adj. R2 = 0.37) for the model predicting mean TRT of the easy text. This is also the only regression model for which we observed a strong and positive relationship between SLS-Berlin scores and eye-tracking measures. Correspondingly, the word level analyses of TRT reported in the Supplementary Table 3 show a significant interaction between text type and SLS-Berlin score. The negative relationship between SLS-Berlin and TRT was slightly more pronounced in the easy compared to the difficult text (see Supplementary Figure S3).
The observed linear relationship in the easy text condition indicates the SLS-Berlin’s sensitivity to differentiate both within the scope of less skilled readers and within the scope of skilled readers. Nevertheless, all other models for which we observed quadratic relationships, suggest that SLS-Berlin scores are more sensitive for distinguishing within less skilled readers as it is within skilled readers. As depicted in Figure 4, for SLS-Berlin scores of 60 and above, the curves flatten. Taken together, comparing SLS-Berlin scores with measures sensitive to early basic processes of reading and measures also sensitive to later comprehension processes underscore the applicability of the SLS-Berlin as a well-suited screening tool to assess individual differences in reading proficiency. The SLS-Berlin seems to be very good in differentiating skilled from less skilled readers whereby its highest sensitivity seems to be within the scope of less skilled readers.
Figure 4. SLS-Berlin Scores predicting Mean First Fixation Duration (A) and Mean Total Reading Time (B) for reading an easy and a difficult text.
General Discussion
Within this article, we introduced a computerized screening test for adults to assess individual differences in reading proficiency. The test structure of the SLS-Berlin follows that of a well-established screening test for German children, the SLS (Mayringer and Wimmer, 2003; Auer et al., 2005; Wimmer and Mayringer, 2014) which measures basic reading ability by simultaneously evaluating reading comprehension and reading speed. Even though, a developmental study with German speaking first graders reading word lists suggests reading speed to differentiate fairly well between good and poor readers (Wimmer and Hummer, 1990), word-based measures of reading speed and/or accuracy underrepresent the construct of reading fluency (Cunningham, 2001; Kuhn and Stahl, 2003). As reading proficiency, in our eyes, comprises the successful integration of early and basic processes of word decoding and lexical access as well as higher-order processes of comprehension to establish local and global coherence, a sentence-based screening test, such as the SLS-Berlin should firstly be able to capture both of these essential processes and secondly, differentiate well between and within poor and good adult readers.
Norms and Validity of the SLS-Berlin
The norm study of the SLS-Berlin attested very high reliability (r = 0.95) for the entire range of the norm sample covering the ages of 16 to 88 years. Comparisons of SLS-Berlin scores with other measures of basic reading fluency and eye movement data during text reading supports the notion that the SLS-Berlin is indeed sensitive for both early word decoding and later comprehension processes during reading. The SLS-Berlin is therefore an excellent screening tool to assess reading proficiency across the life span. The screening is just as intended to pave the way for future studies exploring individual differences in reading as it can be used in educational settings. With a duration of around 7 min, the SLS-Berlin fills the gap for a short and easy to use screening tool for reading proficiency, particularly for adult readers of all ages, which is still missing for the German language. The computer-based application allows for easily realizable group testing which can be particularly important in settings that require a large number of participants or when environmental factors limit testing time, for example, in clinical settings. While for these specific instances, norms for patients still need to be established, the SLS-Berlin currently provides information which can be used to distinguish between general reading ability levels. More specifically, if a person yields an SLS-Berlin score well below the expected age specific norm, it may serve as an indication for further investigation.
Age Related Effects of the SLS-Berlin and Response Times
Although not being the focus of the present article, having collected data of 655 participants below the age of 60 and 1,493 participants being at least 60 years or older, gave us the possibility to further investigate written language processing across the life span in German adults. For this purpose, we analyzed the relationship of SLS-Berlin scores and response times separately for the subsamples of younger and older adults. With the exception of Complexity Ratings and Sentence ID, the order of significant sentence characteristics predicting SLS-Berlin scores by the amount of variance explained was similar for both age groups. However, both Sentence ID and Complexity Ratings explained the most variance in younger and older adults with younger adults relying more heavily on Sentence ID and older adults more on Complexity Ratings. This suggests that both, the younger as well as the older subsample draw in large parts on the same fundamental mechanisms while reading sentences, making the SLS-Berlin an ideal tool for investigating both younger and older readers.
However, there seems to be subtle differences with respect to age when looking at the nature of the relationship between SLS-Berlin scores and response times. As the regression functions differ for five of the seven significant predictors between the younger and older subsample, the nature of the relationship between predictors and sentence reading time seems to slightly change across the life span. Additionally, comparing the adj. R2 of the two subsamples, the biggest differences in explained variance were found for Number of Words (12% advantage for the older subsample) and Sentence ID (8% advantage for the younger subsample). This yields two important implications. First, in line with previous research regarding single word recognition in both English and German, older adults seem to have no deficits in identifying and assigning meaning to words (e.g., Ratcliff et al., 2004; Spaniol et al., 2006; Froehlich et al., 2016). In fact, it has been reported, that older compared to younger adults spend more time on identifying words (lexical access) to compensate for age related cognitive deficits, such as declines in working memory or processing speed (Stine-Morrow et al., 2008). Second, as the test progresses, the impact of Sentence ID on sentence response time becomes larger for the younger subsample than for the older subsample suggesting possible differences in reading strategies. While younger readers may initially emphasize speed, older readers possibly tend to read the sentences in a more constant speed. With the increasing complexity of the sentences younger adults are forced to read more thoroughly and thus slow down toward the end of the test. Comparing the differences in reading speed (measured in words per minute) and age group between the first and second half of the SLS-Berlin (sentences 1–38 vs. 39–77) seem to confirm this assumption. Reading speed of the younger subsample decelerates to a greater degree than that of the older subsample [younger adults: MSLS–Berlin1st –SLS–Berlin2nd = 28.2, SD = 27.2; older adults: MSLS–Berlin1st –SLS–Berlin2nd = 11.6, SD = 16.1], t(2136) = 17.6, p < 0.001, d = 0.82. However, we acknowledge that this is rather speculative. Given the sparsity of available data for German adults, further research on the mechanics of reading and possible changes in reading strategies across the life span is necessary.
Limitations and Outlook
The sentences for the SLS-Berlin were constructed to initiate normal reading in an ecologically valid setting while minimizing shallow processing and skim reading. The analyses of the relationships between sentence characteristics and response times of all 77 sentences indicated that this implementation was successful. Nevertheless, the slightly left-skewed distributions of the SLS-Berlin scores suggest better discriminability within poorer than within skilled to very skilled readers. With the intended application in mind, i.e., within reading research, educational and perhaps clinical settings, we are convinced, that the sensitivity of the SLS-Berlin within the lower performance range is adequate. An extension of the test sentences to differentiate more precisely within the upper performance spectrum would further improve the SLS-Berlin. However, prolonging the test would also lead to enhanced test times which would then restrict its applicability as a short and easy to use screening tool.
The chosen time frame of 3 min to calculate the final test score marks the SLS-Berlin as a somewhat speeded screening test. We understand that this is rather critical considering the intended application of the SLS-Berlin across the whole life span. Independently from looking at reading specific processes at word, sentence or text level, factors related to processing speed are among the biological and behavioral variables with the strongest association to age (Salthouse, 1996). Sensorimotor and cognitive slowing increases with increasing age, resulting in generally larger RTs for older adults compared to younger adults (e.g., Salthouse, 1996; Verhaeghen and Salthouse, 1997). To counteract possible disadvantages of especially older participants, the instruction of SLS-Berlin was changed compared to the children’s version. Neither is the 3-min cut-off mentioned nor are the participants interrupted while working on the 77 test sentences. Instead, in the computerized version participants read all sentences within a self-paced paradigm. This allows particularly older readers to use strategies to compensate for processing deficiencies such as greater reliance on discourse context and increased resource allocation (Stine-Morrow et al., 2008). Nevertheless, we estimated the possible influence of processing speed on the SLS-Berlin in Study 2, where we compared SLS-Berlin scores with a measure for general processing speed. The slightly positive yet non-significant correlation (r = 0.23, p = 0.18) indicated that our adjustments of the instruction were successful. Accordingly, age related differences in SLS-Berlin scores were relatively small and were restricted to the older subsample. Whether the observed decline in SLS-Berlin scores starting around an age of 60 is related to an increase in age related slowing or to changes in reading specific processes cannot be answered within the scope of the present study and should be subject of further research.
In a first step to resolve this issue, Froehlich et al. (2016) investigated the impact of aging on the subprocesses of reading by using hierarchical diffusion modeling. Hierarchical diffusion modeling bears the advantage to disentangle processes underlying decision tasks, such as stimulus encoding (a priori) decision making and processing efficiency, i.e., the speed of information uptake (Ratcliff, 1978; Vandekerckhove et al., 2011; Voss et al., 2013). With regard to the latter, results showed that younger adults outperformed older adults (> 60 years) in letter identification and phonological processing. However, with respect to orthographic and semantic processing, younger adults were slower in information uptake than high performing older adults. Yet, low performing older adults performed below the level of younger adults. These results are a firm indication, that besides age, proficiency impacts the degree to which a reader may rely on the different subprocesses of reading. The SLS-Berlin, as it is presented here, treats reading skill as a unitary construct. One important topic for further research is to explore the relationship between single subprocesses of reading and how they are reflected in the SLS-Berlin.
Conclusion
The very high Spearman-Brown corrected reliability (r = 0.95) and the results from Study 2 demonstrate that the SLS-Berlin is a valuable screening tool for the systematic exploration of individual differences in reading proficiency. The test differentiates very well between poor and good readers. The huge number and extensive age range of the norm sample with an emphasis on age groups above 60 ensures the test to be used for a variety of important (developmental) research questions. Further developments are planned to establish a parallel test version as well as increase the relatively small amount of norm data for the age groups between 40 and 60 years of age. However, we think, the SLS-Berlin’s short duration and easy administration makes it a highly potential tool for the investigation of reading abilities in not only underresearched populations such as older adults and individuals with brain lesions but also in developing and proficient readers.
Data Availability
The raw data supporting the conclusions of this manuscript will be made available by the authors, without undue reservation, to any qualified researcher. The norm data for the SLS-Berlin are part of the Supplementary Table 2. Both, the SLS-Berlin (available for all standard platforms such as Linux, macOS, and Windows as it is written in PsychoPy3; Peirce and MacAskill, 2019) as well as continuously updated norms are available at: https://www.ewi-psy.fu-berlin.de/einrichtungen/arbeitsbereiche/allgpsy/Download/index.html.
Ethics Statement
This study was approved by the Ethics Commitees of the Max Planck Institute for Human Development, Berlin and the Freie Universität Berlin, Department of Education and Psychology.
Author Contributions
JL, EF, and AJ contributed to the conception and design of the study. FH developed the test stimuli. JL and EF organized the database, performed the statistical analysis, and wrote the first draft of the manuscript. All authors contributed to the manuscript revision, and read and approved the submitted version.
Funding
This research was supported by the German Federal Ministry of Education and Research (Bundesministerium für Bildung und Forschung, BMBF; Grant Nos. FKZ 01GJ1010A and FKZ 01GJ1010B). Open Access Funding was provided by the Freie Universität Berlin.
Conflict of Interest Statement
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Acknowledgments
We would like to thank Prof. Dr. Ulman Lindenberger and the recruiting office of the Max Planck Institute for Human Development, Berlin, for recruiting the subjects and collecting the data. We are grateful to the student assistants of both, the Department of Experimental and Neurocognitive Psychology at the Freie Universität Berlin as well as the Max Planck Institute for Human Development, Berlin for their hours of tireless work.
Supplementary Material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2019.01682/full#supplementary-material
Footnotes
- ^ To assess the relationship between Sentence ID and all sentence characteristics, we used regression analyses as described for the analysis of the response time data in study 1 (see section “Study 1-Data Analysis”). Due to the relative small sample pool of 77 sentences, cross-validation was omitted.
- ^ http://sr-research.com/pdf/techspec.pdf
References
Adler-Grinberg, D., and Stark, L. (1978). Eye movements, scanpaths, and dyslexia. Am. J. Optom. Physiol. Opt. 55, 557–570. doi: 10.1097/00006324-197808000-00005
Allen, P. A., Madden, D. J., and Crozier, L. C. (1991). Adult age differences in letter- level and word-level processing. Psychol. Aging 6, 261–271. doi: 10.1037/0882-7974.6.2.261
Amstad, T. (1978). Wie Verständlich Sind Unsere Zeitungen? Ph.D. Dissertation, Studenten-Schreib-Service: Zürich.
Ashby, J., Rayner, K., and Clifton, C. (2005). Eye movements of highly skilled and average readers: differential effects of frequency and predictability. Q. J. Exp. Psychol. Section A 58, 1065–1086. doi: 10.1080/02724980443000476
Auer, M., Gruber, G., Mayringer, H., and Wimmer, H. (2005). SLS 5-8. Salzburger Lese-Screening für die Klassenstufen. Bern: Verlag Hans Huber. 5–8.
Bates, D., Maechler, M., Bolker, B., and Walker, S. (2015). Fitting linear mixed-effects models using lme4. J. Statist. Softw. 67, 1–48. doi: 10.18637/jss.v067.i01
Bergmann, J., and Wimmer, H. (2008). A dual-route perspective on poor reading in a regular orthography: eVidence from phonological and orthographic lexical decisions. Cogn. Neuropsychol. 25, 653–676. doi: 10.1080/02643290802221404
Bertram, L., Böckenhoff, A., Demuth, I., Düzel, S., Eckardt, R., Li, S. C., et al. (2014). Cohort profile: the berlin aging study II (BASE-II). Int. J. Epidemiol. 43, 703–712. doi: 10.1093/ije/dyt018
Borella, E., Ghisletta, P., and de Ribaupierre, A. (2011). Age differences in text processing: the role of working memory, inhibition, and processing speed. J. Gerontol. Series B Psychol. Sci. Soc. Sci. 66, 311–320. doi: 10.1093/geronb/gbr002
Boston, M. F., Hale, J., Kliegl, R., Patil, U., and Vasishth, S. (2008). Parsing costs as predictors of reading difficulty: an evaluation using the Potsdam Sentence Corpus. J. Eye Mov. Res. 2, 1–12. doi: 10.16910/jemr.2.1.1
Brickenkamp, R., Schmidt-Atzert, L., and Liepmann, D. (2010). d2-R – Aufmerksamkeits- und Konzentrationstest. Göttingen: Hogrefe.
Brown, J. A., Fishco, V. V., and Hanna, G. (1993). in Nelson–Denny Reading Test: Manual for Scoring and Interpretation, ed. G. H. Forms (Rolling Meadows, IL: Riverside Publishing).
Brown, W. (1910). Some experimental results in the correlation of mental abilities. Br. J. Psychol. 3, 296–322. doi: 10.1111/j.2044-8295.1910.tb00207.x
Carpenter, P. A., Miyake, A., and Just, M. A. (1995). Language comprehension: sentence and discourse processing. Annu. Rev. Psychol. 46, 91–120. doi: 10.1146/annurev.ps.46.020195.000515
Cohen-Shikora, E. R., and Balota, D. A. (2016). Visual word recognition across the adult lifespan. Psychol. Aging 31, 488–502. doi: 10.1037/pag0000100
Cook, R. D. (1977). Detection of influential observations in linear regression. Technometrics 19, 15–18. doi: 10.2307/1268249
Cunningham, J. W. (2001). The national reading panel report. Read. Res. Q. 36, 326–335. doi: 10.1598/RRQ.36.3.5
Daselaar, S., Veltman, D. J., Rombouts, S. A. R. B., Raaijmakers, J. G. W., and Jonker, C. (2003). Deep processing activates the medial temporal lobe in young but not in old adults. Neurobiol. Aging 24, 1005–1011. doi: 10.1016/s0197-4580(03)00032-0
De Beni, R., Borella, E., and Carretti, B. (2007). Reading comprehension in aging: the role of working memory and metacomprehension. Aging Neuropsychol. Cogn. 14, 189–212. doi: 10.1080/13825580500229213
DeDe, G., Caplan, D., Kemtes, K., and Waters, G. (2004). The Relationship Between Age, Verbal Working Memory, and Language Comprehension. Psychol. Aging 19, 601–616. doi: 10.1037/0882-7974.19.4.601
Demoulin, C., and Kolinsky, R. (2016). Does learning to read shape verbal working memory? Psychon. Bull. Rev. 23, 703–722. doi: 10.3758/s13423-015-0956-7
Diedenhofen, B., and Musch, J. (2015). cocor: a comprehensive solution for the statistical comparison of correlations. PloS One 10:e0121945. doi: 10.1371/journal.pone.0121945
Duffy, S. A. (1992). “Eye movements and complex comprehension processes,” in Eye Movements and Visual Cognition, ed. K. Rayner (New York: Springer), 462–471. doi: 10.1007/978-1-4612-2852-3_28
Engbert, R., Nuthmann, A., Richter, E. M., and Kliegl, R. (2005). SWIFT: a dynamical model of saccade generation during reading. Psychol. Rev. 112, 777–813. doi: 10.1037/0033-295X.112.4.777
Fletcher, J. M. (2006). Measuring reading comprehension. Sci. Stud. Read. 10, 323–330. doi: 10.1207/s1532799xssr1003_7
Flom, P. L., Supino, P. G., and Ross, N. P. (2012). “Constructing and evaluating self-report measures,” in Principles of Research Methodology, eds P. Supino and J. Borer (New York: Springer), 147–175. doi: 10.1007/978-1-4614-3360-6_8
Frick, A., Wahlin, T. B. R., Pachana, N. A., and Byrne, G. J. (2011). Relationships between the national adult reading test and memory. Neuropsychology 25, 397–403. doi: 10.1037/a0021988
Frith, U., Wimmer, H., and Landerl, K. (1998). Differences in phonological recoding in German- and English-speaking children. Sci. Stud. Read. 2, 31–54. doi: 10.1207/s1532799xssr0201-2
Froehlich, E., and Jacobs, A. M. (2016). Verändert sich Lesen im Alter? alterseffekte in der visuellen worterkennung. Lernen und Lernstörungen 5, 95–109. doi: 10.1024/2235-0977/a000130
Froehlich, E., Liebig, J., Morawetz, C., Ziegler, J. C., Braun, M., Heekeren, H. R., et al. (2018). Same same but different: processing words in the aging brain. Neuroscience 371, 75–95. doi: 10.1016/j.neuroscience.2017.11.042
Froehlich, E., Liebig, J., Ziegler, J. C., Braun, M., Lindenberger, U., Heekeren, H. R., and Jacobs, A. M. (2016). Drifting through basic subprocesses of reading: a hierarchical diffusion model analysis of age effects on visual word recognition. Front. Psychol. 7:1863. doi: 10.3389/fpsyg.2016.01863
Gerstorf, D., Bertram, L., Lindenberger, U., Pawelec, G., Demuth, I., Steinhagen-Thiessen, E., et al. (2016). Editorial. Gerontology 62, 311–315. doi: 10.1159/000441495
Graesser, A., Singer, M., and Trabasso, T. (1994). Constructing inferences during narrative comprehension. Psychol. Rev. 101, 371–395. doi: 10.1037/0033-295X.101.3.371
Grainger, J., and Jacobs, A. M. (1996). Orthographic processing in visual word recognition: a multiple read-out model. Psychol. Rev. 103, 518–565. doi: 10.1037//0033-295x.103.3.518
Hannon, B. (2012). Understanding the relative contributions of lower-level word processes. Read. Res. Q. 47, 125–152. doi: 10.1002/RRQ.013
Hannon, B. (2014). Are there gender differences in the cognitive components of adult reading comprehension? Learn. Individ. Differ. 32, 69–79. doi: 10.1016/j.lindif.2014.03.017
Hannon, B., and Frias, S. (2012). A new measure for assessing the contributions of higher level processes to language comprehension performance in preschoolers. J. Educ. Psychol. 104, 897–921. doi: 10.1037/a0029156
Hawelka, S., Gagl, B., and Wimmer, H. (2010). A dual-route perspective on eye movements of dyslexic readers. Cognition 115, 367–379. doi: 10.1016/j.cognition.2009.11.004
Hawelka, S., Schuster, S., Gagl, B., and Hutzler, F. (2015). On forward inferences of fast and slow readers. An eye movement study. Sci. Rep. 5:8432. doi: 10.1038/srep08432
Hedden, T., and Gabrieli, J. D. E. (2004). Insights into the ageing mind: a view from cognitive neuroscience. Nat. Rev.. Neurosci. 5, 87–96. doi: 10.1038/nrn1323
Hittner, J. B., May, K., and Silver, N. C. (2003). A monte carlo evaluation of tests for comparing dependent correlations. J. Gen. Psychol. 130, 149–168. doi: 10.1080/00221300309601282
Hofmann, M. J., and Jacobs, A. M. (2014). Interactive activation and competition models and semantic context: from behavioral to brain data. Neurosci. Biobehav. Rev. 46, 85–104. doi: 10.1016/j.neubiorev.2014.06.011
Hutzler, F., and Wimmer, H. (2004). Eye movements of dyslexic children when reading in a regular orthography. Brain Lang. 89, 235–242. doi: 10.1016/S0093-934X(03)00401-2
Hyönä, J., Lorch, R. F., and Rinck, M. (2003). “Eye movement measures to study global text processing,” in The Mind’s Eye, eds J. Hyönä, R. Radach, and H. Deubel (Amsterdam: Elsevier), 313–334. doi: 10.1016/B978-044451020-4/50018-9
Hyönä, J., and Olson, R. K. (1995). Eye fixation patterns among dyslexic and normal readers: effects of word length and word frequency. J. Exp. Psychol. Learn. Mem. Cogn. 21, 1430–1440. doi: 10.1037/0278-7393.21.6.1430
Jacobs, A. M. (2015a). Neurocognitive poetics: methods and models for investigating the neuronal and cognitive-affective bases of literature reception. Front. Hum. Neurosci. 9:186. doi: 10.3389/fnhum.2015.00186
Jacobs, A. M. (2015b). The scientific study of literary experience: sampling the state of the art. Sci. Study Lit. 5, 139–170. doi: 10.1075/ssol.5.2.01jac
Jacobs, A. M. (2017). “Affective and aesthetic processes in literary reading: a neurocognitive poetics perspective,” in Cognitive Literary Science: Dialogues Between Literature and Cognition, eds M. Burke and E. Troscianko (Oxford: Oxford Univ. Press), doi: 10.1093/acprof:oso/9780190496869.003.0016
Jacobs, A. M., and Grainger, J. (1994). Models of visual word recognition: sampling the state of the art. J. Exp. Psychol.: Hum. Percept. Perform. 20, 1311–1334. doi: 10.3758/s13428-012-0236-7
Jacobs, A. M., Rey, A., Ziegler, J. C., and Grainger, J. (1998). “MROM-p: an interactive activation, multipleread-out model of orthographic and phonological processes in visual word recognition,” in Localist Connectionist Approaches to Human Cognition, eds J. Grainger and A. M. Jacobs (Mahwah, NJ: Erlbaum), 147–188.
Jacobs, A. M., and Willems, R. (2018). The fictive brain: neurocognitive correlates of engagement in literature. Rev. Gen. Psychol. 22, 147–160. doi: 10.1037/gpr0000106
Jacobs, A. M., and Ziegler, J. C. (2015). “Visual word recognition, neurocognitive psychology of,” in International Encyclopedia of the Social & Behavioral Sciences, 2nd Edn, ed. J. D. Wright (Oxford,: Elsevier), 214–219. doi: 10.1016/B978-0-08-097086-8.57021-4
Junker, M. (2005). Der verflixte Akkusativ: Altersunterschiede und Altersinvarianz beim Verstehen von Sätzen mit unterschiedlich komplexer syntaktischer Struktur [Tricky accusative: age-related differences in comprehension of sentences with different syntactical structure]. Available at: https://publishup.uni-potsdam.de/opus4-ubp/frontdoor/deliver/index/docId/360/file/junker_m.pdf (accessed August 2, 2019).
Kemper, S., Crow, A., and Kemtes, K. (2004). Eye-fixation patterns of high- and low-span young and older adults: down the garden path and back again. Psychol. Aging 19, 157–170. doi: 10.1037/0882-7974.19.1.157
Kintsch, W. (1988). The role of knowledge in discourse comprehension: a construction-integration model. Psychol. Rev. 95, 163–182. doi: 10.1037/0033-295X.95.2.163
Kintsch, W. (1998). Modeling Comprehension Processes: The Construction-Integration Model. In W. Kintsch, Comprehension. A Paradigm for Cognition. Cambridge: Cambridge University Press, 93–120.
Kliegl, R., Grabner, E., Rolfs, M., and Engbert, R. (2004). Length, frequency, and predictability effects of words on eye movements in reading. Eur. J. Cogn. Psychol. 16, 262–284. doi: 10.1080/09541440340000213
Krieber, M., Bartl-Pokorny, K. D., Pokorny, F. B., Einspieler, C., Langmann, A., Körner, C., et al. (2016). The relation between reading skills and eye movement patterns in adolescent readers: evidence from a regular orthography. PloS One 11:e0145934. doi: 10.1371/journal.pone.0145934
Kuhn, M. R., and Stahl, S. A. (2003). Fluency: a review of developmental and remedial practices. J. Educ. Psychol. 95, 3–21. doi: 10.1037/0022-0663.95.1.3
Kuperman, V., and Van Dyke, J. A. (2011). Effects of individual differences in verbal skills on eye-movement patterns during sentence reading. J. Mem. Lang. 65, 42–73. doi: 10.1016/j.jml.2011.03.002
Landerl, K. (2000). Influences of orthographic consistency and reading instruction on the development of nonword reading skills. Eur. J. Psychol. Educ. 15, 239–257. doi: 10.1007/BF03173177
Landerl, K. (2001). Word recognition deficits in German: more evidence from a representative sample. Dyslexia 7, 183–196. doi: 10.1002/dys.199
Lenhard, W., Lenhard, A., and Schneider, W. (2018). ELFE II. Ein Leseverständnistest Für Erst- bis Siebtklässler – Version II. Göttingen: Hogrefe.
Lüdtke, J., and Jacobs, A. M. (2015). The emotion potential of simple sentences: additive or interactive effects of nouns and adjectives? Front. Psychol. 6:1137. doi: 10.3389/fpsyg.2015.01137
Lüdtke, J., Meyer-Sickendieck, B., and Jacobs, A. M. (2014). Immersing in the stillness of an early morning: testing the mood empathy hypothesis of poetry reception. Psychol. Aesthet., Creat. Arts 8, 363–377. doi: 10.1037/a0036826
Madden, D. J., Spaniol, J., Whiting, W. L., Bucur, B., Provenzale, J. M., Cabeza, R., et al. (2007). Adult age differences in the functional neuroanatomy of visual attention: a combined fMRI and DTI study. Neurobiol. Aging 28, 459–476. doi: 10.1016/j.neurobiolaging.2006.01.005
Marx, C., Hutzler, F., Schuster, S., and Hawelka, S. (2016). On the development of parafoveal preprocessing: evidence from the incremental boundary paradigm. Front. Psychol. 7:514. doi: 10.3389/fpsyg.2016.00514
Mayringer, H., and Wimmer, H. (2003). Salzburger Lese-Screening für die Klassenstufen 1-4. Bern. Switzerland: Huber.
McClelland, J. L., St John, M., and Taraban, R. (1989). Sentence comprehension: a parallel distributed processing approach. Lang. Cogn. Process. 4, SI287–SI335. doi: 10.1080/01690968908406371
Meyer, B. J. F., and Pollard, C. K. (2006). “Applied learning and aging: a closer look at reading,” in Handbook of the Psychology of Aging, eds J. E. Birren and K. W. Schaie (New York, NY: Academic Press), 233–260. doi: 10.1016/b9-78-012101-2/64950-0148
Moll, K., and Landerl, K. (2010). SLRT-II: Lese-und Rechtschreibtest; Weiterentwicklung des Salzburger Lese-und Rechtschreibtests. Bern: Huber.
Moore, M., and Gordon, P. C. (2015). Reading ability and print exposure: item response theory analysis of the author recognition test. Behav. Res. Methods 47, 1095–1109. doi: 10.3758/s13428-014-0534-3
Myers, J. L., and O’Brien, E. J. (1998). Assessing the discourse representation during reading. Discourse Process. 26, 131–157. doi: 10.1080/01638539809545042
Oganian, Y., Froehlich, E., Schlickeiser, U., Hofmann, M. J., Heekeren, H. R., and Jacobs, A. M. (2016). Slower perception followed by faster lexical decision in longer words: a diffusion model analysis. Front. Psychol. 5:1958. doi: 10.3389/fpsyg.2015.01958
Radach, R., and Kennedy, A. (2004). Theoretical perspectives on eye movements in reading: Past controversies, current issues, and an agenda for the future. Eur. J. Cogn. Psychol. 16, 3–26. doi: 10.1080/09541440340000295
Radvansky, G. A. (1999). Aging, memory, and comprehension. Curr. Dir. Psychol. Sci. 8, 49–53. doi: 10.1111/1467-8721.00012
Radvansky, G. A., and Copeland, D. E. (2001). Working memory and situation model updating. Mem. Cogn. 29, 1073–1080. doi: 10.3758/BF03206375
Radvansky, G. A., and Dijkstra, K. (2007). Aging and situation model processing. Psychon. Bull. Rev. 14, 1027–1042. doi: 10.3758/BF03193088
Ratcliff, R. (1978). A theory of memory retrieval. Psychol. Rev. 85, 59–108. doi: 10.1037/0033-295x.85.2.59
Ratcliff, R., Thapar, A., Gomez, P., and McKoon, G. (2004). A diffusion model analysis of the effects of aging in the lexical-decision task. Psychol. Aging 19, 278–289. doi: 10.1037/0882-7974.19.2.278
Rau, A. K., Moll, K., Snowling, M. J., and Landerl, K. (2015). Effects of orthographic consistency on eye movement behavior: German and english children and adults process the same words differently. J. Exp. Child Psychol. 130, 92–105. doi: 10.1016/j.jecp.2014.09.012
Rayner, K., Chace, K. H., Slattery, T. J., and Ashby, J. (2006). Eye movements as reflections of comprehension processes in reading. Sci. Stud. Read. 10, 241–255. doi: 10.1207/s1532799xssr1003-3
Rayner, K., Kambe, G., and Duffy, S. A. (2000). The effect of clause wrap-up on eye movements during reading. Q. J. Exp. Psychol. Hum. Exp. Psychol. 53, 1061–1080. doi: 10.1080/713755934
Rayner, K., and Reichle, E. D. (2010). Models of the reading process. Wiley Interdiscip. Rev.: Cogn. Sci. 1, 787–799. doi: 10.1002/wcs.68
Reichle, E. D., Pollatsek, A., Fisher, D. L., and Rayner, K. (1998). Toward a model of eye movement control in reading. Psychol. Rev. 105, 125–157. doi: 10.1037/0033-295X.105.1.125
Reichle, E. D., Rayner, K., and Pollatsek, A. (2003). The E-Z reader model of eye-movement control in reading: comparisons to other models. Behav. Brain Sci. 26, 445–476. doi: 10.1017/S0140525X03000104
Rönnlund, M., Nyberg, L., Bäckman, L., and Nilsson, L. G. (2005). Stability, growth, and decline in adult life span development of declarative memory: cross-sectional and longitudinal data from a population-based study. Psychol. Aging 20, 3–18. doi: 10.1037/0882-7974.20.1.3
Salthouse, T. A. (1996). The processing-speed theory of adult age differences in cognition. Psychol. Rev. 103, 403–428. doi: 10.1037//0033-295X.103.3.403
Salthouse, T. A. (2000). Aging and measures of processing speed. Biol. Psychol. 54, 35–54. doi: 10.1016/S0301-0511(00)00052-1
Schrank, F. A., McGrew, K. S., and Mather, N. (2014). Woodcock-Johnson IV Tests of Cognitive Abilities. Rolling Meadows, IL: Riverside.
Schroeder, S., Hyönä, J., and Liversedge, S. P. (2015). Developmental eye-tracking research in reading: introduction to the special issue. J. Cogn. Psychol. 27, 500–510. doi: 10.1080/20445911.2015.1046877
Singer, M., Graesser, A., and Trabasso, T. (1994). Minimal or global inference during reading. J. Mem. Lang. 33, 421–441. doi: 10.1006/jmla.1994.1020
Spaniol, J., Madden, D. J., and Voss, A. (2006). A diffusion model analysis of adult age differences in episodic and semantic long-term memory retrieval. J. Exp. Psychol. Learn. Mem. Cogn. 32, 101–117. doi: 10.1037/0278-7393.32.1.101
Spearman, C. (1910). Correlation calculated from faulty data. Br. J. Psychol. 3, 271–295. doi: 10.1111/j.2044-8295.1910.tb00206.x
Spieler, D. H., and Balota, D. A. (2000). Factors influencing word naming in younger and older adults. Psychol. Aging 15, 225–231. doi: 10.1037//0882-7974.15.2.225
Stine-Morrow, E. A. L., Miller, L. M. S., and Hertzog, C. (2006). Aging and self-regulated language processing. Psychol. Bull. 132, 582–606. doi: 10.1037/0033-2909.132.4.582
Stine-Morrow, E. A. L., Soederberg Miller, L. M., Gagne, D. D., and Hertzog, C. (2008). Self-regulated reading in adulthood. Psychol. Aging 23, 131–153. doi: 10.1037/0882-7974.23.1.131
Thaler, V., Urton, K., Heine, A., Hawelka, S., Engl, V., and Jacobs, A. M. (2009). Different behavioral and eye movement patterns of dyslexic readers with and without attentional deficits during single word reading. Neuropsychologia 47, 2436–2445. doi: 10.1016/j.neuropsychologia.2009.04.006
Torgesen, J. K., Wagner, R. K., and Rashotte, C. A. (2012). Test of Word Reading Efficiency - Second Edition (TOWRE-2). Austin, TX: Pro-Ed.
Tzeng, Y., Van Den Broek, P., Kendeou, P., and Lee, C. (2005). The computational implementation of the landscape model: modeling inferential processes and memory representations of text comprehension. Behav. Res. Methods 37, 277–286. doi: 10.3758/BF03192695
Vandekerckhove, J., Tuerlinckx, F., and Lee, M. D. (2011). Hierarchical diffusion models for two-choice response times. Psychol. Methods 16, 44–62. doi: 10.1037/a0021765
Verhaeghen, P., and Salthouse, T. A. (1997). Meta-analyses of age-cognition relations in adulthood: estimates of linear and nonlinear age effects and structural models. Psychol. Bull. 122, 231–248. doi: 10.1037/0033-2909.122.3.231
Voss, A., Nagler, M., and Lerche, V. (2013). Diffusion models in experimental psychology: a practical introduction. Exp. Psychol. 60, 385–402. doi: 10.1027/1618-3169/a000218
Willems, R., and Jacobs, A. M. (2016). Caring about dostoyevsky: the untapped potential of studying literature. Trends Cogn. Sci. 20, 243–245. doi: 10.1016/j.tics.2015.12.009
Wimmer, H., and Hummer, P. (1990). How German-speaking first graders read and spell: Doubts on the importance of the logographic stage. Appl. Psycholinguist. 11, 349–368. doi: 10.1017/S0142716400009620
Wimmer, H., and Mayringer, H. (2014). Salzburger Lese-Screening für die Schulstufen 2-9. Bern: Huber.
Wimmer, H., Schurz, M., Sturm, D., Richlan, F., Klackl, J., Kronbichler, M., et al. (2010). A dual-route perspective on poor reading in a regular orthography: an fMRI study. Cortex 46, 1284–1298. doi: 10.1016/j.cortex.2010.06.004
Wolf, M., and Katzir-Cohen, T. (2001). Reading fluency and its intervention. Sci. Stud. Read. 5, 211–239. doi: 10.1207/S1532799XSSR050-2
Ziegler, J. C., Castel, C., Pech-Georgel, C., George, F., Alario, F. X., and Perry, C. (2008). Developmental dyslexia and the dual route model of reading: simulating individual differences and subtypes. Cognition 107, 151–178. doi: 10.1016/j.cognition.2007.09.004
Ziegler, J. C., and Goswami, U. (2005). Reading acquisition, developmental dyslexia, and skilled reading across languages: a psycholinguistic grain size theory. Psychol. Bull. 131, 3–29. doi: 10.1037/0033-2909.131.1.3
Ziegler, J. C., Perry, C., Jacobs, A. M., and Braun, M. (2001). Identical words are read differently in different languages. Psychol. Sci. 12, 379–384. doi: 10.1111/1467-9280.00370
Keywords: reading proficiency, screening test, aging, life span, sentence reading, eyetracking, text comprehension
Citation: Lüdtke J, Froehlich E, Jacobs AM and Hutzler F (2019) The SLS-Berlin: Validation of a German Computer-Based Screening Test to Measure Reading Proficiency in Early and Late Adulthood. Front. Psychol. 10:1682. doi: 10.3389/fpsyg.2019.01682
Received: 28 April 2019; Accepted: 03 July 2019;
Published: 14 August 2019.
Edited by:
Jon Andoni Dunabeitia, Nebrija University, SpainReviewed by:
Aaron Veldre, The University of Sydney, AustraliaVictoria Anne McGowan, University of Leicester, United Kingdom
Copyright © 2019 Lüdtke, Froehlich, Jacobs and Hutzler. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Jana Lüdtke, jana.luedtke@fu-berlin.de
†Shared first authorship