Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 30 June 2023
Sec. Educational Psychology
This article is part of the Research Topic Fluency and Reading Comprehension in Typical Readers and Dyslexic Readers, volume II View all 17 articles

Reading-to-Writing Mediation model of higher-order literacy

  • 1Texas Institute for Measurement, Evaluation, and Statistics, University of Houston, Houston, TX, United States
  • 2College of Education and Behavioral Sciences, Houston Christian University, Houston, TX, United States
  • 3Psychological, Health, and Learning Sciences Department, University of Houston, Houston, TX, United States

Introduction: Writing difficulties frequently manifest comorbidly with reading challenges, and reading is implicated in particular acts of writing, such as reviewing and editing. Despite what is known, however, there remain significant barriers to understanding the nature of reading-writing relations, as few studies are comprehensive in the number and types of literacy skills evaluated. This study consists of a secondary data analysis of two studies employing structural equation modeling (SEM) to evaluate relations among reading and writing components skills independently, using the Direct and Inferential Mediation Model (DIME) of reading comprehension and Not-so-Simple View of Writing (NSVW) as theoretical frameworks.

Methods: We examine relations between reading and writing components from these models with a sample of upper elementary students with/at-risk for learning disabilities (n = 405). Lower-order components included word reading, vocabulary, handwriting and spelling. Higher-order components included background knowledge, reading strategies, inferencing, planning, editing, and revision. The literacy outcomes were oral and silent reading fluency, reading comprehension, and writing quality and productivity. We systematically build a Reading-to-Writing Mediation (RWM) model by first merging the DIME and NSVW components in a direct effects model (Aim 1), expanding the joint model to include reading and writing fluency (Aim 2), evaluating indirect effects between DIME and NSVW component skills (Aim 3), and finally, evaluating indirect effects with reading and writing fluency (Aim 4).

Results: The findings suggest that higher order fluency and comprehension skills are differentially related to writing activities and products.

Discussion: The pattern of results helps elucidate the mechanisms of how various reading and writing skills transfer and relate. The results have implications for targeted and implicit instruction in multicomponent interventions and the use of screeners to identify areas of risk.

Introduction

National data on student performance in the United States indicates that reading and writing (R-W) continue to be areas of concern, particularly for children with or at-risk for learning difficulties. Just over one-third of the nation’s students in grades 4, 8, and 12 demonstrated proficient reading comprehension in the latest National Assessment of Educational Progress (NAEP; McFarland et al., 2019). The NAEP Oral Reading Study (White et al., 2021) showed that reading fluency is also a concern for grade 4 students with reading difficulties; specifically, students who performed at the basic or below basic level on the NAEP reading assessment performed significantly lower on measures of passage fluency, accuracy, and expression in comparison to students at the proficient and advanced reading levels. Students at the lowest level of performance averaged just 71 words read correct per minute and 82% accuracy. Historically, an even lower percentage (~25%) of students have attained proficiency in writing (Aud et al., 2012). Beyond the K-12 setting, writing serves as a gatekeeper to college access for underrepresented students. Of note, 58% of employers rated recent graduates as not proficient in writing, and proficiency in written communication skills was considered essential by nearly 96% of employers, who report often considering writing skills when making decisions about hiring and promotions (National Association of Colleges and Employers, 2017). The National Commission on Writing (2004, 2005) estimated that $3.1 billion are spent annually remediating writing skills in the private sector and $250 million in the public sector.

Many individuals who have difficulty with reading also have challenges in the area of writing, highlighting the established connection across these skill areas; however, there is limited research on the relationships among R-W component skills. One group of students particularly likely to demonstrate lower performance in R-W are those with learning difficulties (LD; Fletcher et al., 2018). The most commonly occurring difficulties for students with LDs are word reading, fluency, and comprehension. Students with word-level reading difficulties, such as those with dyslexia, exhibit difficulties not only with handwriting and spelling but also demonstrate deficits with composition skills such as editing (e.g., Berninger et al., 2008; Carretti et al., 2016; Hebert et al., 2018). Difficulty with reading comprehension has been linked to difficulty with composition quality (Cragg and Nation, 2006; Re and Carretti, 2016). Given this reality, and the need to better understand the R-W connection in this group, the primary purpose of this study was to conduct an exhaustive examination of the associations among R-W skills to better inform research and practice. Theoretical and empirical accounts of literacy (described below) suggest that multiple skills contribute to the inter-connectedness of R-W. In an era when multiple interventions are available and easily accessible, it is important for researchers and practitioners alike to understand the complex patterns in which literacy skills interact with each other and how reading skills can be leveraged to explicitly teach skills in the academic domain of writing (and vice versa). Understanding the connections across these skill areas is critical given not only the opportunities for better conceptual or theoretical understanding of the relationships but also the potential for direction regarding the provision of literacy supports broadly.

Reading-to-Writing directionality

The instructional context in which R-W skills are taught can significantly impact the connection between reading and writing, as writing is both shaped and constrained by socio-cultural factors (Graham, 2018). Research has demonstrated that when R-W skills are integrated and taught together, rather than in isolation, students develop stronger R-W skills (Hebert et al., 2018). For example, writing-to-learn approaches emphasize using writing as a tool for understanding and learning new information through activities like summarizing or creating concept maps, which can help students organize and make sense of texts, which in turn can improve reading comprehension. R-W instruction can also be integrated through reading-to-write approaches, which emphasize using reading as a tool for developing writing skills. When students read a wide range of texts, they are exposed to a variety of text genres and structures, which can improve their own writing skills. We focus on the Reading-to-Writing directionality because the instructional context of this study was business-as-usual (i.e., reading instruction was more easily and frequently implemented than writing instruction; see Ahmed et al., 2022a). While we acknowledge that directionality of influence between R-W is not necessarily unidirectional, we emphasize the need for careful consideration of contextual factors, such as the nature of R-W instruction, orthography, and other relevant factors, in determining the directionality of the influence of one set of skills on another. In this study, we present an alternative Writing-to-Reading model in Supplementary Appendix B to acknowledge the potential for bidirectional influence, noting that the current study is limited in its capacity to establish causal connections because doing so necessitates an experimental research design. In the next sections, we start by presenting piecemeal evidence of R-W associations from experimental and correlational studies and end with component skills models that incorporate multiple skills and their interrelations.

Word-, sentence-, and text-level reading-writing

Robust relations between word reading and transcription skills (handwriting and spelling) have been demonstrated (e.g., Berninger et al., 2002; Abbott et al., 2010; Georgiou et al., 2020), and word-level literacy has been established as an important precursor of production and quality of writing. In addition, there is ample evidence for the relation between reading comprehension and writing skills, including word-level skills like spelling (Berninger et al., 2002) and text-level writing outcomes (e.g., Cragg and Nation, 2006; Carretti et al., 2016). There is little evidence, however, of the relationship between oral and silent reading fluency and the various levels of writing performance at the letter/word, sentence, and discourse levels.

Automaticity in R-W is a general issue affecting children with LDs, although little is known about whether rate-subtypes of disability can be reliably identified as separate subgroups of LD (Fletcher et al., 2018). Compromised accuracy and automaticity of word-level skills result in problems of automaticity at the sentence and discourse level of R-W fluency by reducing access to processes required for constructing meaning (e.g., inferencing or revision), as conscious attention to decoding or spelling makes R-W slow and laborious (Fletcher et al., 2018). Consequently, children with LDs are limited to proofreading texts for mechanics but not substance or content (MacArthur, 2016).

Oral reading fluency (ORF)—the ability to read aloud with speed, accuracy, and proper expression— is heavily used both in research and practice as an overall indicator of performance in reading because it is highly predictive of reading problems in children with LD (Deno, 2003). ORF is an overall indicator of reading performance in early elementary grades where the number of words read correctly in 1 min is the outcome observed. Measures of ORF are used as part of screening efforts in the context of multi-tiered systems of support (MTSS) where performance in reading is measured periodically and used to identify those students at-risk for poor performance as well as to monitor progress in response to instruction or intervention.

Limited studies investigate the relationship between reading fluency and writing —specifically, planning, translating, and revising—in typically developing and children with LDs (Graham and Hebert, 2011) despite the clear correlations of reading fluency with writing outcomes and the regular use of ORF in practice (e.g., Shinn et al., 1992; Fewster and MacMillan, 2002; Cragg and Nation, 2006; Berninger et al., 2008; Codding et al., 2015). There is support for the notion that the rate, accuracy, and prosody in ORF may relate to spelling at the word-level (Bear, 1991; Lefly and Pennington, 1991; Ritchey and Coker, 2014), although other researchers have found that ORF did not relate to spelling after controlling for other foundational reading skills (Morris et al., 2017). Two studies found that ORF was related to the total number of words written at the text-level for children in elementary grades (Ahmed et al., 2014; Tortorelli and Truckenmiller, 2023), but these studies did not explore the relations of ORF with sentence-level writing or text-level writing quality. Bear (1991) hypothesized that ORF may play a role in one writing process (planning) because word- and phrase/sentence-level planning are especially evident in oral expression (e.g., phrasal intonation and placement of accent in reading unfamiliar words). However, von Koss Torkildsen et al. (2016) found that ORF was not related to another writing process (revision) after controlling for executive function (working memory and attention) and spelling. To better understand individual differences in R-W, a more complete understanding of the role of ORF is necessary, when individuals engage in foundational writing skills at the word-level (e.g., spelling), self-monitoring during writing processes at the sentence-level (e.g., editing), and general writing outcomes at the text-level (quality and productivity).

Silent reading fluency (SRF) —the ability to read silently with speed and comprehension—emerges as a more important skill as students progress to higher grade levels, and ultimately adulthood, because SRF is required for more advanced texts. It is possible that SRF plays a more critical role in written expression when children are in the transitional phase from ORF to SRF and when the focus of instruction shifts from sentence- to text-level (Bear, 1991; Berninger et al., 2013; van den Boer et al., 2022). Notably, children with LD exhibit deficits in SRF that are commensurate with, or more pronounced than, their deficits in ORF (van den Boer et al., 2022). Research shows that SRF training results in better spelling for children with dyslexia (Berninger et al., 2013). Further, SRF contributes to children’s ability to revise sentences, and revision also impacts SRF (Ahmed et al., 2014). To our knowledge no studies have systematically examined the contributions of both ORF and SRF to higher-order writing processes (Shanahan, 2012). We propose that when evaluating sentence level R-W together with word- and text-level R-W, sentence level skills will have greater predictive power for text level writing quality and fluency than word level literacy.

Higher-order reading-writing connections

Reading skills are needed when individuals engage in self-monitoring during the planning, revision, and reviewing states of writing. That is, one’s ability to accurately and efficiently decode, scan, and comprehend what has been written are pre-requisite skills for revising the composition (McCutchen, 1996). Fitzgerald and Shanahan (2000) outline four areas of shared knowledge: (a) content or domain knowledge; (b) meta-knowledge about written language (i.e., functions and purposes); (c) pragmatic knowledge of text attributes (e.g., words, syntax, and usage); and (d) procedural knowledge for accessing information purposively, setting goals, analyzing, etc. For example, text is extended according to background knowledge and the writer’s hypotheses about the readers’ knowledge (Flower and Hayes, 1980), particularly in later grades, when students are required to write about topics outside of themselves (Davis and Winek, 1989). Knowledge of text structures help students understand the purpose for presenting information, the organization of ideas, and the use of similar schema across texts. A meta-analysis of 45 studies (Hebert et al., 2016), showed that text structure instruction (measured as strategies, such as evaluation of text) improved expository reading comprehension, particularly when including writing in that instruction. Furthermore, it has long been recognized that vocabulary plays a key role in writing development (Olinghouse and Wilson, 2013) with significant relationships evident for vocabulary to spelling and to planning before writing (Vanderberg and Swanson, 2006). Vocabulary knowledge is also related to individuals’ written production and text quality (e.g., Carretti et al., 2016; Kim and Schatschneider, 2017). Finally, planning involves goal-setting and knowledge mobilization, requiring students to evaluate their own knowledge of the topic, and narrow their topics and goals (Tierney and Shanahan, 1991).

Older theoretical models devoted solely to the interaction among R-W processes (Pearson and Tierney, 1984; Langer, 1986) and broader frameworks of writing in adults also specify various mechanisms of co-development (Bereiter and Scardamalia, 1987; Tierney and Shanahan, 1991; Fitzgerald and Shanahan, 2000; Deane et al., 2008). For example, inferencing allows writers to elaborate a new representation from a former one and is related to writing for children in first grade (Kim and Schatschneider, 2017) and in college (Connelly et al., 2006). Overall, higher-order reading skills (background knowledge, inferencing, strategies for reading) predict writing-specific processes such as planning, editing, and revising (e.g., Tierney and Pearson, 1983; Kirby et al., 1986; Singer and Bashir, 2004; Weston-Sementelli et al., 2018). These reading skills are also related to the quality of written composition (e.g., Fitzgerald and Shanahan, 2000; Decker et al., 2016; Kim and Schatschneider, 2017; Weston-Sementelli et al., 2018). The important conclusion from the theoretical literature is that higher-order reasoning processes of R-W are text-based (i.e., require interaction with text).

Component-skills models of reading and writing

To examine the above-mentioned R-W relationships, it is important to situate the study within the specific component models of R-W focused on in the present study. Although several models exist in both areas, there is significant overlap in the component skills represented in each. For the current study, we chose to frame our examination of the R-W relationship using the Direct and Inferential Mediation Model (DIME; Cromley and Azevedo, 2007) and the Not-so-Simple View of Writing (NSVW; Berninger and Winn, 2006) because they are well aligned with cognitive theories of reading and writing, respectively.

Direct and Inferential Mediation model of reading comprehension

The DIME model posits that the several components work together for the end goal of comprehension, and account for virtually all the variance in reading comprehension (Cromley and Azevedo, 2007; Ahmed et al., 2016). The elements of the DIME model include: (1) Decoding and (2) Vocabulary, because students who have adequate word reading skills and word knowledge can better understand text (Hoover and Gough, 1990), (3) Background knowledge, because readers who possess high levels of general knowledge perform better on reading comprehension and retain the information for longer periods of time (Chiesi et al., 1979; Kintsch, 1988), (4) Inferences (knowledge-to-text and text-to-text integration) are automatically generated when students understand what is implied by the text without explicitly being stated (Cain and Oakhill, 1999; Barnes and Dennis, 2001), and (5) Reading strategies, refers to engagement in cognitive and meta-cognitive strategies, such as summarizing, structuring, drawing conclusions, and evaluating text (O’Reilly and McNamara, 2007). The DIME model can be seen as an extension of the Simple View of Reading (SVR; Gough and Tunmer, 1986). While the SVR conceptualized comprehension as a product of word reading/decoding and oral language/linguistic comprehension, in the DIME model, the components of linguistic comprehension are further specified as lower (i.e., word reading, vocabulary) and higher-level (background knowledge, strategies for reading, inferencing) component skills; reading comprehension is thought to be influenced directly and indirectly via these skills. To enhance understanding of relations among the DIME components, four studies included silent reading fluency (SRF) or efficiency as an additional predictor (Smith, 2013; Oslund et al., 2016, 2018; Völlinger et al., 2018). In general, SRF was a strong predictor of reading comprehension for children in upper elementary or middle school, but vocabulary had the largest direct effect, followed by inferences. The relation of SRF and comprehension was dependent on reader proficiency.

Not-so-Simple View of Writing

An early component skill model of writing was the simple view of writing, which posited that transcription and ideation (i.e., text generation) together were necessary for writing (see Juel et al., 1986; Berninger et al., 2002). As a follow-up and extension, the NSVW holds that transcription skills (e.g., spelling and handwriting fluency), along with ideation, interact with higher-order, executive, and self-regulatory functions to produce writing through planning, composing, and revision. That is, proficient writers possess linguistic knowledge of grammar and syntax to create coherent and well-structured sentences and also engage in multiple cycles of reviewing, revising, and editing their work to improve their content, organization, and language use. Further, working memory is intrinsic and is responsible for storing and manipulating information needed during planning, composing, and revision processes. Our recent study (Ahmed et al., 2022a) using structural equation modeling (SEM) showed the NSVW can be deconstructed into key correlates (cognitive resources: self-efficacy and executive function), components (lower-order writing: handwriting and spelling; higher-order writing: planning, editing, and revising), and attributes of writing (productivity, quality, complexity, etc.) with multiple relations within and across the model. Similar to Ahmed et al. (2022a), the present study operationally defines the editing component of the NSVW as a broad construct that includes the knowledge of grammar, spelling, punctuation, and capitalization rules, as well as the ability to effectively apply these rules during the writing process. Likewise, in our study, we operationalized the concept of revision as a comprehensive construct that involves an understanding of syntax and structure, such as the development and organization of sentences and paragraphs, and the effective integration of this knowledge into the child’s written work. These definitions emphasize that the NSVW considers planning, editing, and revising as executive functions that necessitate the manipulation of information during the writing process beyond a mere understanding of grammar and syntax in oral language. The significance of these definitions lies in their recognition of the critical role played by higher-order cognitive processes in the writing process, such as the capacity to plan and organize ideas, pay attention to details, and revise written work for clarity and coherence. Consequently, the NSVW places equal emphasis on both declarative and procedural linguistic knowledge, highlighting the importance of not only understanding the rules of language but also applying them effectively in written expression.

Joint models of reading-writing

An early component skills model linking R-W development is the Simple View of Reading and Writing (SVRW; Juel et al., 1986), which specified common predictors of word recognition and spelling (e.g., lexical knowledge) but did not find support for connections among spelling and word recognition or among reading comprehension and writing. In the SVRW, oral language and IQ were exogenous factors which indirectly influenced spelling through their effect on phonemic awareness (i.e., the effect of oral language on spelling was completely mediated by phonemic awareness). More recently, Kim (2020) developed the Interactive Dynamic Literacy Model (IDL) and the Direct and Indirect Effects Model of Writing (DIEW; Kim and Graham, 2022). The premise of the IDL and DIEW is that several related, yet separate, systems support R-W and include oral language, knowledge, domain-general and higher-order cognition, and sociocultural systems. The IDL model is situated within a levels of language framework, including discourse (text reading fluency, text writing or composition fluency), sentence level (sentence comprehension and sentence writing fluency), lexical (word reading fluency, spelling fluency), and sub-lexical (phoneme-grapheme correspondence, transcription fluency) levels. Interactive relations are highlighted in both models with R-W skills developing interdependently within and across a hierarchy. For example, reading comprehension influences composition and the experience of generating compositions can enhance comprehension through promoting awareness of structure and meaning of text. Reading comprehension is also expected to vary as a function of dimension of written composition (e.g., writing quality, productivity, correctness in writing, syntax, story structure, etc.; Shanahan and Lomax, 1986).

The IDL and DIEW models are broad frameworks that build on older R-W models described above. The empirically tested versions of the IDL and DIEW models are narrower in the number and types of components included and the associations among them. The models do not specify direct relations among higher-order R-W (e.g., vocabulary or inferencing to written expression) because the tested models postulated a complete mediation of higher-order skills through their effects on oral language. In the DIME model described above, higher-order skills (vocabulary, background knowledge, inferencing, and strategies) are pre-requisite reading skills, and collectively replace the linguistic comprehension component of the Simple View of Reading. In contrast, in the empirically tested IDL and DIEW, oral language plays a central role. The same higher-order skills (vocabulary, background knowledge, inferencing, and strategies) are specified as pre-requisite oral language skills, such that higher-order skills only influence reading comprehension (and writing) indirectly through their effect on oral language. Additional research is needed to better understand the nature of the indirect and direct relations between higher-order R-W processes to provide critical information that informs interventions for students with LDs.

Current study

The purpose of this study is to examine the relationships between multiple component skills of R-W in a sample of children with LD, focusing on two areas of research that have received the least attention: (1) connections among higher-order component skills of R-W (e.g., inferencing and revision); (2) connections among reading fluency and higher-order R-W. Our central research question was: What are the direct (Aims 1 and 2) and mediated (Aims 3 and 4) relations among R-W skills in comprehensive component-skills models? With the above-mentioned theoretical models and existing evidence as our foundation for understanding component skills involved in the R-W connection, the following four aims guided this question:

Aim 1: To build and test a Reading-to-Writing Skills (RWS) model of literacy by joining the DIME and NSVW component skills (Model 1, Figure 1). As shown in Figure 1, this model evaluates direct effects only, with lower- and higher-order DIME skills on the left-hand side and lower- and higher-order NSVW skills on the right-hand side.

FIGURE 1
www.frontiersin.org

Figure 1. Reading-to-Writing Skills (RWS) Model without reading and writing fluency (Model 1). The component skills are grouped under (1) letter or word level, which encompasses code- and meaning-based skills mainly at the word level; (2) the sentence level, which encompasses meaning-making linguistics skills; and (3) general literacy outcomes at the discourse level. The component skills roughly correspond with the levels of languages specified in the figure because the granularity of a component skill may be dependent on a child’s ability (e.g., planning may consist of single words for some students and sentences for others), the nature of the task (e.g., editing may involve correcting words or sentences), or the nature of scoring (e.g., legibility of letters and words were both considered for scoring handwriting quality). Small-dashed lines are 8 paths from lower-order reading skills to writing skills; long-dashed lines are 8 paths from higher-order reading skills to writing skills; solid lines are 4 paths from reading outcomes to writing skills. Double headed arrows are correlations or variances. Small single-headed arrows are residual variances.

1.1. We hypothesized word- and text-level connections among word reading and transcription skill (spelling and handwriting), and reading comprehension and written expression (e.g., Juel et al., 1986; Berninger et al., 2002; Abbott et al., 2010). Although decoding contributes to writing quality and productivity (Connelly et al., 2006; Decker et al., 2016), we hypothesized that after controlling for higher-order skills, word reading would not relate to distal, higher-order writing processes or overall writing quality or productivity (Shanahan and Lomax, 1986; Morris et al., 2017), but vocabulary would significantly predict writing quality (e.g., Shanahan and Lomax, 1986; Olinghouse and Wilson, 2013; Allen et al., 2014; Carretti et al., 2016; von Koss Torkildsen et al., 2016; Neumann et al., 2020; Truckenmiller and Petscher, 2020).

1.2. We hypothesized word-sentence level connections of vocabulary with planning and editing (e.g., Vanderberg and Swanson, 2006) but not revision (von Koss Torkildsen et al., 2016).

1.3. We hypothesized sentence-text level connections among knowledge, inferencing, and reading strategies, and writing quality (Allen et al., 2014), along with direct connections among higher-order reading skills and composition processes gleaned from early theoretical models (e.g., Tierney and Pearson, 1983; Kucer, 1985): background knowledge with planning; reading strategies with planning, editing, and revision; and inferencing with revision. We hypothesized that inference would relate to revision rather than editing, which requires superficial changes to text compared to revision.

Aim 2: To expand the Reading-to-Writing Skills model to include connections with R-W fluency (e.g., ORF and writing productivity; Model 2, Figure 2). As shown in Figure 2, we include ORF and SRF as measures of reading fluency and we incorporate writing productivity as the measure of discourse-level writing fluency.1

FIGURE 2
www.frontiersin.org

Figure 2. Full Reading-to-Writing Skills (RWS) Model with reading and writing fluency (Model 2). Silent reading fluency was measured using a sentence-level task, but it is included under discourse level because it measures comprehension of connected text and serves as a proxy for silent reading of longer texts. Small-dashed lines are 10 paths from lower-order reading skills to writing skills; long-dashed lines are 8 paths from higherorder reading skills to writing skills; solid lines are 15 paths from reading outcomes to writing skills. Double headed arrows are correlations or variances. Small single-headed arrows are residual variances.

2.1. We hypothesized word- and text-level connections of vocabulary with writing productivity but not of word reading with productivity (von Koss Torkildsen et al., 2016). Furthermore, we hypothesized that for children with LDs ORF would relate to spelling (Bear, 1991; Lefly and Pennington, 1991), writing quality (Ritchey and Coker, 2014), and productivity (Ahmed et al., 2014; Tortorelli and Truckenmiller, 2023). We expected that after controlling for word-reading and ORF, SRF would not relate to word-level writing.

2.2. We hypothesized sentence-text level connections among reading fluency, writing processes (e.g., editing and revision). We expected that after controlling for ORF and SRF, reading comprehension would predict planning and writing quality but not productivity or self-regulatory processes of editing and revision, and that ORF would relate to planning (Bear, 1991) but not revision (von Koss Torkildsen et al., 2016), and SRF would relate to revision (Ahmed et al., 2014).

Aim 3: To build and test a Reading-to-Writing Mediation (RWM; Figure 3) model with multiple direct and indirect paths between DIME and NSVW component skills (Model 3). Figure 3 depicts the domain-specific direct and indirect effects specified by the DIME or NSVW models (cross-domain associations are omitted for illustration purposes).

FIGURE 3
www.frontiersin.org

Figure 3. Reading-to-Writing Mediation (RWM) model (with cross-domain associations omitted for illustration; Model 3). The word literacy factor was measured by indicators of word reading (WJ-III Letter Word Identification and TOWRE Sight Word Efficiency) and spelling (WJ-III Spelling and percent words spelled correctly). Paths D1-D13 are correlations, direct, and indirect effects from the DIME model. D4 was added to account for the relation between spelling and summary writing (i.e., reading strategies). Paths N1-N11 are direct and indirect effects from the NSVW model.

3.1. We hypothesized connections among reading skills such that reading comprehension would be related to vocabulary, knowledge, and inferencing, but reading strategies would not be a significant predictor of reading comprehension (Ahmed et al., 2016). Word literacy (i.e., word decoding and encoding) would relate to reading comprehension but also to reading strategies because this factor was measured using writing-for-reading tasks (i.e., summarizing), which are related to spelling (Bahr et al., 2020).

3.2. We hypothesized connections among writing skills such that word literacy would relate to handwriting quality and editing, but not distal, higher-order writing skills (planning and revision) and that direct and indirect effects among handwriting, planning, editing, revision, and writing quality would remain significant in the mediation model.

3.3. We hypothesized cross-domain connections among higher-order reading (vocabulary, knowledge, inferencing, and strategies) and higher-order writing (planning, editing, revision) as well as reading comprehension and written expression. Figure 4 shows a mediation model in which (a) higher-order reading mediated the relations of basic literacy and knowledge with higher-order writing (e.g., vocabulary ➔ inferencing ➔ revision), (b) higher-order writing mediated the relations of higher-order reading and writing quality (e.g., reading strategies ➔ editing ➔ writing quality), and (c) higher order writing mediated the relations of basic literacy and knowledge with writing quality (e.g., background knowledge ➔ planning ➔ writing quality).

FIGURE 4
www.frontiersin.org

Figure 4. Reading-to-Writing Mediation (RWM) model (with cross-domain associations included; Model 3). Solid lines = within domain associations. Dashed lines = paths C1-C16 are cross-domain associations among reading and writing skills. Gray dashed lines were tested but omitted from the final model; black dashed lines were included in the final Model 3.

Aim 4: To expand the Reading-to-Writing Mediation model to include direct and indirect connections with R-W fluency (e.g., ORF, writing productivity; Model 4). Figure 5 depicts the domain-specific direct and indirect effects after including oral and silent reading fluency and writing productivity (cross-domain associations are omitted for illustration purposes).

FIGURE 5
www.frontiersin.org

Figure 5. Full Reading-to-Writing Mediation (RWM) model (with cross-domain associations omitted for illustration; Model 4). Paths F1-F12 are paths to/from oral or silent reading Fluency; paths P1-P5 are paths to/from writing Productivity (fluency).

4.1. We hypothesized connections among reading skills such that the direct effects of basic literacy and knowledge on reading comprehension would no longer be significant, but indirect effects would be found via oral and silent reading fluency. We hypothesized that connections among higher-order DIME skills would also change after controlling for SRF and ORF.

4.2. We hypothesized connections among writing skills such the direct and indirect effects of writing skills would be significant for writing productivity and would remain significant for writing quality.

4.3. We hypothesized cross-domain connections would change as a function of R-W fluency. Figure 6 shows a mediation model in which higher-order reading mediated the relations of ORF and SRF with writing skills, and higher-order writing mediated the relations of ORF and SRF and writing quality because we hypothesized that fluent reading foments deeper cognitive processing (e.g., reading comprehension and revision), which in turn influence writing quality and productivity (e.g., silent reading fluency ➔ inferencing ➔ revision; oral reading fluency ➔ editing ➔ writing quality). In general, we expected SRF to mediate the effects of ORF on other R-W skills, noting that the literature on indirect effects is limited.

FIGURE 6
www.frontiersin.org

Figure 6. Full Reading-to-Writing Mediation (RWM) model (with cross-domain associations included; Model 4). Within-domain variables and paths are the same as Figure 5 but were rearranged to include cross-domain paths. Long-dashed lines = paths C17-C28 are cross domain associations to/from oral and silent fluency and writing productivity. Solid lines = within domain associations. Small-dashed lines = the model did not converge with the inclusion of these paths.

Method

Participants and procedures

Data were collected as part of a larger RCT of an after-school reading intervention (see Roberts et al., 2018) with a 2×2 factorial treatment design with 2 levels of reading intervention (foundational reading skills and text-processing or text-processing only) and 2 levels of modality of small group instruction (writing or self-regulation). All intervention conditions included individualized computer-based instruction and small-group instruction in the first and second phase, respectively, of each instructional session. However, the present study includes data from pre-test only (i.e., the interventions took place after the collection of the baseline battery of measures included in the present study and did not impact performance on any tests). Thus, in this study, we do not differentiate among experimental condition because no differences were apparent in the groups randomized to treatment and the business-as-usual control condition on the assessments or on demographic variables. The sample for the present study consisted of 405 children in Grades 3–5 who were identified as struggling readers using the 25th percentile cutoff on the Test of Silent Reading Efficiency and Comprehension (TOSREC). These students also struggled with written expression as evidenced by low scores on the Test of Written Language (TOWL). As Table 1 indicates, the average age of the sample was 10 and ranged from 6 to 12 years old. Most of the sample was economically disadvantaged (69% free/reduced lunch) and 20% were in special education and/or had limited English proficiency. The majority of the sample was White (52%), followed by Black (41%), multiple races (22%), American Indian or Alaskan Native (1.5%), and Asian (0.5%).

TABLE 1
www.frontiersin.org

Table 1. Demographic characteristics by grade level.

Measures

Reading measures

Word reading

Word reading was assessed using the Woodcock Johnson III Letter Word Identification (WJ-LWID; Woodcock et al., 2001) and the Sight Word Efficiency (SWE) subtest of the Test of Word Reading Efficiency (TOWRE; Torgesen et al., 1999), with both measures demonstrating adequate reliability (α = 0.91 and α = 0.90–0.93, respectively).

Vocabulary

The verbal knowledge subtest of the Kaufman Brief Intelligence Test (KBIT-2; Kaufman, 2004) was used to measure picture vocabulary. The students are required to point to a picture that shows the meaning of a word or provides the answer to a question. Reliability is adequate for verbal knowledge subtest (α = 0.86–93).

Inferencing

The Bridge-It (Pike et al., 2010) test of inferencing measures the ability to integrate information presented in a statement sentence and a continuation sentence. Students are asked to read 4 sentences, one of which is the statement sentence, and a continuation sentence, which can either be a correct continuation (i.e., consistent with the situation model) or an incorrect continuation (i.e., inconsistent with the situation model). The statement sentence and the continuation sentence were separated by 3 sentences in the far condition and were adjacent in the near condition. This measure has adequate reliability (α = 0.73).

Background knowledge

The Assessment of Writing, Self-Monitoring and Reading (AWSM Reading; Gioia et al., 2023) is a paper-and-pencil experimental test developed for the larger study. Background knowledge items were tied directly to three passages that students read for comprehension and were not tied to the topic the students wrote about. The background knowledge items (e.g., What is found inside Yellowstone National Park?) were presented prior to reading the passages. A composite score of the knowledge items was used in this study (α = 0.62–0.69).

Reading strategies

Items from the AWSM Reader were also used to form a latent variable for strategies. Students read passages and provided short summaries of the passages as a performance measure of reading strategies. Reliability was high for summary writing (κ = 0.92–0.97) in this sample. The strategies factor also included a self-report measure of contextualized learning, Student Contextual Learning Scale (SCL; Cirino, 2012). The Strategies sub-scale of the SCL asks students to rate their beliefs, attitudes, and habits related to reading and learning strategies, with adequate reliability (α = 0.71–0.82) in our sample.

Reading comprehension

The Gates–MacGinitie Reading Test (GMRT; MacGinitie et al., 2000) was used to measure reading comprehension. The GMRT requires students to read short passages and answer multiple choice questions. The test has adequate reliability (Kuder–Richardson 20 [K-R 20] = 0.93–0.94 for grades 3–5).

Oral reading fluency

Two forms of the AIMSweb Oral Reading Fluency (Shinn and Shinn, 2002) were administered. Students were asked to read appropriate grad-level passages and the number of words read correctly in 1 min were recorded. AIMSweb reports adequate alternate-forms reliability (α = 0.80–0.81).

Silent reading fluency

Two forms of the Test of Silent Reading Efficiency and Comprehension (TOSREC; Wagner et al., 2010) were used. The TOSREC requires students to read sentences and verify the veracity of sentences (e.g., Do birds fly?). Alternate-forms reliability is high (0.86–0.93) in grades 3–5.

Writing measures

Spelling

The Woodcock Johnson III (WJ-III; Mather et al., 2001) Spelling subtest required students to spell phonetically regular (e.g., under) and irregular (e.g., beautiful) words. Reliability is high for grades 3–5 (α = 0.93). Spelling was also measured by counting the percent of total words spelled correctly on the TOWL Story Composition subtest (described below), with high inter-rater reliability (κ = 0.98).

Handwriting

Scores were derived using the presentation domain (and handwriting subdomain) of the 6 + 1 traits rubric (described below). Inter-rater reliability was high (κ = 0.91).

Planning

Students were given 5 min to plan their TOWL Story Composition responses following TOWL administration guidelines (Hammill and Larsen, 2009). Because the TOWL does not include a separate rubric for scoring planning, we adapted a planning rubric from Olinghouse and Graham (2009), which consists of scores ranging from 1 (little or no planning) to 5 (detailed story elements). Inter-rater reliability was κ = 0.75.

Editing and revision

Items for the editing and revision measures were derived from the TOWL Contextual Conventions subscale. Editing items required knowledge of mechanics and Revising items required knowledge of writing elements that enhance meaning. Two independent raters classified the 21 items into two categories (editing or revision), with perfect agreement (κ = 1.00). The editing category included items related to grammar, capitalization, spelling, and punctuation. The revision category included items related to content, structure, syntax and organization (see Supplementary Appendix C for the list of items coded as editing or revision). Internal consistency in this sample was α = 0.62 for Editing and α = 0.72 for Revision. As additional evidence for internal validity, we present a factor model for the TOWL Editing and Revision sub-scales as supplementary analyses in Supplementary Appendix C. Further, external validity was established with the State of Texas Assessments of Academic Readiness (STAAR; Texas Education Agency, 2016) Editing and Revising subtests with a sub-sample of fourth grade students (n = 73) from the present study for whom data were available on the state-wide assessment. The STAAR test was administered in the semester following the administration of the TOWL. Data were only available for a subsample of students because the STAAR high stakes writing assessment is not administered in grades 3 or 5 in Texas and because the STAAR Writing data were obtained for a smaller project (see Reid et al., in press). The STAAR Editing and Revision subtests require reading grade-level compositions embedded with errors and answering multiple-choice questions to identify and/or correct the errors in the text (see Reid et al., in press, for additional details). The STAAR and TOWL Editing sub-scales (r = 0.51, p < 0.001) and the STAAR and TOWL Revision sub-scales (r = 0.43, p < 0.001) were moderately correlated.

Writing quality

The Story Composition subtest of the Test of Written Language – 4th Edition (TOWL; Hammill and Larsen, 2009) requires students to write a story in response to a picture prompt in 15 min. We used the TOWL scoring guidelines to obtain the Story Composition score, which is scored on criteria such as plot (storyline), if characters show feelings/emotions, and story action or energy level. In addition, we used the 6 + 1 traits rubric (Culham, 2003) to score the essays. The 6 + 1 traits rubric includes the following domains: (1) Ideas: whether the essay is focused and clearly communicates ideas; (2) Organization: if the logical structure makes ideas easy to follow; (3) Voice: whether the author writes in an engaging manner; (4) Word Choice: relates to how the student’s choice of words creates a clear vision for the reader; (5) Sentence Fluency: how the author uses sentences and phrases to communicate; and (6) Conventions: errors related to punctuation, spelling, capitalization, grammar/usage. Inter-rater reliability ranged from κ = 0.80 for ideas to κ = 0.91 for word choice.

Writing productivity

The total words written were obtained for the TOWL Story Composition responses, as were correct minus incorrect word sequences (CIWS). CIWS is a curriculum-based measure of written grammar and mechanics (Espin et al., 2008). If two adjacent words are correctly spelled, capitalized, and punctuated that bigram results in a correct word sequence; otherwise, the bigram results in an incorrect word sequence. CIWS is calculated as the correct sequences minus any incorrect word sequences. Inter-rater reliability was κ = 0.995 for TWW and κ = 0.98 for CIWS.

Analytic approach

The present study consists of a secondary data analysis employing SEM to separately evaluate relations among reading components using the DIME model of reading comprehension as a theoretical framework (Ahmed et al., 2022b) and writing components using the NSVW as the theoretical framework (Ahmed et al., 2022a). We evaluate the relations between R-W components from these models, as shown in Figures 1, 2, 4, 6. The lack of empirical support for a path may reflect that research is lacking in a specific area, rather than support for a null relationship. Therefore, several paths were evaluated that have theoretical support but little empirical support (e.g., reading fluency and writing). Paths were not estimated if they were not significant in prior studies and there was no support in the theoretical literature (e.g., inference to handwriting and inference to spelling; Kim, 2020).

The SEM models were fit using Full Information Maximum Likelihood (FIML) in M-plus 8.6 (Muthen and Muthen, 1998–2017) to handle missing data (in the current sample, covariance coverage ranged from 0.81 to 0.99). Multiple criteria were considered to evaluate a model fit function (i.e., the extent to which the model fits the data) given a specific estimation method. Absolute model fit was evaluated using Akaike Information Criteria (AIC), Bayesian Information Criteria (BIC) and sample-size adjusted BIC, which take sample size, model fit, and number of parameters into account, with lower values reflecting a better fit. The root mean square error of approximation (RMSEA) compensates for model complexity and standardized root mean square residual (SRMR) is the standardized difference between the observed and predicted correlations. RMSEA and SRMR values ≤0.05 indicate an adequate fit. The comparative fit index (CFI) and Tucker-Lewis index (TLI) are incremental indices that compare the fit of the hypothesized model with a more restricted, baseline model (i.e., a model in which all observed variables are uncorrelated). CFI and TLI values ≥0.95 indicate a good fit and values ≥0.90 indicate an acceptable fit (Hu and Bentler, 1999).

Aims 1 and 2 (direct effects models)

Aim 1 examined the direct effects of DIME to NSVW component skills shown in Figure 1 (Model 1). Aim 2 examined the direct effects of R-W skills with three nested models that systematically incorporated R-W fluency: a full2 R-to-W Skills model depicted in Figure 2, and two nested models in which relations of each fluency skill (ORF or SRF) with writing skills were estimated independently of the other fluency skill. For example, the ORF model excluded any hypothesized relations of SRF with writing skills. The SRF model included all paths from the full model (Figure 2) but excluded any relations of ORF with writing skills. Nested models were compared with chi-square difference tests.

Aims 3 and 4 (mediation models)

The RWM models explored indirect effects of basic literacy (i.e., decoding and encoding) and knowledge (i.e., word and world knowledge) on writing skills via the indirect effects on reading skills. In addition, several indirect effects were evaluated within the reading domain (e.g., vocabulary ➔ inference ➔ reading comprehension) and writing domain (e.g., handwriting ➔ editing ➔ writing quality). The measurement models were similar to the RWS models of Aims 1–2, except that measures of word reading and spelling loaded on a single factor (Mehta et al., 2005). Consequently, in the RWM models, the word literacy factor predicted multiple R-W skills. As a first step, the RWM models included associations among DIME skills (paths D1-D14 in Figure 3) and NSVW skills (paths N1-N11 in Figure 3). We then evaluated the cross-domain associations shown in Figure 4 (paths C1-C17) by testing competing structural models. Due to the specification of a word literacy factor, these cross-domain associations are the 17 paths from Model 1 (Figure 1) which did not involve handwriting or spelling. The RWM model without fluency (Model 3) was generated by trimming paths without strong empirical support (i.e., gray dashed arrows in Figure 4; Mulaik and Millsap, 2000). The full RWM model with fluency (Model 4) retained the same variables and paths from the trimmed Model 3 and incorporated three R-W fluency variables (ORF, SRF, and writing productivity; Figure 5). Several direct and indirect effects were evaluated within the reading domain (e.g., ORF ➔ SRF ➔ inference ➔ reading comprehension; paths F1-F12 in Figure 5) and writing domain (e.g., handwriting ➔ productivity ➔ quality; paths P1-P5 in Figure 5) based on the literature reviewed. Cross-domain associations from Model 3 were retained (including ten cross-domain paths C2, C8-C12, C15-C17 which were not trimmed), and additional cross-domain associations were evaluated with R-W fluency skills (paths C18-C29 in Figure 6). The additional cross-domain associations are the 12 paths from Model 2 (Figure 2) originating from ORF or SRF, or going into writing productivity (with the exception of ORF ➔ spelling because in the RWM models spelling was combined with word reading in the word literacy factor). All indirect effects were estimated under FIML in Mplus and bias-corrected bootstrapped confidence intervals were obtained based on 1000–3000 bootstrap samples.

Alternative direct effects models

For space considerations the diagrams and results for the alternative models are presented in the Appendices. First, the Reading-to-Writing Domains model explored how reading skills differentially relate to writing dimensions depending on the skills assessed (Shanahan and Lomax, 1986; Kim and Graham, 2022). For example, ORF may predict sentence fluency because this dimension taps into the ability to use varied sentence structures that invite expressive oral reading. Similarly, ORF may predict scores on the voice dimension because this dimension taps into the ability to address the reader in an engaging way. In the present study, we measured six dimensions using the 6 + 1 traits rubric (Culham, 2003). In our approach to examining the R-W relationship, we also use correct minus incorrect word sequences (CIWS) as an overall indicator of writing. As a production-dependent metric, CIWS captures the amount of written text a student produces but also captures writing quality through consideration of spelling, grammar, and punctuation of adjacent words in the context of a sentence. Three R-to-W Dimensions models were evaluated: a full model depicted in Supplementary Appendix A and two reduced models for ORF and SRF, respectively. The R-to-W Dimensions Model included a general factor for writing which reflects the common variance across specific writing dimensions. Thus, the path from a specific reading skill to a specific dimension can be interpreted as a one unit change in the writing dimension as a function of the reading skill after controlling for (a) other reading skills and (b) for variance shared with other dimensions. Second, we evaluated a Writing-to-Reading Model because it is possible that the opposite directionality could fit the data equally well (i.e., due to model equivalence). Like the R-to-W models, paths for the W-to-R models were specified based on prior literature. For example, if there were no theoretical, experimental, or correlational studies surmising that better planning influences vocabulary then this path was omitted from the model. The W-to-R model (see Supplementary Appendix B) specified that: (a) handwriting and spelling predicted word reading; (b) spelling also predicted vocabulary, knowledge, ORF, and SRF; (c) higher-order writing (planning, editing, and revision) predicted reading strategies and inferencing; (d) planning predicted reading comprehension (e) editing and revision predicted reading comprehension, ORF, and SRF; (f) writing productivity predicted reading comprehension, ORF, and SRF; and (g) writing quality predicted all reading skills.

Results

Data were first screened for assumptions of normality and outliers, defined as data points with studentized residuals ±3 and high leverage. As the outliers did not represent minor or major reliability concerns (e.g., equipment failure), and the inclusion of the outliers did not change the results substantively, these data points were retained for the final analyses. As shown in Table 2, all assumptions of univariate normality were supported. The higher kurtosis on the AWSM Reader Summary 3 (5.43) is due to this passage’s higher text difficulty (readability) in comparison to summary 1 and 2 (Gioia et al., 2023).

TABLE 2
www.frontiersin.org

Table 2. Descriptive statistics.

Reading-to-Writing Skills models

The reduced model depicted in Figure 12 (108) = 215.96, p < 0.001; RMSEA [90% CI] = 0.05 [0.04, 0.06]; CFI = 0.96; TLI = 0.93; SRMR = 0.04) and the full model depicted in Figure 22 (177) = 294.58, p < 0.001; RMSEA [90% CI] = 0.04 [0.03, 0.05]; CFI = 0.97; TLI = 0.95; SRMR = 0.04) provided a good fit to the data. These models explained 66–68% of variance in writing quality, 57–63% in spelling, 41–42% in editing, and a smaller percentage of variance in revision (22%), handwriting (12–16%), and planning (10%). The full RWS model also explained 16% variance in writing productivity (see Table 3).

TABLE 3
www.frontiersin.org

Table 3. Variance explained in the direct effects models (RWS), mediation models (RWM), and alternative models.

All measures loaded significantly on their hypothesized factors (see Table 4). For the reading strategies factor, the loading of the self-report measure was smaller in magnitude (λ=0.13–0.14, p < 0.05) because all other loadings on this factor were from performance measures of strategies (summarizing). The correlations among reading variables in the full model are reported in Table 5, and correlations among residuals of the writing variables are reported in Table 6. Most reading variables were moderately to highly correlated, ranging from 0.25 for word reading and vocabulary (and reading strategies and vocabulary) to 0.90 for ORF and decoding. The largest residual correlation for the writing variables was between revision and writing quality (r = 0.63), and the smallest correlation was between spelling and total words written (r = 0.03; see Table 6). In addition, the disturbances of the WJ spelling and word reading subtests were allowed to correlate in all models because both subtests belong to the same family of tests (r = 0.46–0.54, p < 0.05).

TABLE 4
www.frontiersin.org

Table 4. Standardized solutions for the measurement models.

TABLE 5
www.frontiersin.org

Table 5. Correlations among exogenous variables in the full Reading-to-Writing Skills model (Model 2).

TABLE 6
www.frontiersin.org

Table 6. Correlations among disturbances of writing variables from the full Reading-to-Writing Skills model (Model 2).

Aim 1: RWS model without fluency

Several effects were in the expected range in the reduced model without R-W fluency (Model 1; see Table 7). Word reading predicted handwriting (β = 0.39, SE = 0.06, p < 0.001), spelling (β = 0.76, SE = 0.04, p < 0.001), and writing quality (β = 0.22, SE = 0.09, p < 0.05). Vocabulary predicted spelling (β = 0.09, SE = 0.05, p < 0.05), editing (β = 0.20, SE = 0.05, p < 0.001), and writing quality (β = 0.11, SE = 0.05, p < 0.05). Reading strategies predicted editing (β = 0.44, SE = 0.05, p < 0.001), revision (β = 0.33, SE = 0.08, p < 0.001), and writing quality (β = 0.30, SE = 0.09, p < 0.001). As hypothesized, higher-order reading skills such as reading strategies were related to writing quality, but contrary to our expectations, inferencing (β = 0.14, SE = 0.10, p > 0.05) and knowledge (β = 0.06, SE = 0.05, p > 0.05) were not related to writing quality. Similarly, reading strategies were related to higher-order writing skills (editing and revision), but other higher-order reading skills (background knowledge and inferencing) were not related to planning, editing, or revising (see results for Model 1 in Table 7). Finally, reading comprehension predicted planning (β = 0.22, SE = 0.06, p < 0.001), editing (β = 0.18, SE = 0.05, p < 0.001), revision (β = 0.25, SE = 0.07, p < 0.001), and writing quality (β = 0.24, SE = 0.07, p < 0.001).

TABLE 7
www.frontiersin.org

Table 7. Standardized solutions for the structural portion of the Reading-to-Writing Skills models.

Aim 2: RWS model with fluency

In the full Reading-to-Writing Skills model (Model 2, Figure 2) there were similarities in the pattern of associations with Model 1 and one notable difference. Word reading was still related to handwriting (β = 0.35, SE = 0.05, p < 0.05). Reading strategies still predicted editing (β = 0.20, SE = 0.07, p < 0.05), revision (β = 0.20, SE = 0.08, p < 0.05), and writing quality (β = 0.24, SE = 0.08, p < 0.05), and reading comprehension predicted planning (β = 0.18, SE = 0.07, p < 0.05), revision (β = 0.23, SE = 0.07, p < 0.05), and writing quality (β = 0.31, SE = 0.11, p < 0.05). However, vocabulary was only related to editing (β = 0.16, SE = 0.06, p < 0.05), and in this model ORF was related to spelling (β = 0.50, SE = 0.17, p < 0.05), planning (β = 0.15, SE = 0.08, p < 0.05), and editing (β = 0.25, SE = 0.10, p < 0.05). Writing productivity was predicted by reading comprehension (β = 0.20, SE = 0.10, p < 0.05), but not decoding, vocabulary, ORF, or SRF. As Table 7 shows, the nested model for SRF (Δχ2 (6) = 14.35, p < 0.05) was significantly worse fitting than the full model in terms of overall fit, whereas the model for ORF (Δχ2 (4) = 0.86, p > 0.05) was not significantly different from the full model, but these solutions pointed to a key difference: when the hypothesized direct effects of ORF on writing skills are not controlled for, SRF predicted the higher-order writing skills with heavier cognitive load, editing (β = 0.40, SE = 0.09, p < 0.001) and revising (β = 0.23, SE = 0.10, p < 0.05), whereas when the hypothesized direct effects of SRF on writing skills are not controlled for, ORF predicted the writing skills with lower (spelling; β = 0.50, SE = 0.16, p < 0.05) and higher cognitive load (planning, β = 0.14, SE = 0.07, p = 0.05; editing, β = 0.31, SE = 0.06, p < 0.05; and revising, β = 0.18, SE = 0.07, p < 0.05), and vocabulary predicted writing quality (β = 0.12, SE = 0.05, p < 0.05). However, when both ORF and SRF were evaluated simultaneously in the full model, only the effects of ORF remained statistically significant.

Reading-to-Writing Mediation models

The reduced RWM model without R-W fluency (Figure 4) provided a good fit to the data (χ2 (df) = 202.80 (121), p < 0.001; RMSEA [90% CI] = 0.04 [0.03, 0.05]; CFI = 0.97; TLI = 0.95; SRMR = 0.04), as did the full RWM model (Figure 6; χ2 (df) = 368.22 (195), p < 0.001; RMSEA [90% CI] = 0.05 [0.04, 0.05]; CFI = 0.96; TLI = 0.94; SRMR = 0.04). Table 3 shows that the RWM models explained a larger proportion of variance compared to the direct effects models (e.g., the full RWM model explained 90% variance in writing quality). These models also explained a large amount of variance in reading skills (e.g., the full RWM model explained 62% variance in inferencing and 94% in ORF). The measurement model solutions were similar to the solution of the RWS models presented above (see Table 4).

Aim 3: RWM model without fluency

First, the results of the RWM model showed that several direct effects of R-to-R skills were in the expected range: inferencing (β = 0.48, SE = 0.14, p < 0.05) and background knowledge (β = 0.14, SE = 0.05, p < 0.05) predicted reading comprehension; vocabulary (β = 0.24, SE = 0.06, p < 0.001) and reading strategies (β = 0.53, SE = 0.08, p < 0.001) predicted inferencing; and word literacy predicted reading strategies (β = 0.64, SE = 0.06, p < 0.001). The effects of several W-to-W skills were also in the expected range: word literacy was related to handwriting (β = 0.43, SE = 0.05, p < 0.05) and editing (β = 0.59, SE = 0.10, p < 0.001); handwriting was related to planning (β = 0.18, SE = 0.06, p < 0.001), editing (β = 0.10, SE = 0.04, p < 0.05), and writing quality (β = 0.25, SE = 0.05, p < 0.001). Planning was related to revision (β = 0.11, SE = 0.05, p < 0.05) and writing quality (β = 0.11, SE = 0.04, p < 0.05), and editing was related to revision (β = 0.31, SE = 0.06, p < 0.001), which in turn was related to writing quality (β = 0.31, SE = 0.05, p < 0.001). Second, Table 8 shows that several cross-domain direct effects from the RWS model (Model 1) also remained significant (e.g., vocabulary ➔ editing, reading strategies ➔ writing quality, reading comprehension ➔ planning, reading comprehension ➔ revision, and reading comprehension ➔ writing quality), with some exceptions: reading strategies were no longer predictive of editing or revision. However, the total indirect effects of several variables were significant as the associations were driven by one or more mediators, as shown in Table 9. Three effects were partially mediated: the effects of reading strategies (via inferencing and reading comprehension), reading comprehension (via revision), and handwriting, on writing quality because zero was not included in the 95% confidence intervals for these effects (see Table 9). Five effects were completely mediated, indicating that the mediators explained all of the relationship between the variables: (1) vocabulary to reading comprehension (specifically, via inferencing), (2) reading strategies to reading comprehension (specifically, via inferencing), (3) inferencing to revision (specifically, via reading comprehension), (4) background knowledge to writing quality, and (5) editing to writing quality.

TABLE 8
www.frontiersin.org

Table 8. Standardized solutions for the structural portion of the Reading-to-Writing Mediation (RWM) models.

TABLE 9
www.frontiersin.org

Table 9. Total indirect and specific indirect effects of the full Reading-to-Writing Mediation (RWM) models.

Aim 4: RWM model with fluency

The full RWM model (Model 4) included all the variables from Model 3 and specified relations with R-W fluency (ORF, SRF, and total words written). In this model, additional R-to-R paths were in the expected range (word literacy predicted ORF [β = 0.94, SE = 0.04, p < 0.001] and SRF [β = 0.66, SE = 0.34, p < 0.05]; vocabulary [β = 0.27, SE = 0.07, p < 0.001] and background knowledge [β = 0.21, SE = 0.07, p < 0.001] predicted SRF, but vocabulary was not related to ORF [β = 0.04, SE = 0.05, p > 0.05]). Contrary to our expectations, ORF and SRF were not significant predictors of inferencing, strategies, or comprehension after controlling for all other variables in the model. Two additional effects of W-to-W skills were in the expected range: revision predicted productivity (β = 0.18, SE = 0.05, p < 0.05), and productivity predicted writing quality (β = 0.18, SE = 0.05, p < 0.001). Unlike the RWS Model 2, writing productivity was not significantly predicted by reading comprehension in the RWM model (see Table 8).

Table 8 shows that most cross-domain effects from the RWS model (Model 2) also remained significant, with some exceptions: background knowledge did not have a direct effect on writing quality, and ORF was no longer predictive of planning or editing. However, the total indirect effect of background knowledge to writing quality was significant as these variables were indirectly related through multiple variables in the R-W system (e.g., knowledge ➔ SRF ➔ writing quality, as well as knowledge ➔ SRF ➔ reading comprehension ➔ writing quality), but none of these specific indirect effects were statistically significant. Overall, Table 9 shows few indirect effects were statistically meaningful (i.e., did not include 0 in the confidence intervals). While editing was not directly related to writing productivity in the RWM model (see Table 8), this effect was completely mediated by revision (β = 0.06, 95% CI [0.01, 0.14]). Finally, nested models that evaluated the hypothesized relations of SRF with writing variables independently of the relation of ORF with writing variables (and vice versa) did not yield a different pattern of results.

Alternative models

The Reading-to-Writing Domains model (alternative model 1) and the Writing-to-Reading Skills model (alternative model 2) provided a good fit to the data (e.g., CFI = 0.97, SRMR = 0.03; see Supplementary Appendixes A,B). The measurement models were similar to that of the RWS model, but the R-to-W Domains model also included a general factor for writing because all the writing dimensions and the CIWS shared method variance (i.e., required human ratings and were derived from the same written response). The R-to-W Domains model explained over half of the variance in writing dimensions (e.g., 83% for ideas; see Table 3). The model also explained 43% of the variance in CIWS. However, the R-to-W Domains model showed that multiple reading skills were not differentially related to specific writing dimensions, except for ORF, which predicted CIWS (β = 0.26, SE = 0.11, p < 0.05) but not the six traits (see Supplementary Appendix A). The nested models for ORF (Δχ2 (5) = 4.29, p > 0.05) and SRF (Δχ2 (4) = 5.55, p > 0.001) were not significantly different from the full model in terms of overall fit. The reduced model solutions again pointed to a key difference: when ORF did not make direct contributions to specific domains, word reading predicted word choice (β = 0.10, SE = 0.04, p < 0.05) and SRF also predicted conventions (β = 0.21, SE = 0.09, p < 0.05) and CIWS (β = 0.44, SE = 0.06, p < 0.001). When paths from SRF to writing domains were omitted, ORF predicted conventions (β = 0.23, SE = 0.12, p < 0.05) and CIWS (β = 0.37, SE = 0.05, p < 0.05).

The Writing-to-Reading Model specified regressions of reading skills on writing skills. The diagram and results of this model are presented in Supplementary Appendix B. Several effects were in the expected range: spelling predicted word reading (β = 0.61, SE = 0.07, p < 0.001), ORF (β = 0.59, SE = 0.07, p < 0.001), and SRF (β = 0.40, SE = 0.08, p < 0.05), and interestingly, spelling also predicted background knowledge (β = 0.21, SE = 0.07, p < 0.001); editing predicted vocabulary (β = 0.26, SE = 0.07, p < 0.001), inferencing (β = 0.23, SE = 0.08, p < 0.05), reading strategies (β = 0.19, SE = 0.07, p < 0.05), SRF (β = 0.11, SE = 0.07, p < 0.001), and reading comprehension (β = 0.17, SE = 0.06, p < 0.05); and writing quality predicted all reading skills, except for word reading. This model explained over half of the variance in reading strategies and ORF and SRF, and a smaller percent of variance in vocabulary and knowledge (see Table 3).

Discussion

A better understanding of the connection of R-W is of vital importance for supporting struggling writers considering the continuing difficulties exhibited and the documented relationships among these skills. Students who experience reading difficulties are increasingly likely to also experience difficulties in writing but teachers report sideling evidence-based writing instruction in the classroom (Graham et al., 2014). Understanding the skill patterns between R-W in ways that support the identification of other skill areas of need is critical.

Increasingly there are theoretical models that highlight R-W connections given their overlap in use of skills (Costa et al., 2016). In this study, we joined DIME components with NVSW to evaluate how lower- and higher-order skills in one domain impact counterpart skills in the other domain, including both ORF and SRF in the models. We limited the scope of the RWM model to malleable skills (i.e., those amenable to training) to increase its practical utility. Thus, we excluded the executive function (including attention, working memory, cognitive control, motivation, and self-efficacy) components of the NSVW model because there is a lack of compelling evidence that executive function training improve academic outcomes or predict response-to-intervention (Fletcher et al., 2018). We found support for the relations among DIME skills: mainly, vocabulary and strategies predicted inferencing, and higher-order knowledge and inferencing predicted reading comprehension. However, with the addition of fluency to the model, vocabulary predicted SRF instead of inferencing, and knowledge also predicted SRF instead of comprehension in line with recent research on this model (e.g., Oslund et al., 2018). We also found support for associations among component skills derived from the NSVW model: mainly among word literacy, handwriting and editing, as well as among planning, editing, revision, and writing quality and productivity. In the next sections, we highlight the findings of the cross-domain associations and their alignment with prior research.

Reading-to-Writing connections

Lower-order reading skills

The results revealed that decoding is related to transcription, specifically spelling and handwriting. This is consistent with previous research which has shown that word-level R-W are connected due to a shared set of skills and knowledge that influence both. Fitzgerald and Shanahan (2000) describe several universal text attributes that help explain the relationship between spelling and decoding, including letter knowledge, phonological and morphological awareness, and knowledge of the orthography of the language. Other studies have also shown that word reading is a correlate of spelling skills (e.g., Abbott et al., 2010), and that it can predict spelling in languages varying in orthographic transparency (Georgiou et al., 2020).

We hypothesized that decoding and vocabulary would not predict writing quality, after controlling for higher-order cognition and comprehension. However, we found that decoding played a role in writing quality, highlighting this fundamental connection for children with LD. Interestingly, the opposite direction did not hold in the W-to-R Model because writing quality predicted all reading skills except for decoding. Collectively, these findings suggest that decoding is an active self-regulatory process in writing, beyond higher-order self-regulation (e.g., editing). Further, when students can read words accurately and fluently, they are more likely to use those words correctly in their own writing, which can help them expand their vocabulary and spelling skills, leading to overall writing quality. The study also found that decoding predicted word choice, a specific component of composition, possibly due to shared knowledge of components involved in the R-W process and could reflect an artifact of print exposure. As students are exposed to print and words, and their meaning, they become stored in the mental lexicon, and thus, more accessible during the writing process. However, it’s possible that students may select words that they know how to spell, thus reinforcing a potential W-to-R pathway. Future research is needed to disentangle these different mechanisms and to provide a more comprehensive understanding of the relationship between decoding, encoding, and word choice in composition.

We found that vocabulary was related to writing quality and editing, but not planning or revision. This suggests that stronger vocabulary facilitates conveying the intended meaning and identifying and correcting errors of spelling and usage effectively, but may not necessarily help revise (for content, organization, tone, and syntax), or help organize and structure ideas before beginning to write. Surprisingly, reading comprehension, rather than vocabulary, was related to productivity, possibly because better vocabulary allows students to express ideas succinctly and precisely but may relate less to total words written than other productivity measures that account for accuracy. On the other hand, comprehension may facilitate understanding the ideas and concepts students are writing about and avoid errors and misunderstandings in their writing that could slow them down.

Higher-order reading skills

In our analysis, we looked at the role of higher-level reading skills, such as background knowledge, inferencing, and reading strategies. Our hypothesis was that knowledge is important for planning because writers need to verbalize their knowledge of a topic before they start writing to focus their writing (Tierney and Pearson, 1983). This was possibly not supported because inexperienced writers simply retrieve ideas prompted by the topic and translate them into text without purposeful engagement in planning, while experienced writers develop a set of goals and generate ideas from their knowledge to achieve these goals (Kucer, 1985). It is also possible that the knowledge assessment in our study affected the results (i.e., it measured knowledge from the reading passages but that was not specific to the writing task). Importantly, we found evidence of an indirect relationship between general knowledge and writing quality. This is not surprising given the importance of knowledge access, use, and generation during the writing process (e.g., Bereiter and Scardamalia, 1987; Allen et al., 2014; Kim, 2020), but our findings suggest that general knowledge influenced writing through its indirect effect on other literacy skills although not via any specific indirect path. Future research should evaluate general and topic-specific knowledge in the context of multiple literacy skills.

Contrary to our hypothesis, inferencing did not predict revision possibly because students with LD are not sophisticated in their revision process to use inferencing skills to detect errors in meaning. Limpo et al. (2013) found that revision skills increase from elementary to secondary grades, but in general, students detect and correct mechanical errors more than substantive meaning errors, and further, students are significantly better at correcting than detecting errors in stories with errors deliberately embedded in them. Nonetheless, we found support for complete mediation in which the effect of inferencing on revision was mediated by reading comprehension. It is possible that students who are better at making inferences while reading (i.e., making logical and reasonable assumptions based on information that is not explicitly stated in texts) are better able to understand the deeper meaning of a text and draw connections between different pieces of information. Understanding and interpreting written text effectively may, in turn, allow them to make revisions so their writing is more effective.

Lastly, reading strategies were related to editing, revising, and quality, but the findings did not support our hypothesis that strategies would relate to planning. The performance measure of reading strategies involved reading a passage to compose a summary; students also completed a survey of reading and learning strategies. More strategic students demonstrated better editing and revising skills and produced higher quality compositions, as expected, because editing and revising are strategic and involve self-monitoring, although research on this is limited. Self-regulation (a strategic process) also distinguishes novice from expert writers and employing more strategies during writing likely results in higher quality compositions. Although the measure focused on reading strategies, there may be overlap with writing strategies due to shared knowledge and skills between the two. Procedural knowledge, purposive information access, and goal-setting could all be influencing factors.

Oral and silent reading fluency

ORF predicted spelling possibly because letter-sound knowledge (Paige et al., 2019) and conceptual word knowledge (Zutell and Rasinski, 1989) underlie both component skills. Zutell and Rasinski (1986) found that the correlation of spelling was higher with oral reading accuracy than rate or prosody, further emphasizing the important role of phonological and orthographic knowledge above and beyond speed or expressiveness of reading aloud. Further, Paige and colleagues evaluated the opposite directionality (spelling ➔ ORF) in a sample of third graders at risk for LDs and found a small, non-hypothesized direct effect of spelling on oral reading fluency after controlling for word- and non-word reading but they did not estimate indirect effects via word and non-word reading. Although these finding contribute to our understanding of the complex interactions among reading fluency and spelling, further research is needed to better understand the underlying mechanisms and directionality of the observed relationships.

ORF also predicted the sentence-level writing component with less cognitive load (editing), whereas SRF predicted planning, which requires deeper processing (Alamargot et al., 2006). Interestingly, both ORF and SRF were related to reviewing, and neither were predictive of discourse-level writing outcomes (productivity and quality), suggesting that ORF/SRF are implicated in specific writing processes (De Smet et al., 2018; Conijn et al., 2022). However, it should be noted that there is a lack of research available to definitively interpret the results of ORF/SRF with writing processes and dimensions.

Finally, ORF/SRF were related to specific dimensions of writing (CIWS and conventions). Interestingly, when paths for both SRF/ORF were included, SRF no longer related to writing dimensions. However, when paths from ORF were excluded, both SRF and decoding predicted conventions, and word reading predicted word choice. These findings emphasize the differential relations of fluency skills with writing dimensions, further emphasizing (a) the importance of the construct of fluency and the connections between fluency in R-W, and (b) the value of ORF/SRF not only as indicators of overall reading but as broader language indicators that capture dimensions such as CIWS, which is thought to reflect both fluency and accuracy of writing. It is therefore not surprising that ORF/SRF were also related to conventions (Conijn et al., 2022), given its design to capture elements of capitalization and punctuation, as these elements are also captured in the CIWS metric. Although based on fluency, ORF was originally designed to serve as an overall indicator of reading performance (Deno, 2003), capturing fluency and related skills (e.g., vocabulary and reading comprehension). The intention of reading fluency to function in this capacity can be seen particularly in the outcomes for our alternative Reading-to-Writing Domains model. To further advance our understanding of the complex relationship between reading fluency and writing processes/domains, future research should aim to explore the co-development of these skills in students with LDs. This research could benefit from utilizing experimental measures and procedures borrowed from the discourse-processing literature, such as eye-tracking technology to investigate concurrent, silent reading and writing processes (Anson and Schwegler, 2012) in addition to standardized measures of ORF/SRF. Such methods can provide valuable insights into the underlying cognitive mechanisms involved in the development of reading fluency and writing skills and could ultimately inform more effective interventions for students with LDs.

Reading comprehension

Contrary to our expectations, we found that writing quality and productivity were not related to ORF and SRF, but rather to reading comprehension. This means that just knowing more words or reading with accuracy, speed, and expression does not automatically lead to a deeper level of written expression or self-regulatory writing activities. Similarly, we found that reading comprehension (and not ORF/SRF, inferencing or strategies) played a significant role in multiple indirect effects. First, we expected that higher-order reading would mediate the relationship between word/world knowledge and writing processes (e.g., vocabulary ➔ inference ➔ revision, or vocabulary ➔ strategies ➔ planning), as students with stronger foundational skills could apply those skills better in R-W activities like planning. However, higher-order skills were found to mediate other higher-order skills and writing processes or outcomes (e.g., strategies ➔ inference ➔ reading comprehension ➔ writing quality), although these effects were small. Revision mediated the relationship between reading comprehension and writing quality, suggesting that comprehension is related to writing quality because it facilitates making revisions. Inferencing and reading comprehension mediated the relationship between reading strategies and writing quality, indicating that reading strategies facilitate better inferencing and understanding of the text, which in turn improves writing.

Our findings also showed that adequate reading comprehension relates to planning, editing, and revision and writing quality and productivity. Stronger comprehension facilitates generating and organizing ideas, communicating thoughts effectively in writing, and analyzing one’s own writing to identify and correct errors. These findings demonstrate that the connection between reading comprehension and writing goes beyond shared content or domain knowledge. The comprehension-planning link points to the connection with procedural knowledge for accessing information purposely, setting goals, and analyzing. The comprehension-editing link points to the pragmatic knowledge of text attributes (words, syntax, usage). The comprehension-revision link points to meta-knowledge about written language (functions and purposes).

The results suggest knowledge and abilities in R-W skills may transfer across domains. Teaching a skill in one domain (e.g., reading comprehension) can directly impact another (e.g., written expression). Implicit R-W connections may also develop (e.g., improved reading comprehension through targeted inferencing instruction may lead to better writing and revision). These findings have important implications for targeted interventions, particularly when paired with screeners like ORF in the context of MTSS.

Writing-to-Reading connections

The W-to-R model fit well and provided evidence for robust effects of spelling, editing, and writing quality on word/letter-, sentence-, and text-level reading. Although the reading skills of inferencing and knowledge did not predict writing quality, the opposite directions held in the W-to-R model, which is an important finding. Effective writing requires anticipating the reader’s potential inferences and identifying important themes and connections from knowledge to produce coherent text. Better writers likely produced clearer and logically consistent texts, although we did not score the essays for logical coherence. Similarly, editing predicted inferencing, but planning and revision were not predictive of reading skills, suggesting that editing (the ability to correct errors in spelling, grammar, punctuation, and syntax) facilitates meaning making processes (i.e., inferencing) because both skills require detecting inconsistencies in meaning using context clues and background knowledge whereas planning does not necessarily rely on context cues. Revision involves making changes to the content or structure of text, which may not necessarily rely on the use of context clues or background knowledge in the same way as editing. Our findings suggest that writing activities (e.g., writing-to-learn) may build word/world knowledge and meaning-making processes (e.g., inferencing and comprehension monitoring) by providing opportunities to write about new topics, using new vocabulary, and monitoring inconsistencies and meaning in their texts. Additional research is needed to fully understand the W-to-R directionality in multiple instructional contexts (e.g., integrated R-W instruction using writing-to-learn or reading-to-write approaches) in ways that direct both theory and practice for specific populations like those with LDs.

Limitations and future directions

The findings of this study are limited to students with/at-risk for LDs and should be replicated and systematically contrasted with both typically developing and other special populations. Further, the nature of the measures impacted our results. For example, planning was not affected by knowledge, possibly because the knowledge measure was not aligned with the writing prompt. The planning measure may have been insufficient to capture the student’s necessary background knowledge, as students were only given 5 min to plan without explicit instructions or tools. Also, writing requirements in most strategy assessments may align more with editing and revision than with planning. Alternative performance measures, such as comprehension monitoring, and the inclusion of writing-specific self-reported strategy use could potentially yield different results. In the RWM Model specifically, editing and revision are framed as activities that take place during writing, but their measures were based on a TOWL subtest that assessed the use of contextual conventions in student essays. Future research should incorporate experimental indicators for capturing inter- and intra-individual editing and revision patterns alongside offline measures of grammar, syntax, and organization, for example. It is possible that sentence-level measurement of SRF is insufficient and alternative indicators are necessary to capture the SRF needed for advanced text reading and composition. Similarly, we measured handwriting quality, not fluency (the ability to write quickly and legibly). Future research using measures of handwriting fluency could help further our understanding of how handwriting accuracy and rate impact associations among R-W, particularly in conjunction with ORF, SRF, and writing productivity. Lastly, regarding measurement used in the study, the individual traits of the 6 + 1 rubric have poor reliability and multicollinearity issues, as shown by high correlations among the traits in the present study (r = 0.53–0.81). Future research should assess writing dimensions using measures/indicators with high content and face validity. Future studies should also aim to include multiple prompts, genres, and types of measures and evaluate the common method variance in component skills models that make use of a single writing prompt to derive multiple indicators.

Another limitation is that insufficient data were available on the specific components, frequency, and duration of the writing instruction each student received. Bi-directional relations among malleable component-skills, such as those in the R-to-W models of the present study, should be systematically evaluated by introducing variations in (a) instruction and (b) executive function and motivation requirements of the task. For example, the DIME model’s direct/indirect relations changed as a function of the focus of intervention (e.g., foundational skills vs. text processing; Ahmed et al., 2022b). We expect that means in R-W skills, and pattern of direct/indirect effects in R-to-W and W-to-R models will be disrupted as a function of the instruction (e.g., writing-for-reading, text-structure) and task requirements. Future research is needed to evaluate the proposed model and alternative specifications (including W-to-R) in multiple instructional and assessment contexts.

Conclusion

Despite existing evidence of the general connection between reading and writing for children with learning difficulties, an evaluation of the specific relationships between different literacy sub-skills is crucial for identifying potential areas for improvement and understanding instructional challenges. This study investigated higher-order reading skills of the Direct and Inferential Mediation (DIME) model as mediators between basic literacy and knowledge, and the writing processes and products of the Not-so-Simple View of Writing (NSVW) model. This study offers a comprehensive evaluation of connections among higher-order skills in both reading and writing domains in component-skills models, considering a range of model constellations that systematically explore the role of oral and silent reading fluency and comprehension skills. The three reading-to-writing models evaluated in the current study (Reading-to-Writing Skills, Reading-to-Writing Dimensions, and Reading-to-Writing Mediation) and the Writing-to-Reading alternative models are comprehensive in scope (in terms of the number of higher-order skills included) and depth (in terms of the granularity of constructs). The findings provide ongoing support for the importance of the constructs of higher-order reading and writing, and fluency, and add to the empirical literature on the direct and indirect effects among components skills at the sub-word/word, sentence, and text-levels. Specifically, the study’s findings highlight the intricate interplay between reading and writing skills, emphasizing the explicit roles of decoding, comprehension, fluency, and strategies in shaping writing quality and productivity, and the implicit roles of background knowledge and inferencing in shaping writing processes and quality. We found that, as opposed to vocabulary, decoding predicted word choice, and reading comprehension had an impact on writing productivity. Background knowledge exerted an indirect influence on writing through its effects on other literacy skills, and we identified a mediation effect of inferencing impacting revision through reading comprehension. Additionally, distinctions between oral and silent reading fluency emerged, with the former predicting spelling and editing and the latter predicting planning. Both types of fluency were linked to reviewing and specific dimensions of writing. Finally, writing quality was found to predict inferencing and knowledge, while editing predicted inferencing. These findings highlight the importance of fostering strong reading abilities and writing skills to enhance students’ overall literacy performance, emphasizing the need for a comprehensive approach to literacy education that nurtures both reading and writing competencies. Studying the relationship between reading fluency and writing skills across different levels of language is critical, given the reliance on ORF for screening and progress monitoring in practice. The study’s findings suggest that ORF may serve as an indicator of both reading difficulty and writing performance across different levels of writing. However, the results of the present study should serve as the basis for further studies on the relationship between reading fluency and component skills of writing.

Data availability statement

The data are available upon request from the Texas Institute for Measurement, Evaluation, and Statistics (TIMES). Requests to access these datasets should be directed to yusra.ahmed@times.uh.edu.

Ethics statement

The studies involving human participants were reviewed and approved by the Institutional Review Boards of the University of Houston and the University of Texas at Austin. Written informed consent to participate in this study was provided by the participants’ legal guardian/next of kin.

Author contributions

YA conceived of the presented idea and performed the computations. YA, SCK, and MKM refined the ideas, discussed the results, and contributed to the final manuscript.

Funding

This work was supported by the Eunice Kennedy Shriver National Institute of Child Health and Human Development (NICHD) under Grants R01HD096262 and P50HD052117.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Author disclaimer

The claims expressed in this article are solely those of the authors and do not necessarily represent those of the National Institute of Child Health and Human Development.

Supplementary material

The Supplementary material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2023.1033970/full#supplementary-material

Footnotes

1. ^Additional writing dimensions were evaluated in the Reading-to-Writing Dimensions model (see Supplementary Appendix A).

2. ^Full models refer to structural models that include all the R-W constructs we evaluated (14 or 15 observed and latent variables). This is different from fully saturated models, which outline every possible path among the constructs. Thus, the term full here refers to the number of factors rather than the number of paths.

References

Abbott, R. D., Berninger, V. W., and Fayol, M. (2010). Longitudinal relationships of levels of language in writing and between writing and reading in grades 1 to 7. J. Educ. Psychol. 102, 281–298. doi: 10.1037/a0019318

CrossRef Full Text | Google Scholar

Ahmed, Y., Francis, D. J., York, M., Fletcher, J. M., Barnes, M., and Kulesz, P. (2016). Validation of the direct and inferential mediation (DIME) model of reading comprehension in grades 7 through 12. Contemp. Educ. Psychol. 44-45, 68–82. doi: 10.1016/j.cedpsych.2016.02.002

CrossRef Full Text | Google Scholar

Ahmed, Y., Kent, S., Cirino, P. T., and Keller-Margulis, M. (2022a). The not-so-simple view of writing in struggling readers/writers. Read. Writ. Q. 38, 272–296. doi: 10.1080/10573569.2021.1948374

PubMed Abstract | CrossRef Full Text | Google Scholar

Ahmed, Y., Miciak, J., Taylor, W. P., and Francis, D. J. (2022b). Structure altering effects of a multicomponent reading intervention: an application of the direct and inferential mediation (DIME) model of reading comprehension in upper elementary grades. J. Learn. Disabil. 55, 58–78. doi: 10.1177/0022219421995904

PubMed Abstract | CrossRef Full Text | Google Scholar

Ahmed, Y., Wagner, R. K., and Lopez, D. (2014). Developmental relations between reading and writing at the word, sentence, and text levels: a latent change score analysis. J. Educ. Psychol. 106, 419–434. doi: 10.1037/a0035692

PubMed Abstract | CrossRef Full Text | Google Scholar

Alamargot, D., Chesnet, D., Dansac, C., and Ros, C. (2006). Eye and pen: a new device for studying reading during writing. Behav. Res. Methods 38, 287–299. doi: 10.3758/BF03192780

PubMed Abstract | CrossRef Full Text | Google Scholar

Allen, L. K., Snow, E. L., Crossley, S. A., Jackson, G. T., and McNamara, D. S. (2014). Reading comprehension components and their relation to writing. L’Annee Psychologique 114, 663–691. doi: 10.3917/anpsy.144.0663

CrossRef Full Text | Google Scholar

Anson, C. M., and Schwegler, R. A. (2012). Tracking the mind’s eye: a new technology for researching twenty-first-century writing and reading processes. Coll. Compos. Commun. 64, 151–171.

Google Scholar

Aud, S., Hussar, W., Johnson, F., Kena, G., Roth, E., Manning, E., et al. (2012). The condition of education 2012 (NCES 2012–045). U.S. Department of Education, National Center for Education Statistics. Washington, DC.

Google Scholar

Bahr, R. H., Lebby, S., and Wilkinson, L. C. (2020). Spelling error analysis of written summaries in an academic register by students with specific learning disabilities: phonological, orthographic, and morphological influences. Read. Writ. 33, 121–142. doi: 10.1007/s11145-019-09977-0

CrossRef Full Text | Google Scholar

Barnes, M. A., and Dennis, M. (2001). Knowledge-based inferencing after childhood head injury. Brain Lang. 76, 253–265. doi: 10.1006/brln.2000.2385

PubMed Abstract | CrossRef Full Text | Google Scholar

Bear, D. R. (1991). “Learning to fasten the seat of my union suit without looking around”: the synchrony of literacy development. Theory Pract. 30, 149–157. doi: 10.1080/00405849109543495

CrossRef Full Text | Google Scholar

Bereiter, C., and Scardamalia, M. (1987). An attainable version of high literacy: approaches to teaching higher-order skills in reading and writing. Curric. Inq. 17, 9–30. doi: 10.2307/1179375

CrossRef Full Text | Google Scholar

Berninger, V. W., Lee, Y. L., Abbott, R. D., and Breznitz, Z. (2013). Teaching children with dyslexia to spell in a reading-writers’ workshop. Ann. Dyslexia 63, 1–24. doi: 10.1007/s11881-011-0054-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Berninger, V. W., Nielsen, K. H., Abbott, R. D., Wijsman, E., and Raskind, W. (2008). Writing problems in developmental dyslexia: under-recognized and under-treated. J. Sch. Psychol. 46, 1–21. doi: 10.1016/j.jsp.2006.11.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Berninger, V. W., Vaughan, K., Abbott, R. D., Begay, K., Coleman, K. B., Curtin, G., et al. (2002). Teaching spelling and composition alone and together: implications for the simple view of writing. J. Educ. Psychol. 94, 291–304. doi: 10.1037/0022-0663.94.2.291

CrossRef Full Text | Google Scholar

Berninger, V. W., and Winn, W. (2006). “Implications of advancements in brain research and technology for writing development, writing instruction, and educational evolution” in Handbook of writing research. eds. C. A. MacArthur, S. Graham, and J. Fitzgerald (New York: The Guilford Press), 96–114.

Google Scholar

Cain, K., and Oakhill, J. V. (1999). Inference making ability and its relation to comprehension failure in young children. Read. Writ. 11, 489–503. doi: 10.1023/A:1008084120205

CrossRef Full Text | Google Scholar

Carretti, B., Motta, E., and Re, A. M. (2016). Oral and written expression in children with reading comprehension difficulties. J. Learn. Disabil. 49, 65–76. doi: 10.1177/0022219414528539

PubMed Abstract | CrossRef Full Text | Google Scholar

Chiesi, H. L., Spilich, G. J., and Voss, J. F. (1979). Acquisition of domain-related information in relation to high and low domain knowledge. J. Verbal Learn. Verbal Behav. 18, 257–273. doi: 10.1016/s0022-5371(79)90146-4

CrossRef Full Text | Google Scholar

Cirino, P. T. (2012). The contextual learning scale. Houston, TX: Author.

Google Scholar

Codding, R. S., Petscher, Y., and Truckenmiller, A. (2015). CBM reading, mathematics, and written expression at the secondary level: examining latent composite relations among indices and unique predictions with a state achievement test. J. Educ. Psychol. 107, 437–450. doi: 10.1037/a0037520

PubMed Abstract | CrossRef Full Text | Google Scholar

Conijn, R., Speltz, E. D., Zaanen, M. V., Waes, L. V., and Chukharev-Hudilainen, E. (2022). A product-and process-oriented tagset for revisions in writing. Writ. Commun. 39, 97–128. doi: 10.1177/07410883211052104

CrossRef Full Text | Google Scholar

Connelly, V., Campbell, S., MacLean, M., and Barnes, J. (2006). Contribution of lower order skills to the written composition of college students with and without dyslexia. Dev. Neuropsychol. 29, 175–196. doi: 10.1207/s15326942dn2901_9

PubMed Abstract | CrossRef Full Text | Google Scholar

Costa, L. J. C., Edwards, C. N., and Hooper, S. R. (2016). Writing disabilities and reading disabilities in elementary school students: rates of co-occurrence and cognitive burden. Learn. Disabil. Q. 39, 17–30. doi: 10.1177/0731948714565461

CrossRef Full Text | Google Scholar

Cragg, L., and Nation, K. (2006). Exploring written narrative in children with poor reading comprehension. Educational Psychology, 26, 55–72. doi: 10.1080/01443410500340991

CrossRef Full Text | Google Scholar

Cromley, J. G., and Azevedo, R. (2007). Testing and refining the direct and inferential mediation model of reading comprehension. J. Educ. Psychol. 99, 311–325. doi: 10.1037/0022-0663.99.2.311

CrossRef Full Text | Google Scholar

Culham, R. (2003). 6+ 1 traits of writing: The complete guide grades 3 and up Scholastic Inc.

Google Scholar

Davis, S. J., and Winek, J. (1989). Improving expository writing by increasing background knowledge. J. Read. 33, 178–181.

Google Scholar

Deane, P., Odendahl, N., Quinlan, T., Fowles, M., Welsh, C., and Bivens-Tatum, J. (2008). Cognitive models of writing: writing proficiency as a complex integrated skill. ETS Res. Rep. Ser. 2008, 1–36. doi: 10.1002/j.2333-8504.2008.tb02141.x

CrossRef Full Text | Google Scholar

Decker, S. L., Roberts, A. M., Roberts, K. L., Stafford, A. L., and Eckert, M. A. (2016). Cognitive components of developmental writing skill. Psychol. Sch. 53, 617–625. doi: 10.1002/pits.21933

CrossRef Full Text | Google Scholar

Deno, S. L. (2003). Developments in curriculum-based measurement. J. Spec. Educ. 37, 184–192. doi: 10.1177/00224669030370030801

CrossRef Full Text | Google Scholar

De Smet, M. J., Leijten, M., and Van Waes, L. (2018). Exploring the process of reading during writing using eye tracking and keystroke logging. Writ. Commun. 35, 411–447. doi: 10.1177/0741088318788070

CrossRef Full Text | Google Scholar

Espin, C., Wallace, T., Campbell, H., Lembke, E. S., Long, J. D., and Ticha, R. (2008). Curriculum-based measurement in writing: predicting the success of high-school students on state standards tests. Except. Child. 74, 174–193. doi: 10.1177/001440290807400203

CrossRef Full Text | Google Scholar

Fewster, S., and MacMillan, P. D. (2002). School-based evidence for the validity of curriculum-based measurement of reading and writing. Remedial Spec. Educ. 23, 149–156. doi: 10.1177/07419325020230030301

CrossRef Full Text | Google Scholar

Fitzgerald, J., and Shanahan, T. (2000). Reading and writing relations and their development. Educ. Psychol. 35, 39–50. doi: 10.1207/s15326985ep3501_5

CrossRef Full Text | Google Scholar

Fletcher, J. M., Lyon, G. R., Fuchs, L. S., and Barnes, M. A. (2018). Learning disabilities: From identification to intervention. New York: Guilford Publications.

Google Scholar

Flower, L., and Hayes, J. R. (1980). The cognition of discovery: defining a rhetorical problem. Coll. Compos. Commun. 31, 21–32. doi: 10.2307/356630

CrossRef Full Text | Google Scholar

Georgiou, G. K., Torppa, M., Landerl, K., Desrochers, A., Manolitsis, G., de Jong, P. F., et al. (2020). Reading and spelling development across languages varying in orthographic consistency: do their paths cross? Child Dev. 91, e266–e279. doi: 10.1111/cdev.13218

PubMed Abstract | CrossRef Full Text | Google Scholar

Gioia, A. R., Ahmed, Y., Woods, S. P., and Cirino, P. T. (2023). Properties of a combined measure of Reading and writing: the assessment of writing, self-monitoring, and Reading (AWSM reader). Read. Writ. 36, 723–744. doi: 10.1007/s11145-022-10274-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Gough, P. B., and Tunmer, W. E. (1986). Decoding, reading, and reading disability. Remedial Spec. Educ. 7, 6–10. doi: 10.1177/074193258600700104

CrossRef Full Text | Google Scholar

Graham, S. (2018). A revised writer (s)-within-community model of writing. Educ. Psychol. 53, 258–279. doi: 10.1080/00461520.2018.1481406

CrossRef Full Text | Google Scholar

Graham, S., Capizzi, A., Harris, K. R., Hebert, M., and Morphy, P. (2014). Teaching writing to middle school students: a national survey. Read. Writ. 27, 1015–1042. doi: 10.1007/s11145-013-9495-7

CrossRef Full Text | Google Scholar

Graham, S., and Hebert, M. (2011). Writing to read: a meta-analysis of the impact of writing and writing instruction on reading. Harv. Educ. Rev. 81, 710–744. doi: 10.17763/haer.81.4.t2k0m13756113566

CrossRef Full Text | Google Scholar

Hammill, D. D., and Larsen, S. C. (2009). Test of written language-(TOWL-4). Austin, TX: Pro-Ed.

Google Scholar

Hebert, M., Bohaty, J. J., Nelson, J. R., and Brown, J. (2016). The effects of text structure instruction on expository reading comprehension: a meta-analysis. J. Educ. Psychol. 108, 609–629. doi: 10.1037/edu0000082

CrossRef Full Text | Google Scholar

Hebert, M., Kearns, D. M., Hayes, J. B., Bazis, P., and Cooper, S. (2018). Why children with dyslexia struggle with writing and how to help them. Lang. Speech Hear Schools 49, 843–863. doi: 10.1044/2018_lshss-dyslc-18-0024

PubMed Abstract | CrossRef Full Text | Google Scholar

Hoover, W. A., and Gough, P. B. (1990). The simple view of reading. Read. Writ. 2, 127–160. doi: 10.1007/BF00401799

CrossRef Full Text | Google Scholar

Hu, L., and Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: conventional criteria versus new alternatives. Struct. Equ. Model. 6, 1–55. doi: 10.1080/10705519909540118

CrossRef Full Text | Google Scholar

Juel, C., Griffith, P. L., and Gough, P. B. (1986). Acquisition of literacy: a longitudinal study of children in first and second grade. J. Educ. Psychol. 78, 243–255. doi: 10.1037/0022-0663.78.4.243

CrossRef Full Text | Google Scholar

Kaufman, A. S. (2004). Kaufman brief intelligence test–second edition (KBIT-2). Circle Pines, MN: American Guidance Service.

Google Scholar

Kim, Y. S. G., and Graham, S. (2022). Expanding the direct and indirect effects model of writing (DIEW): dynamic relations of component skills to various writing outcomes. J. Educ. Psychol. 114, 215–238. doi: 10.1037/edu0000564

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, Y. S. G. (2020). “Interactive dynamic literacy model: an integrative theoretical framework for reading-writing relations” in Reading-writing connections. eds. R. Alves, T. Limpo, and R. M. Joshi (Cham: Springer), 11–34.

Google Scholar

Kim, Y. S. G., and Schatschneider, C. (2017). Expanding the developmental models of writing: a direct and indirect effects model of developmental writing (DIEW). J. Educ. Psychol. 109, 35–50. doi: 10.1037/edu0000129

PubMed Abstract | CrossRef Full Text | Google Scholar

Kintsch, W. (1988). The role of knowledge in discourse comprehension: a construction-integration model. Psychological Review, 95:163. doi: 10.1037/0033-295X.95.2.163

CrossRef Full Text | Google Scholar

Kirby, K., Nist, S. L., and Simpson, M. L. (1986). The Reading journal: a bridge between Reading and writing. Forum Read 18, 13–19.

Google Scholar

Kucer, S. L. (1985). The making of meaning: Reading and writing as parallel processes. Writ. Commun. 2, 317–336. doi: 10.1177/0741088385002003006

CrossRef Full Text | Google Scholar

Langer, J. A. (1986). Children reading and writing: structures and strategies. Norwood, NJ: Ablex Publishing.

Google Scholar

Lefly, D. L., and Pennington, B. F. (1991). Spelling errors and reading fluency in compensated adult dyslexics. Ann. Dyslexia 41, 141–162. doi: 10.1007/bf02648083

PubMed Abstract | CrossRef Full Text | Google Scholar

Limpo, T., Alves, R. A., and Fidalgo, R. (2013). Children’s high-level writing skills: development of planning and revising and their contribution to writing quality. Br. J. Educ. Psychol. 84, 177–193. doi: 10.1111/bjep.12020

PubMed Abstract | CrossRef Full Text | Google Scholar

MacArthur, C. A. (2016). “Instruction in evaluation and revision” in Handbook of writing research. eds. C. A. MacArthur, S. Graham, and J. Fitzgerald. 2nd ed (New York, NY: The Guilford Press), 272–287.

Google Scholar

MacGinitie, W. H., MacGinitie, R. K., Maria, K., and Dryer, L. (2000). Gates-MacGinitie reading tests, 4th. Itasca, IL: Riverside Publishing.

Google Scholar

Mather, N., Wendling, B. J., and Woodcock, R. W. (2001). Essentials of WJ III [TM] Tests of Achievement Assessment. Essentials of psychological assessment series. New York, NY: John Wiley & Sons.

Google Scholar

McCutchen, D. (1996). A capacity theory of writing: working memory in composition. Educ. Psychol. Rev. 8, 299–325. doi: 10.1007/bf01464076

CrossRef Full Text | Google Scholar

McFarland, J., Hussar, B., Zhang, J., Wang, X., Wang, K., Hein, S., et al. (2019). The condition of education 2019. NCES 2019-144. Jessup, MD: National Center for Education Statistics.

Google Scholar

Mehta, P. D., Foorman, B. R., Branum-Martin, L., and Taylor, W. P. (2005). Literacy as a unidimensional multilevel construct: Validation, sources of influence, and implications in a longitudinal study in grades 1 to 4. Scientific Studies of Reading, 9, 85–116. doi: 10.1207/s1532799xssr0902_1

CrossRef Full Text | Google Scholar

Morris, D., Trathen, W., Perney, J., Gill, T., Schlagal, R., Ward, D., et al. (2017). Three DIBELS tasks vs. three informal reading/spelling tasks: a comparison of predictive validity. Read. Psychol. 38, 289–320. doi: 10.1080/02702711.2016.1263700

CrossRef Full Text | Google Scholar

Mulaik, S. A., and Millsap, R. E. (2000). Doing the four-step right. Struct. Equ. Model. 7, 36–73. doi: 10.1207/s15328007sem0701_02

CrossRef Full Text | Google Scholar

Muthen, L. K., and Muthen, B. O. (1998–2017). Mplus user’s guide. 8th. Los Angeles, CA: Muthen & Muthen.

Google Scholar

National Association of Colleges and Employers (2017). Job Outlook 2018. Bethlehem, PA: National Association of Colleges and Employers.

Google Scholar

National Commission on Writing. (2004). Writing: a ticket to work or a ticket out: a survey of business leaders. Available at: https://archive.nwp.org/cs/public/print/resource/2540

Google Scholar

National Commission on Writing. (2005). Writing: a powerful message from state government. Available at: https://archive.nwp.org/cs/public/print/resource/2541

Google Scholar

Neumann, H., Leu, S., McDonough, K., and Crawford, B. (2020). Improving students’ source integration skills: does a focus on reading comprehension and vocabulary development work? J. Engl. Acad. Purp. 48:100909. doi: 10.1016/j.jeap.2020.100909

CrossRef Full Text | Google Scholar

O’Reilly, T., and McNamara, D. S. (2007). The impact of science knowledge, reading skill, and reading strategy knowledge on more traditional “high-stakes” measures of high school students’ science achievement. Am. Educ. Res. J. 44, 161–196. doi: 10.3102/0002831206298171

CrossRef Full Text | Google Scholar

Olinghouse, N. G., and Graham, S. (2009). The relationship between the discourse knowledge and the writing performance of elementary-grade students. J. Educ. Psychol. 101, 37–50. doi: 10.1037/a0013462

CrossRef Full Text | Google Scholar

Olinghouse, N. G., and Wilson, J. (2013). The relationship between vocabulary and writing quality in three genres. Read. Writ. 26, 45–65. doi: 10.1007/s11145-012-9392-5

CrossRef Full Text | Google Scholar

Oslund, E. L., Clemens, N. H., Simmons, D. C., and Simmons, L. E. (2018). The direct and indirect effects of word reading and vocabulary on adolescents’ reading comprehension: comparing struggling and adequate comprehenders. Read. Writ. 31, 355–379. doi: 10.1007/s11145-017-9788-3

CrossRef Full Text | Google Scholar

Oslund, E. L., Clemens, N. H., Simmons, D. C., Smith, S. L., and Simmons, L. E. (2016). How vocabulary knowledge of middle-school students from low socioeconomic backgrounds influences comprehension processes and outcomes. Learn. Individ. Differ. 45, 159–165. doi: 10.1016/j.lindif.2015.11.013

CrossRef Full Text | Google Scholar

Paige, D. D., Smith, G. S., Rasinski, T. V., Rupley, W. H., Magpuri-Lavell, T., and Nichols, W. D. (2019). A path analytic model linking foundational skills to grade 3 state reading achievement. J. Educ. Res. 112, 110–120. doi: 10.1080/00220671.2018.1445609

CrossRef Full Text | Google Scholar

Pearson, P. D., and Tierney, R. J. (1984). Chapter VI: on becoming a thoughtful reader: learning to read like a writer. Teach. Coll. Rec. 85, 144–173. doi: 10.1177/016146818408500506

CrossRef Full Text | Google Scholar

Pike, M. M., Barnes, M. A., and Barron, R. W. (2010). The role of illustrations in children’s inferential comprehension. J. Exp. Child Psychol. 105, 243–255. doi: 10.1016/j.jecp.2009.10.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Re, A. M., and Carretti, B. (2016). Further evidence of poor comprehenders’ difficulty with expressive writing: exploring the role of inferences. Res. Dev. Disabil. 51-52, 145–152. doi: 10.1016/j.ridd.2016.01.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Reid, E., Ahmed, Y., and Keller-Margulis, (in press). Contributions of attentional control, hyperactivity, and reading skills to performance on a fourth-grade state writing test. J. Sch. Psychol. 99C:101220.

Google Scholar

Ritchey, K. D., and Coker, D. L. Jr. (2014). Identifying writing difficulties in first grade: an investigation of writing and reading measures. Learn. Disabil. Res. Pract. 29, 54–65. doi: 10.1111/ldrp.12030

CrossRef Full Text | Google Scholar

Roberts, G. J., Capin, P., Roberts, G., Miciak, J., Quinn, J. M., and Vaughn, S. (2018). Examining the effects of afterschool reading interventions for upper elementary struggling readers. Remedial Spec. Educ. 39, 131–143. doi: 10.1177/0741932517750818

PubMed Abstract | CrossRef Full Text | Google Scholar

Shanahan, T., and Lomax, R. G. (1986). An analysis and comparison of theoretical models of the reading–writing relationship. J. Educ. Psychol. 78, 116–123. doi: 10.1037/0022-0663.78.2.116

CrossRef Full Text | Google Scholar

Shanahan, T. (2012). “Developing fluency in the context of effective literacy instruction” in Fluency instruction: Research-based best practices. eds. T. Rasinski, C. Blachowicz, and K. Lems (New York: Guilford Press), 17–34.

Google Scholar

Shinn, M. M., and Shinn, M. R. (2002). AIMSweb training workbook: Administration and scoring of reading curriculum-based measurement (R-CBM) for use in general outcome measurement. Eden Prairie, MN: Edformation, Inc.

Google Scholar

Shinn, M. R., Knutson, N., Good, R. H. III, Tilly, W. D. III, and Collins, V. L. (1992). Curriculum-based measurement of oral reading fluency: a confirmatory analysis of its relation to reading. Sch. Psychol. Rev. 21, 459–479. doi: 10.1080/02796015.1992.12085629

CrossRef Full Text | Google Scholar

Singer, B. D., and Bashir, A. S. (2004). “Developmental variations in writing composition skills” in Handbook of language and literacy: development and disorders. eds. C. A. Stone, E. R. Stillman, B. J. Ehren, and K. Akpel (New York, NY: The Guilford Press), 559–582.

Google Scholar

Smith, S. R. (2013). Testing a multicomponent model of reading comprehension for seventh- and eighth-grade students (Doctoral dissertation). College Station, TX: Texas A&M University Libraries.

Google Scholar

Tierney, R. J., and Pearson, P. D. (1983). Toward a composing model of reading. Language Arts 60, 568–580.

Google Scholar

Tierney, R. J., and Shanahan, T. (1991). “Research on the reading-writing relationship: interactions, transactions, and outcomes” in Handbook of reading research. eds. R. Barr, M. L. Kamil, P. B. Mosenthal, and P. D. Pearson, vol. 2 (New York: Lawrence Erlbaum Associates, Inc), 246–280.

Google Scholar

Torgesen, J. K., Rashotte, C. A., and Wagner, R. K. (1999). TOWRE: test of word reading efficiency. Austin, TX: Pro-Ed.

Google Scholar

Tortorelli, L. S., and Truckenmiller, A. J. (2023). Automaticity in writing in response to reading: relations between oral reading fluency and compositional writing fluency in grades 3–5. Read. Writ. Q. 1-15, 1–15. doi: 10.1080/10573569.2023.2172757

CrossRef Full Text | Google Scholar

Truckenmiller, A. J., and Petscher, Y. (2020). The role of academic language in written composition in elementary and middle school. Read. Writ. 33, 45–66. doi: 10.1007/s11145-019-09938-7

CrossRef Full Text | Google Scholar

van den Boer, M., Bazen, L., and de Bree, E. (2022). The same yet different: oral and silent reading in children and adolescents with dyslexia. J. Psycholinguist. Res. 51, 803–817. doi: 10.1007/s10936-022-09856-w

PubMed Abstract | CrossRef Full Text | Google Scholar

Vanderberg, R., and Swanson, H. L. (2006). Which components of working memory are important in the writing process? Read. Writ. 20, 721–752. doi: 10.1007/s11145-006-9046-6

CrossRef Full Text | Google Scholar

Völlinger, V. A., Spörer, N., Lubbe, D., and Brunstein, J. C. (2018). A path analytic test of the reading strategies mediation model: relating cognitive competences and motivational influences to individual differences in fifth-grade students’ reading comprehension. J. Educ. Res. 111, 733–745. doi: 10.1080/00220671.2017.1412930

CrossRef Full Text | Google Scholar

von Koss Torkildsen, J., Morken, F., Helland, W. A., and Helland, T. (2016). The dynamics of narrative writing in primary grade children: writing process factors predict story quality. Read. Writ. 29, 529–554. doi: 10.1007/s11145-015-9618-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Wagner, R. K., Torgesen, J. K., Rashotte, C. A., and Pearson, N. A. (2010). TOSREC: Test of silent reading efficiency and comprehension. Austin, TX: Pro-Ed.

Google Scholar

Weston-Sementelli, J. L., Allen, L. K., and McNamara, D. S. (2018). Comprehension and writing strategy training improves performance on content-specific source-based writing tasks. Int. J. Artif. Intell. Educ. 28, 106–137. doi: 10.1007/s40593-016-0127-7

CrossRef Full Text | Google Scholar

White, S., Sabatini, J., Park, B. J., Chen, J., Bernstein, J., and Li, M. (2021). The 2018 NAEP oral reading fluency study. NCES 2021-025. Jessup, MD: National Center for Education Statistics.

Google Scholar

Woodcock, R. W., McGrew, K. S., and Mather, N. (2001). Woodcock-Johnson tests of achievement. (3rd). Itasca, IL: Riverside Publishing.

Google Scholar

Zutell, J., and Rasinski, T. (1986). Spelling ability and reading fluency. Natl. Read. Conf. Yearbook 35, 109–112.

Google Scholar

Zutell, J., and Rasinski, T. (1989). Reading and spelling connections in third and fifth grade students. Read. Psychol. 10, 137–155. doi: 10.1080/0270271890100203

CrossRef Full Text | Google Scholar

Keywords: written expression, text writing, compositional fluency, reading comprehension, text reading fluency, learning disabilities

Citation: Ahmed Y, Kent SC and Keller-Margulis M (2023) Reading-to-Writing Mediation model of higher-order literacy. Front. Psychol. 14:1033970. doi: 10.3389/fpsyg.2023.1033970

Received: 01 September 2022; Accepted: 30 May 2023;
Published: 30 June 2023.

Edited by:

Giseli Donadon Germano, São Paulo State University, Brazil

Reviewed by:

Airil Haimi Mohd Adnan, MARA University of Technology, Malaysia
Erikson Saragih, University of North Sumatra, Indonesia

Copyright © 2023 Ahmed, Kent and Keller-Margulis. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yusra Ahmed, yusra.ahmed@times.uh.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.