Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 10 June 2019
Sec. Educational Psychology
This article is part of the Research Topic Advancements in Technology-Based Assessment: Emerging Item Formats, Test Designs, and Data Sources View all 21 articles

Making the Psychological Dimension of Learning Visible: Using Technology-Based Assessment to Monitor Students’ Cognitive Development

  • 1Department of Learning and Instruction, University of Szeged, Szeged, Hungary
  • 2MTA–SZTE Research Group on the Development of Competencies, University of Szeged, Szeged, Hungary

Technology-based assessment offers unique opportunities to collect data on students’ cognitive development and to use that data to provide both students and teachers with feedback to improve learning. The aim of this study was to show how the psychological dimension of learning can be assessed in everyday educational practice through technology-based assessment in reading, mathematics and science. We analyzed three related aspects of the assessments: cognitive development, gender differences and vertical scaling. The sample for the study was drawn from primary school students in Grades 1–8 (ages 7 to 14) in Hungary. There were 1500 to 2000 students in each grade cohort. Online tests were constructed from 1638 items from the reading, mathematics, and science domains in the eDia system. The results confirmed that the disciplinary, application and psychological dimensions of learning can be distinguished empirically. Students’ cognitive development was the most steady (and effective) in mathematics, where the greatest development occurred in the first years of schooling. Path models suggested that the psychological dimension of learning can be predicted at a moderate level based on students’ level of school knowledge consisting of the disciplinary and application dimensions of learning as latent constructs. The predictive power was almost the same in both dimensions. Generally, girls developed faster in the psychological dimension of reading, mathematics and science learning; however, the size of gender differences varied by age and domain. This study (1) provides evidence that the psychological dimension of learning can be made visible even in an educational context, (2) highlights the importance of the explicit development of the psychological dimension of learning during school lessons, and (3) shows that there are gender differences in the developmental level of the psychological dimension of learning in favor of girls but that this varies by grade and domain.

Introduction

Improving students’ cognitive abilities has always been a goal of schooling since the very beginning of formalized education (Hattie and Anderman, 2013). However, despite the theoretical foundations, assessment instruments and pedagogical practices that have evolved over time, this aim has not yet been met; in many school systems students’ cognitive abilities are not optimally enhanced. In the 20th century, several research schools and paradigms sought to conceptualize cognition, define its key constructs and make them measurable (see e.g., Binet and Simon, 1916; Inhelder and Piaget, 1958; Adey, 2007). Among these, research on intelligence and the related psychometric tradition, Piaget and his school, and the cognitive revolution have all had a major impact on redefining the goals of education. The implications of the research within these paradigms were drawn for educational practice, and a number of mostly stand-alone programs were initiated in the 1970s, outside classroom instruction (Feuerstein et al., 1980; Klauer, 1989a,b, 1991, 1993, 1997). Later on, in the 1990s, developmental effects were embedded in school subjects using the content of learning (Adey and Shayer, 1994; Shayer, 1999; Adey et al., 2001; Shayer and Adey, 2002; Shayer and Adhami, 2007). The related research, including a number of experiments, resulted in a better understanding of the role that cognitive processes play in school learning, but it has had a modest impact on educational practice.

At the beginning of this millennium, more or less the same ideas emerged in a new wave of teaching 21st-century skills. Several projects sought to define, operationalize, measure and teach these skills (see e.g., Trilling and Fadel, 2009; Griffin and Care, 2014), but the same constraints appeared to hinder progress in putting these ideas into practice in mass education as with previous similar attempts. There were no proper tools for assessing and monitoring changes in students’ cognition. The availability of appropriate assessment instruments is a necessary condition for any pre-test – post-test experimental design as well. However, what can be created and applied in specific experimental conditions cannot always be scaled up for broader practical applications. Similarly, the roots of a number of practical educational challenges can be traced back to the fact that significant determinants of school learning are not visible (Hattie, 2009). They are also not easy to observe, nor can developmental deficiencies always be identified by teachers (MacGilchrist et al., 2004). The lack of thinking skills – the cognitive tools required for successful learning – are not identified; thus, they remain untreated, and this significantly hampers further learning.

Thinking, or more specifically, a set of cognitive skills essential for learning, such skills are not observable in the everyday educational context. Students are not aware of the existence of the required processes, and teachers, even if they receive training in identifying the cognitive processes underlying learning, are not able to observe them, or they simply have no time or capacity to determine each student’s individual needs. Although the developmental levels of crucial thinking skills might be measured with traditional paper-based instruments, the immense costs, the human resources required, and the time between assessment and feedback excludes the possibilities of using them diagnostically. Technology may be a solution for making thinking processes visible by creating simpler, faster, frequently applicable and cost-effective assessments (Mayrath et al., 2012).

In this paper, which is part of a larger project, we present the results of work in identifying cognitive processes relevant for learning, making them measurable in normal educational contexts, and providing students and teachers with frequent feedback. One of the most challenging aspects of this work, is establishing the validity of diagnostic instruments to assess of cognitive processes; showing that the tests measure something more than mastering the current teaching material. To do this, we empirically validated a 3-dimensional framework developed for diagnostic assessment and explored the psychometric characteristics of an item bank devised for the assessment of the psychological dimension of learning.

Theoretical Background

The idea of making learning visible was introduced into educational research and development by John Hattie. He made a great step forward in initiating evidence-based educational practice when he synthesized the results of over 800 meta-analyses (Hattie, 2009). He translated his findings into actual classroom work, and in his book for teachers, he explained:

Visible teaching and learning occurs when learning is the explicit and transparent goal, when it is appropriately challenging, and when the teacher and the student both (in their various ways) seek to ascertain whether and to what degree the challenging goal is attained, when there is deliberate practice aimed at attaining mastery of the goal, when there is feedback given and sought, and when there are active, passionate, and engaging people (teachers, students, peers, and so on) participating in the act of learning (Hattie, 2012, p. 18).

As he emphasizes, feedback plays a central role in successful learning, which at a higher level of learning, includes self-monitoring, self-evaluation and self-assessment. However, he also explains how difficult a task it is to provide proper feedback: “Learners can be so different, making it difficult for a teacher to achieve such teaching acts: students can be in different learning places at various times, using a multiplicity of unique learning strategies, meeting different and appropriately challenging goals” (Hattie, 2012, p. 18).

Student diversity, i.e., students at different levels in different cognitive attributes, is not the most challenging phenomenon when proper feedback is considered. A major problem is that a number of learning outcomes, sometimes the most important ones, are not visible and cannot easily be made visible. While the majority of the studies Hattie reports on deal with organizational issues, methods and classroom practices for teaching curricula, there are far fewer studies that cover the underlying cognitive processes, e.g., reasoning skills, required to understand mathematics and science or precursors of reading, such as phonemic awareness. Some studies have focussed on the most hidden aspects of learning. For example, Ritchhart et al. (2011) identify a broad range of teaching and learning practices to make thinking visible. They identify the crucial problem in a simplified conception of learning (reduced to memorization) and knowledge (reduced to information, facts and figures): “When we demystify the thinking and learning processes, we provide models for students of what it means to engage with ideas, to think, and to learn. In doing so, we dispel the myth that learning is just a matter of committing the information in the textbook to one’s memory” (Ritchhart et al., 2011, p. 28).

Taking into account diversity among students, the limited capacity of teachers and the need to provide feedback on the most relevant but least visible aspects of school learning – promoting students’ cognitive development – we may conclude that students and teachers need a different approach to assessment to improve learning. The online assessment system, eDia, was designed for this purpose. It assesses “thinking,” or “cognitive development,” as a separate dimension, which we call the psychological dimension of learning. We briefly introduce the 3-dimensional theoretical framework that forms the basis for the diagnostic assessment system, and then we elaborate on the psychological dimension in more detail, as that is the focal topic of the present study. Finally, we discuss the crucial role of technology, arguing that its widespread availability in schools makes the time right for such a system to be introduced and integrated into regular educational processes.

Learning and Cognitive Development: A 3-Dimensional Model of Learning Outcomes

An online diagnostic assessment system, eDia, has been constructed to provide teachers and students with relevant feedback information (Csapó and Molnár, unpublished). The eDia system covers the three most frequently assessed domains of school education; reading, mathematics and science. Large item banks have been developed for use in regular classroom assessments in Grades 1 to 6 of primary school, and for Grades 7 and 8 to explore the developmental trends in a broader age range.

The objectives of each item bank are defined in its assessment framework, similarly to international comparative studies, such as Trends in International Mathematics and Science Study (TIMSS; Mullis et al., 2005) and Progress in International Reading Literacy Study (PIRLS; Mullis and Martin, 2015); they are based on a 3-dimensional model of the goals of learning that forms a common foundation for diagnostic assessment. The three dimensions include thinking/reasoning, application and disciplinary knowledge. [The 3-dimensional framework has been published in several articles and book chapters before the assessment frameworks were elaborated (see e.g., Csapó, 2010; Nunes and Csapó, 2011; Adey and Csapó, 2012; Blomert and Csépe, 2012)]. The framework for reading was somewhat different those for mathematics and science (Csapó and Csépe, 2012; Csapó et al., 2015c), which were more similar (Csapó and Szendrei, 2011; Csapó and Szabó, 2012; Csapó et al., 2015a,b).

The intention of “cultivating the mind” – developing cognitive abilities – may be traced back to ancient philosophy. To set goals in this direction, a model of mind is needed; more specifically, knowledge of how internal psychological attributes are structured and how psychological processes play a role in learning (see more details in the next section). In the eDia frameworks, this is the “thinking” (this term is mostly used in the context of mathematics and science), or, more generally, the “psychological dimension.” According to the model, we propose the psychological dimension of knowledge does not only contain “domain-specific reasoning skills,” but also general reasoning skills embedded in different content and contexts, which has lately been referred to as transversal skills; and is not the same as procedural knowledge. We assume that there are natural cognitive developmental (psychological) processes. These processes, as described by Piaget, take place in the interaction between the child and his/her environment. School education may stimulate this development if it provides a student with proper environmental stimuli and if these stimuli are within the zone of proximal development (ZPD) of the child (Vygotsky, 1978). Very often, school instruction is not adjusted to the individual needs of the students; usually the stimuli are far beyond their ZPD. In these cases, students benefit little from instruction; they memorize the rules and develop specific skills through a large amount of drill practice, which have any real impact on their cognitive development. For example, students may learn rules to deal with ratios and proportions without this learning having much impact on the development of proportional reasoning. Schools may teach students a great deal about combinatorics, probability and correlation without having a real impact on the development of combinatorial, probabilistic or correlative reasoning. In this way, we distinguish the psychological dimension from the disciplinary dimension, which may include procedural knowledge (e.g., skills for solving linear equations or proving geometric theorems) or domain-specific reasoning skills. This model and approach opens the door to fostering domain-general reasoning skills in a domain-specific context.

Application deals with another ancient goal – that school should teach something that is applicable beyond the school context. Applying knowledge and transferring it to new contexts require a deeper conceptual understanding and usually specific exercises to facilitate application. Therefore, most knowledge mastered at school remains inert and not applicable in new contexts (Alexander and Murphy, 1999; Bransford and Schwartz, 1999; Csapó, 2010). The PISA conducted by the OECD has focussed on this dimension from the very beginning. The PISA expert groups elaborated the concept of applicable knowledge and defined it as competencies students need in a modern society. To develop such a framework, the social relevance of knowledge, i.e., the needs of societies have also be taken into account. For the frameworks of the first and second PISA assessment, the concept of literacy was extended in include the objects of the assessment in the three domains as reading literacy, mathematical literacy and scientific literacy (OECD, 1999, 2003).

Disciplinary knowledge is the third dimension and is most commonly known as curricular content. Arts and sciences content constitutes the major source of disciplinary knowledge. The first major international comparative studies (e.g., First and Second International Mathematics Study – Husén, 1967; Burstein, 1993; First and Second International Science Study – Bloom, 1969; IEA, 1988), the precursor to the TIMSS, assessed this dimension. The first assessments were based on an analysis of the curricula in the participating countries. More recently, the TIMSS frameworks organize the objects of the assessment into three groups: content, application and reasoning. This classification bears some similarity to the 3-dimensional eDia frameworks (For PISA assessment frameworks, see OECD, 2003).

Education must not be reduced to providing the right answer quickly, but must deal with the ongoing cognitive work of understanding new ideas and information that will serve students as learners in the future (Costa and Kallick, 2009). In modern society, students are expected to apply their knowledge in a wide range of contexts, and they should be able to solve problems in unknown, novel situations. Thus, these goals must reinforce and interact with each other as they are strongly connected (Molnár and Csapó, 2019).

It is reasonable that the earliest efforts to measure knowledge learnt at school focussed on areas that were the easiest to measure: the disciplinary (knowledge) dimension of learning (see e.g., IEA TIMSS). The goal of applying that knowledge in a new context (the application dimension) and assessing students’ ability to do so is a more complicated task (see e.g., OECD PISA). The goal of developing students’ thinking abilities (the psychological dimension) is even more complex. To be able to make thinking visible, we must be clear about, and draw on, our understanding of what thinking is and what types of thinking we want to assess and enhance.

Assessment Beyond the Content of Actual Learning

In the 20th century, several research paradigms have conceptualized the development of thinking and its relationship to school education. Among these, research on “intelligence” was the first that was closely linked to education. The first intelligence test (Binet–Simon test, Binet and Simon, 1916) was constructed to assess children’s preparedness for schooling, and the Scholastic Aptitude Test (SAT) (see Grissmer, 2000) served a similar purpose at the transition from secondary to tertiary education. Several new approaches, models and interpretations of the concept of intelligence have been proposed. From the perspective of education, the more useful ones consider intelligence as (able to be modified, taught, learnt, or improved within educational contexts). Our psychological dimension in each domain may thus overlap with the inductive reasoning components of “fluid” intelligence. The psychological dimension can be embedded within the conception of plastic general ability (see Adey et al., 2007), and a number of cognitive skills covered by the psychological dimension of our frameworks are explicitly identified in Carroll’s three-strata model of abilities (Carroll, 1993) and the Specialized Cognitive Systems of Demetriou’s model (Demetriou et al., 1992, 1993; Adey et al., 2007). On the other hand, we emphasize that all the cognitive skills discussed in the psychological dimension of the frameworks are embedded within the content and context of each particular domain, and the tasks developed from the frameworks are adjusted to the developmental level of the cohort of students to be assessed.

The work of Jean Piaget and his school was characterized by another approach. Piaget described students’ reasoning skills with well-defined operations, which correspond with certain mathematical structures (see e.g., Inhelder and Piaget, 1958). He mostly used basic science content for his experiments (e.g., the pendulum), and the operations he identified may be found in various learning contexts as well as in everyday problems.

The cognitive revolution in psychology provided a new impetus to research efforts in school learning. It led to a more differentiated conception of knowledge and learning, allowing a more precise definition of the goals of education. Recent studies in psychology and education have shown that these skills are especially crucial at the beginning years of schooling, as students’ developmental level determines later success (see Nguyen et al., 2016).

The psychological dimension has been conceptualized as the interaction between the development of students’ thinking skills and learning at school (Nunes and Csapó, 2011; Adey and Csapó, 2012; Blomert and Csépe, 2012) and must address how students learn in reading, mathematics and science.

In this study, we explored the prospects of making the psychological dimension of learning visible by using technology-based assessments to monitor the development of students’ thinking skills. The aim of this study was to show how the psychological dimension of learning (thinking) can contribute to the development of specific reasoning skills.

In reading, assessment of the psychological dimension (thinking and reasoning) covers the cognitive mechanisms of development from laborious phonological decoding to the automatic recognition of whole words, and from prerequisite skills of reading through phonemic, phonological and morphological awareness to metacognitive aspects (Blomert and Csépe, 2012). In mathematics (Nunes and Csapó, 2011) and science (Adey and Csapó, 2012), there are generic objects and domain specific objects. For example, number sense is specific to mathematics, while the control of variables and scientific reasoning are better covered within the science framework. Operational reasoning (e.g., seriation, class inclusion, classification, combinatorial reasoning, probabilistic reasoning, proportional reasoning) and some higher-order thinking skills (e.g., inductive reasoning and problem solving) are more generic and can be assessed in both mathematics and science.

Aims, Research Questions and Hypotheses

In this study, we explored the prospects of making the psychological dimension of learning visible by using technology-based assessments to monitor the development of students’ reasoning skills. The aim of the study was to show how the psychological dimension of learning (thinking) can be assessed in everyday educational practice and how it is related to students’ level of subject matter content knowledge. Three domains were explored from this perspective: reading, mathematics and science. Reading is the basis for all further learning, including mathematics and science, while mathematics provides foundations for learning in various areas of science. These domains are central in many education systems, and large-scale international comparative studies, such as TIMSS, PIRLS, and PISA, have focussed on these areas. We analyzed three aspects of the assessments: cognitive development, gender differences and vertical scaling.

Worldwide, there are many initiatives and computer-based tests available in the domains of reading, mathematics and science worldwide. However, they mainly focus only on disciplinary knowledge dimension (content) or the application dimension (literacy of learning) (e.g., TIMSS, PIRLS, and OECD PISA). There are no regular large-scale assessments that include the psychological dimension of learning in primary school – cognitive development. The available assessment systems in reading, mathematics and science have been designed to assess older students’ reading, mathematics and science knowledge (e.g., TIMSS, PIRLS, and PISA). The present study sought to: (1) define and examine the different dimensions of learning in reading, mathematics and science; (2) monitor and compare cognitive development (the psychological dimension of learning) in the three domains over time; (3) analyze the proportion of unexplained variance in cognitive development if school knowledge (the application and disciplinary dimensions) is taken into account in reading, mathematics and science; and (4) identify any gender differences in the cognitive development in the three domains. We sought to answer five research questions.

RQ1: Can the three dimensions of learning be distinguished empirically? We explored this question to see if cognitive development, the development of reasoning skills, can be assessed separately and be made visible in everyday educational practice. We hypothesize that the psychological, application and disciplinary dimensions of learning can be distinguished empirically, assessed and monitored in everyday educational practice (Csapó and Szendrei, 2011; Csapó and Csépe, 2012; Csapó and Szabó, 2012). We also hypothesize that they will interact and correlate with each other.

RQ2: Is the psychological dimension of learning the same across the three domains? That is, is the same construct being measured in the psychological dimension of learning across the three main domains? The roots of cognitive development may be universal as early neurocognitive development in children is similar across cultures and societies (Molnár and Csapó, 2019). Therefore, based on the conceptualization of the psychological dimension of learning as the interaction between students’ cognitive development and learning at school (Nunes and Csapó, 2011), we hypothesize that the 1-dimensional model will fit the data better than the 3-dimensional model. However, we argue that the 3-dimensional model will take into account results from research on knowledge transfer. According to McKeachie (1987), “Spontaneous transfer is not nearly as frequent as one would expect” (p. 709).

RQ3: How does the psychological dimension of reading, mathematics and science develop over time during primary schooling? Based on previous research results on reasoning skills, we hypothesize that children’s cognitive development is slow (Molnár et al., 2013, 2017), indicating the need for more stimulating school lessons. Based on Polya’s (1981) theory of problem solving, and results from research on mathematics teaching (e.g., Nunes and Csapó, 2011), we hypothesize that the psychological dimension of learning in mathematics will develop the most readily.

RQ4: How can the psychological dimension of learning be explained by students’ level of school knowledge in reading, mathematics and science? That is, how can learning in reading, mathematics and science contribute to the development of the psychological dimension of learning, and how effectively does it stimulate students’ general cognitive development? Research in this field provides rich resources ranging from the classical work of Piaget (see e.g., Inhelder and Piaget, 1958) to the most recent neurocognitive studies (such as Geake and Cooper, 2003; Thomas et al., 2019). We hypothesize that learning reading, mathematics and science will contribute to students’ development in the psychological dimension of learning but that the transfer effect will be low. We base our hypothesis on empirical research that has found that reasoning skills develop relatively slowly during primary and secondary education with the average pace of development being about one quarter of a standard deviation per year (Csapó, 1997; Molnár and Csapó, 2011; Greiff et al., 2013; Molnár et al., 2013, 2017). The development of reasoning skills is a “by-product” of teaching rather than guided by explicit instruction (de Koning, 2000).

RQ5: How does the developmental level of the psychological dimension of learning differ by gender, grade and domain? Based on the most prominent international studies (Martin et al., 2016; Mullis et al., 2016, 2017; OECD, 2016) and the research results on gender differences in students’ development of reasoning skills (Wüstenberg et al., 2014), we hypothesize gender differences in the development of the psychological dimension of learning will vary by grade and domain. The PISA studies indicated that the achievement of 15-year-old Hungarian girls in the application dimension of reading was significantly better than that of boys, while there were no statistically significant gender differences in mathematics and science (OECD, 2016). In contrast, the TIMSS studies that focus on younger students (Grades 4 and 8; 10- to 14-year-olds) mainly assess the disciplinary dimension of mathematics and science knowledge. Their findings indicated that boys significantly outperform girls in mathematics in Grade 8 (Mullis et al., 2016), but there was no statistically significant gender difference in Grade 4. In science, boys significantly outperformed girls at both grade levels (Martin et al., 2016). In PIRLS, Grade 4 Hungarian girls significantly outperformed their boys in reading (Mullis et al., 2017). Please note that the present study focussed on the psychological dimension and not on the application or disciplinary dimensions of learning in mathematics, science or reading.

Materials and Methods

Participants

The sample of students for the study was chosen from the partner school network of the Center for Research on Learning and Instruction at the University of Szeged in Hungary. As schools participated voluntarily in the project, representative sampling of school classes or students was not a goal. However, based on the data collected from the schools, it was possible to generate nationally representative indicators for the main variables. We noted that schools with relatively large numbers of low socioeconomic (SES) students were under-represented in the present study, possibly due to the lack of ICT available in those schools.

The sampling unit was a school class. Classes were drawn from primary and secondary schools from Grades 1–8 (aged 7–14). A total of 656 classes from 134 schools in different regions were involved in the study, resulting in a wide-ranging distribution of students’ background variables. The total number of students involved in the study was 14,062 (Table 1). The proportion of boys and girls was about the same. As participation was voluntary, not all students completed tests in all three domains or in each dimension within each domain. Thus, data was potentially available for students who completed nine elements: the assessment of three dimensions of learning (psychological, application, and disciplinary) in three domains (reading, mathematics, and science). After the scaling procedure, we excluded students from the analyses where, because of missing data. it was not possible to compute an ability level in at least one of the nine elements. Thus, 5,714 students from 310 classes and 97 schools were involved in the analyses.

TABLE 1
www.frontiersin.org

Table 1. The sample for the study.

Tests

An item bank was constructed for diagnostic assessments in reading, mathematics and science based on the three dimensions of learning described in the previous section. These item banks collectively contained almost 17,000 tasks with most tasks having several items. There were 6685 tasks for reading, 6691 for mathematics and 3535 for science. Tests to measure the psychological, application, and disciplinary dimensions of learning in reading, mathematics and science among students in Grades 1–6 (aged 6–7 to 12–13). The tests for the study were drawn from these item banks. Students in Grades 7 and 8 received tasks originally written for students in Grades 5 and 6 (see Table 2).

TABLE 2
www.frontiersin.org

Table 2. The structure of the tests in mathematics by cluster of tasks for each grade level.

For each grade level, nine tasks with different difficulty levels (three easy, three medium-difficulty and three difficult) were chosen from each item bank to assess each dimension. After this procedure, there were 543 tasks in reading, 604 in mathematics and 492 in science.

The tasks were grouped into clusters, with 10–15 items per cluster for students in the lower grades and 15–20 items for students in the higher grades. One 45-min test consisted of four clusters of tasks for students in Grades 1 and 2 (50–55 items) and five clusters for students in Grades 3 to 6 (60–85 items). Each test contained clusters of tasks from each learning dimension with the clusters positioned in a different order to avoid the item-position effect in the scaling procedure. Anchor items were used within and between the different grades for the horizontal and vertical scaling of the data. The clusters contained easier or harder tasks from lower or higher grades. A total of 483 strongly anchored, but different clusters were developed from the items selected.

For optimizing the measurement error of the test, the clusters contained tasks from the same dimension of learning, ranging in task difficulty for the different grade levels. That is, students received more tasks from one learning dimension if those tasks were originally prepared for students in lower or higher grades. The structure of the test of mathematical knowledge is presented in Table 2 paralleled the structure of the reading and science tests. Based on this structure, 162 different tests (nine in each grade and each domain) were constructed from the item banks for the vertical scaling of students in Grades 1–8.

In Grades 1–3, instructions were provided in written form, on-screen, and with a pre-recorded voiceover to avoid any reading difficulties and to ensure greater validity of the assessments. Thus, students used headphones during the administration of the tests. After listening to the instructions, they indicated their answer by using the mouse or keyboard (in the case of desktop computers, which are most commonly used in Hungarian schools) or by directly tapping, typing or dragging the elements of the tasks using their fingers on tablets.

The tasks presented in Figure 1 assess students’ mathematical and scientific reasoning. Based on the framework for the diagnostic assessment of mathematics (Csapó and Szendrei, 2011) and science (Csapó and Szabó, 2012), the main questions in this psychological dimension related to how well mathematics and science education was adjusted to students’ psychological development, how learning mathematics and science could contribute to the development of specific reasoning skills and how effectively they could stimulate students’ general cognitive development. Items developed to measure the psychological dimension of learning encompassed a long list of skills, such as inductive reasoning, deductive reasoning, analogical reasoning, combinatorial reasoning, systematization skills, proportional reasoning and correlative reasoning. Two examples of tasks for assessing students’ inductive reasoning are presented in Figure 1. Students had to discover regularities by detecting dissimilarities with respect to attributes of different objects. They completed the tasks by dragging the elements to different areas, thereby defining the proper sets. The scoring of all tasks was automated, including items with several correct answers.

FIGURE 1
www.frontiersin.org

Figure 1. Measuring the psychological dimension of learning: assessment of students’ inductive reasoning skills in the context of geometry and biology.

Figure 2 presents a task measuring student’s science disciplinary knowledge and a mathematics tasks measuring the application dimension. In the science task, students retrieve disciplinary knowledge of phases of the water cycle. In the mathematics task, students have to select and place flowers – drag and drop – in the vase; only the number of flowers counts. The task measures the application of adding up to 10 in a realistic application context.

FIGURE 2
www.frontiersin.org

Figure 2. Measuring the disciplinary dimension of learning science and the application dimension of learning mathematics.

Procedures

The tests were administered over a period of 7 weeks in computer rooms within the participating schools during regular school hours. Each test lasted approximately 45 min. Test sessions were supervised by teachers who had been thoroughly trained in test administration. The tests were delivered on the eDia online platform. After students entered the system and chose the domain (reading, mathematics, or science), the system randomly selected a test for that student from the nine tests available in the appropriate grade level.

To learn to use the program, students were provided with instructions and a trial (warm-up) task with immediate feedback. This instruction included: (1) a yellow bar at the top of the screen to show how far along they were on the test; (2) they had to click on the speaker icon to listen to the task instructions; (3) they had to click on the “next” button to move on to the next task; (4) pupils in Grades 1 and 2 received extra warm-up tasks to enhance keyboarding and mouse skills; and (5) after completing the last task, participants received immediate visual feedback with a display of 1 to 10 balloons, where the number of balloons was proportionate to their achievement.

The feedback system available for the teacher was more elaborate. Due to the large number of students and items, the Rasch analyses were run with the built-in analytic module in the eDia system. As the tasks in the item bank were scaled using IRT, it was possible to compare students’ achievement. Teachers received feedback on students’ achievement both as a percentage of correct items and as ability scores. For each grade and domain, the national average achievement (ability score) was set at 500 with a standard deviation of 100 (Carlson, 2009; Ferrão et al., 2015; Weeks, 2018). This was the point of reference for interpreting students’ achievement.

We used confirmatory factor analyses (CFA) within structural equation modeling (SEM) (Bollen, 1989) to test the underlying measurement models of reading, mathematics and science knowledge in the three dimensions of learning: psychological (reasoning), application (literacy), and disciplinary knowledge, respectively (RQ 1). We used the preferred estimator for categorical variables; the adjusted weighted least squares mean and variance (WLSMV) (Muthén and Muthén, 2012). We tested a 3-dimensional model to distinguish the three different dimensions of learning, and we also tested a 1-dimensional model with all three dimensions combined under one general factor. In order to test which model fitted the data better, we carried out a special χ2-difference test in Mplus. We also used CFA to test the underlying measurement model, and to determine the invariance behavior of the psychological dimension across the three domains of learning (RQ 2).

To establish a developmentally valid scale, we used the Rasch model with the vertical and horizontal scaling of the data (RQs 2 and 4) and then a linear transformation of the logit metric. As indicated above, for each domain and at each grade level, the mean achievement of each dimension was set to 500 with a standard deviation of 100. We used path models to test the effect and predictive power of school learning on the psychological dimension of learning (RQ 3).

Results

The Psychological Dimension of Learning

Results showed that the psychological (reasoning/thinking), application and disciplinary dimensions of learning can be distinguished empirically and are independent of domain and grade. The χ2-difference test in Mplus showed that the 3-dimensional model fitted significantly better than the 1-dimensional model in each grade and in each domain (see Tables 35 for reading, mathematics and science, respectively). Generally, the 3-dimensional measurement model for each domain showed a good model fit (Tables 35), based on Hu and Bentler’s (1999) recommended cut-off values. The comparative fit index (CFI) and the Tucker–Lewis index (TLI) values above 0.95 and the root mean square error of approximation (RMSEA) below 0.06 indicated a good global model fit.

TABLE 3
www.frontiersin.org

Table 3. Goodness of fit indices for testing the dimensionality of reading from Grades 1 to 8.

TABLE 4
www.frontiersin.org

Table 4. Goodness of fit indices for testing the dimensionality of mathematics from Grades 1 to 8.

TABLE 5
www.frontiersin.org

Table 5. Goodness of fit indices for testing the dimensionality of science from Grades 1 to 8.

In most cases, the 3-dimensional models fitted the data significantly better than that the 1-dimensional models. In some cases, mostly in Grades 7 and 8, the 3-dimensional model fit indices were lower. This could have been because the tasks were originally developed for students in lower grades.

The fit indices dropped in the case of mathematics and science in Grade 8 but were significantly higher than that of the 1-dimensional model. Thus, the psychological, application and disciplinary dimensions of learning could be distinguished. The psychological dimension of learning could be made visible independently of the measured domain in everyday educational settings, thus supporting Hypothesis 1.

The Psychological Dimension of
Learning Across Domains

The bivariate correlations of the psychological dimensions between pairs of domains (mathematics and reading, mathematics and science, and reading and science) ranged from 0.29 to 0.49 and were statistically significant (Table 6). At each grade level, the correlations of the psychological dimension (reasoning/thinking) tended to be the highest between mathematics and reading and lowest between mathematics and science. The strongest set of correlations, independent of the measured domain, was found in Grade 8, indicating that the psychological dimension of learning in reading, mathematics and science were highly correlated, but not identical constructs.

TABLE 6
www.frontiersin.org

Table 6. Correlations of the psychological dimension between pairs of domains from Grades 1 to 8.

The invariance in the psychological dimension of learning across the three domains was supported by comparing the 3-dimensional measurement model, which distinguishes the psychological dimension of reading, mathematics and science, and the 1-dimensional measurement model, which combines the psychological dimension of the different learning domains under a single factor. The special χ2-difference test in Mplus showed that the 3-dimensional model fitted significantly better at each grade level than the 1-dimensional model (Table 7).

TABLE 7
www.frontiersin.org

Table 7. Goodness of fit indices for testing the dimensionality of the psychological dimension in reading, mathematics, and science using 1- and 3-dimensional models for Grades 1 to 8.

The Rate of Development in the Psychological Dimension

Figure 3 presents the mean cognitive development scale scores in the psychological dimension of learning reading, mathematics and science. Please note that in each domain, the mean score of Grade 8 students was set at 500 with a standard deviation of 100, thereby constructing the point of reference for interpreting students’ achievement. This means that we cannot compare the development of the psychological dimension of learning across domains, but we can compare the rate of development.

FIGURE 3
www.frontiersin.org

Figure 3. The speed of the cognitive development in the psychological dimension of learning within the domains of mathematics, science and reading (Please note, that in each measured domain the mean of the 8th graders’ achievement was artificially set to 500 with a standard deviation of 100).

We found that the amount and rate of cognitive development were almost the same in each domain between Grades 6 and 8 and that there was no appreciable development in reading and science between Grades 2 and 6. The greatest rate of progress occurred in Grade 1 in reading and science, but not mathematics. Generally, there was a steady increase in the psychological dimension of learning in mathematics, especially in the first 4 years of schooling. The results confirmed our hypothesis that children’s cognitive development is slow (Molnár et al., 2013; Molnár et al., 2017), thus indicating the importance of the explicit development in this dimension in school lessons. Overall, these results highlighted the importance, sensitivity and potential of the development of thinking skills in the early years of schooling.

Relationship Between the Three Dimension of Learning

The possibility and practical relevance of separating the psychological dimension of learning can be explored from another perspective by examining the proportion of its variance that remains unexplained if the more readily visible disciplinary and application dimensions (referred to together as school knowledge) are taken into account. Technically, these dimensions may be considered as potential predictors of the psychological dimension.

We used continuous factor indicators in SEM analyses to examine the relationships between school knowledge and the psychological dimension of learning in each domain. School knowledge as a latent factor was specified as the application and disciplinary dimensions of learning. According to the results, school knowledge predicted the psychological dimension of learning in all domains, but a significant amount of variance remained unexplained (see Figures 46). This indicates that existing aspects of the psychological dimension of learning can be separated from school knowledge as measured by the disciplinary and application parts of students’ knowledge. That is, it is relevant to measure the psychological dimension of learning in addition to measuring the disciplinary and application dimensions of learning. So our hypothesis was confirmed.

FIGURE 4
www.frontiersin.org

Figure 4. A structural model of mathematics school knowledge as a predictor of students’ cognitive development in the domain of mathematical reasoning.

FIGURE 5
www.frontiersin.org

Figure 5. A structural model of science school knowledge as a predictor of students’ cognitive development in the domain of scientific reasoning.

FIGURE 6
www.frontiersin.org

Figure 6. A structural model of reading school knowledge as a predictor of students’ cognitive development in the domain of the psychological dimension of reading.

The amount of explained variance was statistically significant and almost the same for mathematics and reading and somewhat higher for science. This suggests that there may be more common reasoning aspects in the three dimensions of science. The model for each domain fitted well (CFI = 1.000, TLI = 1.000, RMSEA = 0.000).

Gender Difference in the Psychological Dimension of Learning

In the present study, girls outperformed boys in the psychological dimension of learning in reading, mathematics and science (Mathematics: F = 0.272, t = -6.696, p < 0.001; Science: F = 3.578, t = -11.525, p < 0.001; Reading: F = 3.224, t = -4,370, p < 0.001); however, this varied by grade level (see Table 8). The largest, statistically significant differences in favor of girls were found in Grades 4 and 5, where girls outperformed boys in all three domains, and in Grades 6 to 8, where girls outperformed boys in two of the three domains. Girls also outperformed boys in reading in Grades 3–8, in mathematics in Grades 1 and 4–6, and in science in Grades 1, 4, 5, 7, and 8.

TABLE 8
www.frontiersin.org

Table 8. Gender differences in the psychological dimension of learning in reading, mathematics and science in Grades 1 to 8.

In this section, we examine gender differences among Grade 8 students – the grade level of students in PISA, TIMSS and our study. The results confirm our hypotheses that an assessment which focuses on students’ disciplinary knowledge or application does not replace an assessment of the psychological dimension of learning. In the case of mathematics, no gender differences were detected in the application and psychological dimensions of learning, but girls scored significantly higher, on average, than boys in the disciplinary dimension of learning. The results were different in the case of science. There were no gender differences in the application dimension of science learning. Boys achieved significantly higher in the psychological dimension.

Discussion

Previous research has already identified several characteristics of learning reading, mathematics and science. However, it has mainly focussed on only one dimension; either the disciplinary dimension or the application dimension of learning, and on the reading, mathematics and science learning of older students. There have been significant attempts to concentrate on the application and reasoning dimensions, but educational practice has mostly focussed on the assessment of the content of individual curriculum subjects. The application of knowledge has seldom been assessed, although the PISA assessments have highlighted its importance. Because of the lack of easy-to-use assessments, the psychological dimension of learning (cognitive development and reasoning) remains hidden. Therefore, neither the students nor their teachers receive feedback on level or development in this dimension. This study provides evidence that the psychological dimension of learning can be made visible and that technology-based assessment may be applied in an everyday educational context. This evidence highlights the importance of the assessment and the explicit development of the psychological dimension of learning in a school context. Further, it points to gender differences in the developmental rate of the psychological dimension of learning in favor of girls, although this varies by grade and domain.

Results support our hypotheses that the three dimensions of learning can be distinguished empirically and can be assessed separately. The 3-dimensional frameworks derived from previous research, including international comparative studies (Csapó and Szendrei, 2011; Csapó and Csépe, 2012; Csapó and Szabó, 2012), showed relatively good validity, and the results from the current analyses confirmed that they may form evidence-based foundations for diagnostic assessment. The most important findings from these analyses was that the psychological dimension of learning can be measured at the primary school level in the context of three of the most important domains of learning – reading, mathematics and science.

The present results also confirmed that, although the roots of the psychological development of different domains are universal and the domains of learning build on each other (Molnár and Csapó, 2019), there are still significant developmental differences between them. While there is a close connection between the development of early literacy and numeracy, and later mathematics learning builds on reading, and science builds on both (McKeachie, 1987), our results support the notion that the transfer is not obvious between the different domain contexts. There were statistically significant correlations between the development scores in the psychological dimension of reading, mathematics and science learning, but they were not identical constructs.

Previous studies have indicated that children’s cognitive development is slow (Molnár et al., 2013, 2017) but that it can be taught effectively (de Koning et al., 2002; Klauer and Phye, 2008; Perret, 2015). Our results confirmed both of these notions as there was no appreciable development in the psychological dimension of learning in reading and science for students in Grades 2–6, and students’ cognitive development was the most steady (and effective) in mathematics, where the greatest development took place in the first years of schooling. This confirms previous research findings and highlights the potential of developing thinking skills in the early years of schooling.

The results of the SEM indicated the complex nature of learning in reading, mathematics and science. An examination of the predictive power of school knowledge on the psychological dimension of learning showed that the disciplinary and application dimensions of learning together predicted the psychological dimension of learning at a moderate, but statistically significant level, while a significant amount of variance remained unexplained. This indicates that school knowledge in reading, mathematics, and science can contribute to the development of the psychological dimension of learning and can stimulate students’ general cognitive development, but the transfer effect may not be high. The results suggest that aspects of the psychological dimension of learning exist and can be separated from the learning dimensions assessed most often at school and in international comparative studies. This highlights the importance and relevance of developing measures of the psychological dimension of learning as well.

To provide context to interpret the size of the gender difference in the psychological dimension of learning in reading, mathematics and science, we compared our results to findings on gender differences in the most prominent international comparative studies. The gender differences in the international studies at Grade 4 and 8 were found in our study. We found gender differences in reading over almost all the primary school grade levels, including Grades 4 and 8, indicating that girls perform better in reading, irrespective of the dimension of learning.

Limitations of the Present Study

As the PISA 2015, TIMSS 2015, and PIRLS 2016 studies have also indicated, there are large differences between countries not only on the level of reading, mathematics and science performance, but also in gender differences. Therefore, results found in one country cannot be generalized across countries and cultures. Although general trends have been found, the generalizability of the results may be limited. The method we applied in this study was generalizable and may be useful for making the psychological dimension of learning visible in any educational context. A further limitation of the study could be the results of the “common method bias” and “test motivation” as possible sources of shared variance across tests and domains. Participation in the study was voluntary, and although the large sample sizes and the diversity of the schools made the results sufficiently robust, the actual samples were not nationally representative. Thus, the present study does not provide a complete picture of the Hungarian education system. Nevertheless, the analyses did reveal some generalizable trends.

Conclusion

The 3-dimensional frameworks for the diagnostic assessment used in the present study were devised on the basis of current results from a number of research fields ranging from cognitive neuroscience to research on cognitive development, standard setting and the theoretical frameworks of large-scale international comparative studies. The item banks for assessing reading, mathematics and science were developed through the careful mapping of assessment tasks onto frameworks. The next step in scientifically establishing and further developing the diagnostic system is to empirically validate the 3-dimensional framework. We first presented the results of the comprehensive analyses in this study. In the present analyses, we focused on the psychological dimension of learning, which determines the dimensions of disciplinary knowledge and application, but is less visible or observable in the school context.

The results confirmed the theoretical foundations of the project and made clear that the psychological dimension can be distinguished and measured in the context of the most important domains of learning in the beginning phase of schooling. These findings indicate directions for further research as well. Item development for this study was based on the theoretical frameworks without empirical evidence of dimensionality. Based on the empirical confirmation of the three dimensions in this study, the validity of the assessment scales constructed from the item banks, may be improved by exploring how well the items fit particular scales.

Establishing scales empirically to assess the psychological dimension of learning paves the way to improving learning as well. The evidence that cognitive development is measurable provides a basis for large-scale systematic diagnostic monitoring of the development of students’ thinking skills, one of the most sorely lacking elements in the current spectrum of assessment practices. It also supports different types of intervention studies from teacher-initiated practical improvements to well-controlled, randomized experiments.

Ethics Statement

The authors only had access to anonymized data, and hence an ethics approval and parental consent were not required as per applicable institutional and national guidelines and regulations. The assessment data collected for this study formed integrated parts of the normal educational processes of the participating schools. The coding system for the online platform masked students’ identity, the researchers would thus have been unable to tie the data to the students. The results from the low-stakes diagnostic assessments were only disclosed to the participating students (as immediate feedback) and to their teachers. Because of the anonymity and low-stakes testing design of the assessment process, it was not required or possible to request and obtain written informed parental consent from the participants.

Author Contributions

GM and BC took responsibility for the content, including participation in the concept, design, analysis, drafting the manuscript, writing and final approval of the manuscript, and agreed to be accountable for all aspects of the study.

Funding

This study was funded by OTKA K115497 and EFOP 3.2.15.

Conflict of Interest Statement

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Acknowledgments

We would like to express our appreciation to the members of the eDia Research Group, who provided continuous support for the participating schools during data collection. A special thanks go to Dr. Géza Makay and Dóra Mokri, who assisted us with data filtering and scaling. We also wish to thank the two the reviewers for their valuable comments on earlier versions of this manuscript.

References

Adey, P. (2007). “The CASE for a general factor in intelligence,” in Integrating the Mind: Domain General Versus Domain Specific Processes in Higher Cognition, ed. M. J. Roberts (Hove: Psychology Press), 369–385.

Google Scholar

Adey, P., and Csapó, B. (2012). “Developing and assessing scientific reasoning,” in Framework for Diagnostic Assessment of Science, eds B. Csapó and G. Szabó (Budapest: Nemzeti Tankönyvkiadó), 17–53.

Google Scholar

Adey, P., Csapó, B., Demetriou, A., Hautamaki, J., and Shayer, M. (2007). Can we be intelligent about intelligence? Why education needs the concept of plastic general ability. Educ. Res. Rev. 2, 75–97.

Google Scholar

Adey, P., and Shayer, M. (1994). Really Raising Standards: Cognitive Intervention and Academic Achievement. London: Routledge.

Google Scholar

Adey, P., Shayer, M., and Yates, C. (2001). Thinking Science: The Curriculum Materials of the CASE Project, 3rd Edn. London: Nelson Thornes.

Google Scholar

Alexander, P. A., and Murphy, P. K. (1999). Nurturing the seeds of transfer: a domain-specific perspective. Int. J. Educ. Res. 31, 561–576. doi: 10.1016/s0883-0355(99)00024-5

CrossRef Full Text | Google Scholar

Binet, A., and Simon, T. (1916). “The development of intelligence in children: The Binet–Simon Scale,” in The Development of Intelligence in Children, eds E. S. Kite Trans. (Vineland NJ: Publications of the Training School).

Google Scholar

Blomert, L., and Csépe, V. (2012). “Psychological foundations of reading acquisition and assessment,” in Framework for Diagnostic Assessment of Reading, eds B. Csapó1 and V. Csépe (Budapest: Nemzeti Tankönyvkiadó), 17–77.

Google Scholar

Bloom, B. S. (1969). Cross-National Study of Educational Attainment: Stage I of the I.E.A. Investigation in Six Subject Areas, Vol. 1–2. Washington, DC: Office of Education (DHEW).

Google Scholar

Bollen, K. A. (1989). Structural Equations with Latent Variables. New York, NY: Wiley.

Google Scholar

Bransford, J. D., and Schwartz, D. L. (1999). Rethinking transfer: a simple proposal with multiple implications. Rev. Res. Educ. 24, 61–100. doi: 10.3102/0091732x024001061

CrossRef Full Text | Google Scholar

Burstein, L. (1993). The IEA Study. of Mathematics III: Student Growth and Classroom Processes. Oxford: Pergamon Press.

Google Scholar

Carlson, J. E. (2009). “Statistical models for vertical linking,” in Statistical Models for Test Equating, Scaling, and Linking, ed. A. A. von Davier (New York, NY: Springer), 59–70. doi: 10.1007/978-0-387-98138-3_4

CrossRef Full Text | Google Scholar

Carroll, J. B. (1993). Human Cognitive Abilities. Cambridge: Cambridge University Press.

Google Scholar

Costa, A., and Kallick, B. (2009). Learning and Leading with Habits of Mind: 16 Characteristics of Success. Alexandria, VA: Association for Supervision and Curriculum Development.

Google Scholar

Csapó, B. (1997). The development of inductive reasoning: cross-sectional assessments in an educational context. Int. J. Behav. Dev. 20, 609–626. doi: 10.1080/016502597385081

CrossRef Full Text | Google Scholar

Csapó, B. (2010). Goals of learning and the organization of knowledge. Zeitschrift für Pädagogik 2010(Suppl. 56), 12–27.

Google Scholar

Csapó, B., and Csépe, V. (2012). Framework for Diagnostic Assessment of Reading. Budapest: Nemzeti Tankönyvkiadó.

Google Scholar

Csapó, B., Csíkos, C., and Molnár, G. (2015a). A Matematikai Tudás Online Diagnosztikus Értékelésének Tartalmi Keretei [Framework for Online Diagnostic Assessment of Mathematics]. Budapest: Nemzeti Tankönyvkiadó.

Google Scholar

Csapó, B., Korom, E., and Molnár, G. (2015b). A Természettudományi Tudás Online Diagnosztikus Értékelésének Tartalmi Keretei [Framework for Online Diagnostic Assessment of science]. Budapest: Nemzeti Tankönyvkiadó.

Google Scholar

Csapó, B., Steklács, J., and Molnár, G. (2015c). Az Olvasás-Szövegértés Online Diagnosztikus Értékelésének Tartalmi Keretei [Framework for Online Diagnostic Assessment of Reading]. Budapest: Nemzeti Tankönyvkiadó.

Google Scholar

Csapó, B., and Szabó, G. (2012). Framework for Diagnostic Assessment of Science. Budapest: Nemzeti Tankönyvkiadó.

Google Scholar

Csapó, B., and Szendrei, M. (2011). Framework for Diagnostic Assessment of Mathematics. Budapest: Nemzeti Tankönyvkiadó.

Google Scholar

de Koning, E. (2000). Inductive Reasoning in Primary Education: MEASUREMENT, Teaching, Transfer. Zeist: Kerckebosch.

Google Scholar

de Koning, E., Hamers, J. H. M., Sijtsma, K., and Vermeer, A. (2002). Teaching and transfer of inductive reasoning in primary education. Dev. Rev. 22, 211–241. doi: 10.1080/0361073X.2017.1398802

PubMed Abstract | CrossRef Full Text | Google Scholar

Demetriou, A., Efklides, A., Platsidou, M., and Campbell, R. L. (1993). The architecture and dynamics of a developing mind: experiential structuralism as a frame for unifying cognitive developmental theories. Monogr. Soc. Res. Child Dev. 58:167.

PubMed Abstract | Google Scholar

Demetriou, A., Gustafsson, J. E., Efklides, A., and Platsidou, M. (1992). “Structural systems in developing cognition, science, and education,” in Neo-Piagetian Theories of Cognitive Development: Implications and Applications for Education, eds A. Demetriou, M. Shayer, and A. Efklides (London: Routledge), 79–103.

Google Scholar

Ferrão, M. E., Costa, P. M., and Oliveira, P. N. (2015). Generalized partial credit item response model: linking scales in the assessment of learning. J. Interdisciplinary Mathematics 18, 339–354. doi: 10.1080/09720502.2014.932119

CrossRef Full Text | Google Scholar

Feuerstein, R., Rand, Y., Hoffman, M., and Miller, M. (1980). Instrumental Enrichment: An Intervention Program for Cognitive Modifiability. Baltimore: University Park Press.

Google Scholar

Geake, J., and Cooper, P. (2003). Cognitive neuroscience: implications for education? Westmin. Stud. Educ. 26, 7–20. doi: 10.1080/0140672032000070710

CrossRef Full Text | Google Scholar

Greiff, S., Wüstenberg, S., Molnár, G., Fischer, A., Funke, J., and Csapó, B. (2013). Complex problem solving in educational contexts – something beyond g: concept, assessment, measurement invariance, and construct validity. J. Educ. Psychol. 105, 364–379. doi: 10.1037/a0031856

CrossRef Full Text | Google Scholar

Griffin, P., and Care, E. (2014). Assessment and Teaching of 21st century skills: Methods and (approach). Dordrecht: Springer.

Google Scholar

Grissmer, D. W. (2000). The continuing use and misuse of SAT scores. Psychol. Public Policy Law 6, 223–232. doi: 10.1037/1076-8971.6.1.223

CrossRef Full Text | Google Scholar

Hattie, J. (2009). Visible Learning: A Synthesis of 800 Meta-Analyses Relating to Achievement. New York, NY: Routledge.

Google Scholar

Hattie, J. (2012). Visible Learning for Teachers: Maximizing Impact on Learning. New York, NY: Routledge.

Google Scholar

Hattie, J., and Anderman, E. M. (2013). “Introduction,” in International Guide to Student Achievement, eds J. Hattie and E. M. Anderman (New York, NY: Routledge).

Google Scholar

Hu, L., and Bentler, P. M. (1999). Cutoff criteria for fit indexes in covariance structure analysis: conventional criteria versus new alternatives. Struct. Equ. Model. 6, 1–55. doi: 10.1080/10705519909540118

CrossRef Full Text | Google Scholar

Husén, T. (1967). International. Study of Achievement in Mathematics: A Comparison of Twelve Countries, Vol. 1–2. Stockholm: Almqvist & Wiksell.

Google Scholar

IEA (1988). Science Achievement in Seventeen Countries: A Preliminary Report. Oxford: Pergamon Press.

Google Scholar

Inhelder, B., and Piaget, J. (1958). The Growth of Logical Thinking. London: Routledge and Kegan Paul.

Google Scholar

Klauer, K. J. (1989a). Denktraining für Kinder I. Göttingen: Hogrefe.

Google Scholar

Klauer, K. J. (1989b). Teaching for analogical transfer as a means of improving problem solving, thinking and learning. Instruct. Sci. 18, 179–192. doi: 10.1007/bf00053358

CrossRef Full Text | Google Scholar

Klauer, K. J. (1991). Denktraining für Kinder II. Göttingen: Hogrefe.

Google Scholar

Klauer, K. J. (1993). Denktraining für Jugendliche. Göttingen: Hogrefe.

Google Scholar

Klauer, K. J. (1997). “Training inductive reasoning: A developmental programme of higher-order cognitive skills,” in Teaching Thinking in Europe. Inventory of European programmes, eds J. H. M. Hamers and M. T. Overtoom (Utrecht: Sardes), 77–81.

Google Scholar

Klauer, K. J., and Phye, G. D. (2008). Inductive reasoning: a training approach. Rev. Educ. Res. 78, 85–123. doi: 10.3102/0034654307313402

CrossRef Full Text | Google Scholar

MacGilchrist, B., Reed, J., and Myers, K. (2004). The Intelligent School. London: Sage.

Google Scholar

Martin, M. O., Mullis, I. V. S., Foy, P., and Hooper, M. (2016). TIMSS 2015 International Results in Science. Chestnut Hill, MA: TIMSS & PIRLS International Study.

Google Scholar

Mayrath, M. C., Clarke Midura, J., Robinson, D. H., and Schraw, G. (eds) (2012). Technology Based Assessments for 21st Century Skills: Theoretical and Practical Implications from Modern Research. Charlotte, NC: IAP.

Google Scholar

McKeachie, W. J. (1987). Cognitive skills and their transfer: discussion. Int. J. Educ. Res. 11, 707–712. doi: 10.1016/0883-0355(87)90010-3

CrossRef Full Text | Google Scholar

Molnár, G., and Csapó, B. (2011). Az 1–11 évfolyamot átfogó induktív gondolkodás kompetenciaskála készítése a valószínûségi tesztelmélet alkalmazásával. [Constructing inductive reasoning competency scales for years 1–11 using IRT models]. Magyar Pedagógia 111, 127–140.

Google Scholar

Molnár, G., and Csapó, B. (2019). “Technology-based diagnostic assessments for identifying early learning difficulties in mathematics,” in International Handbook of Mathematical Learning Difficulties, eds A. Fritz-Stratmann, P. Räsänen, and V. Haase (Heidelberg: Springer), 683–707. doi: 10.1007/978-3-319-97148-3_40

CrossRef Full Text | Google Scholar

Molnár, G., Greiff, S., and Csapó, B. (2013). Inductive reasoning, domain-specific and complex problem solving: relations and development. Think. Skills Creat. 9, 35–45. doi: 10.1016/j.tsc.2013.03.002

CrossRef Full Text | Google Scholar

Molnár, G., Greiff, S., Wüstenberg, S., and Fischer, A. (2017). “Empirical study of computer-based assessment of domain-general dynamic problem solving skills,” in The Nature of Problem Solving, eds B. Csapó and J. Funke (Paris: OECD), 123–143.

Google Scholar

Mullis, I. V. S., and Martin, M. O. (eds) (2015). PIRLS 2016 Assessment Framework, 2nd Edn. Boston, MA: TIMSS & PIRLS International Study Center, Boston College.

Google Scholar

Mullis, I. V. S., Martin, M. O., Foy, P., and Hooper, M. (2016). TIMSS 2015 International Results in Mathematics. Boston, MA: TIMSS & PIRLS International Study Center, Boston College.

Google Scholar

Mullis, I. V. S., Martin, M. O., Foy, P., and Hooper, M. (2017). PIRLS 2016 International Results in Reading. Boston, MA: TIMSS & PIRLS International Study Center, Boston College.

Google Scholar

Mullis, I. V. S., Martin, M. O., Ruddock, G. J., O’Sullivan, C. Y., Arora, A., and Eberber, E. (eds) (2005). TIMSS 2007 Assessment Frameworks. Boston, MA: TIMSS & PIRLS International Study Center, Boston College.

Google Scholar

Muthén, L. K., and Muthén, B. O. (2012). Mplus User’s Guide, 7th Edn. Los Angeles: Muthén & Muthén.

Google Scholar

Nguyen, T., Watts, T. W., Duncan, G. J., Clements, D. H., Sarama, J. S., Wolfe, C., et al. (2016). Which preschool mathematics competencies are most predictive of fifth-grade achievement? Early Child. Res. Q. 36, 550–560. doi: 10.1016/j.ecresq.2016.02.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Nunes, T., and Csapó, B. (2011). “Developing and assessing mathematical reasoning,” in Framework for Diagnostic Assessment of Mathematics, eds B. Csapó and M. Szendrei (Budapest: Nemzeti Tankönyvkiadó),17–56.

PubMed Abstract | Google Scholar

OECD (1999). Measuring Student Knowledge and Skills: A New Framework for Assessment. Paris: OECD.

Google Scholar

OECD (2003). The PISA 2003 Assessment Framework: Mathematics, Reading, Science and Problem Solving. Paris: OECD.

Google Scholar

OECD (2016). PISA 2015 Results: Excellence and Equity in Education, Vol. I. Paris: OECD.

Google Scholar

Perret, P. (2015). Children’s inductive reasoning: developmental and educational perspectives. J. Cogn. Educ. Psychol. 14, 389–408. doi: 10.1891/1945-8959.14.3.389

CrossRef Full Text | Google Scholar

Polya, G. (1981). Mathematics Discovery: On Understanding, Learning, and Teaching Problem Solving. New York, NY: John Wiley & Sons.

Google Scholar

Ritchhart, R., Church, M., and Morrison, K. (2011). Making Thinking Visible: How to Promote Engagement, Understanding, and Independence for all Learners. San Francisco: Jossey-Bass.

Google Scholar

Shayer, M. (1999). Cognitive acceleration through science education II: its effect and scope. Int. J. Sci. Educ. 21, 883–902. doi: 10.1080/095006999290345

CrossRef Full Text | Google Scholar

Shayer, M., and Adey, P. (eds) (2002). Learning Intelligence: Cognitive Acceleration Across the Curriculum From 5 to 15 years. Milton Keynes: Open University Press.

Google Scholar

Shayer, M., and Adhami, M. (2007). Fostering cognitive development through the context of mathematics: results of the CAME project. Educ. Stud. Math. 64, 265–291. doi: 10.1007/s10649-006-9037-1

CrossRef Full Text | Google Scholar

Thomas, M. S., Ansari, D., and Knowland, V. C. (2019). Annual research review: educational neuroscience: progress and prospects. J. Child Psychol. Psychiatry 60, 477–492. doi: 10.1111/jcpp.12973

PubMed Abstract | CrossRef Full Text | Google Scholar

Trilling, B., and Fadel, C. (2009). 21st Century Skills: Learning for Life in our Times. San Francisco: John Wiley & Sons.

Google Scholar

Vygotsky, L. S. (1978). Mind in Society: The Development of Higher Psychological Processes. Cambridge, MA: Harvard University Press.

Google Scholar

Weeks, J. P. (2018). An application of multidimensional vertical scaling. Measure. Interdiscipl. Res. Perspect. 16, 139–154. doi: 10.1080/15366367.2018.1502005

CrossRef Full Text | Google Scholar

Wüstenberg, S., Greiff, S., Molnar, G., and Funke, J. (2014). Determinants of cross-national gender differences in complex problem solving competency. Learn. Individ. Diff. 29, 18–29. doi: 10.1016/j.lindif.2013.10.006

CrossRef Full Text | Google Scholar

Keywords: technology-based assessment, online assessment, assessment for learning, visible learning, cognitive development

Citation: Molnár G and Csapó B (2019) Making the Psychological Dimension of Learning Visible: Using Technology-Based Assessment to Monitor Students’ Cognitive Development. Front. Psychol. 10:1368. doi: 10.3389/fpsyg.2019.01368

Received: 07 December 2018; Accepted: 27 May 2019;
Published: 10 June 2019.

Edited by:

Ronny Scherer, University of Oslo, Norway

Reviewed by:

Alison Margaret Gilmore, University of Otago, New Zealand
Andreas Rausch, Universität Mannheim, Germany

Copyright © 2019 Molnár and Csapó. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Gyöngyvér Molnár, Z3ltb2xuYXJAZWRwc3kudS1zemVnZWQuaHU=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.