- School of Education Science, Nanjing Normal University, Nanjing, China
The rapid development of computers and technology affects modern daily life. Individuals in the digital age need to develop computational thinking (CT) skills. Existing studies have shown that programming teaching is conducive to cultivating students’ CT, and various learning models have different effects on the cultivation of CT. This study proposed a problem-oriented learning (POL) model that is closely related to programming and computational thinking. In all, 60 eighth-grade students from a middle school in China were divided into an experimental group (EG) which adopted the POL model, and a control group (CG) which adopted the lecture-and-practice (LAP) learning model. The results showed that the students who were instructed using the POL model performed better than those who were instructed using the LAP model on CT concepts, CT practices, and CT perspectives. Significant differences were found for CT concepts and CT perspectives, but not for CT practices. Findings have implications for teachers who wish to apply new learning models to facilitate students’ CT skills, and the study provides a reference case for CT training and Python programming teaching.
Introduction
The younger generation interacts frequently with technologies that permeate all aspects of their lives on a daily basis (Baruch and Erstad, 2018), and they are increasingly expected to be not only consumers but also producers of technology (Kong et al., 2020). In the digital age, computational thinking (CT) can develop students’ abilities of critical thinking, creative thinking, and problem solving (Ananiadou and Claro, 2009; Mishra and Yadav, 2013; Repenning et al., 2015). CT, just like reading, writing, and arithmetic, is a basic skill for all students (Wing, 2006). According to Bundy (2007), CT influences the research of almost all disciplines in natural science and human science. Many researchers consider that CT should be integrated into the formal education system as a learning objective to cultivate students’ ability to guide their future lives (Grover and Pea, 2013).
Recently CT related to programming has been included in K-12 courses around the world (Shute et al., 2017; Hsu et al., 2018; Sands et al., 2018). There is a consensus that students’ CT can be nurtured via programming education (Rich et al., 2017; Nouri et al., 2020). Most studies use visual programming tools, such as App Inventor and Scratch, which are closer to the representation of human language, helping students concentrate on the logic and structure, and become involved in programming instead of being anxious about the difficulties of writing programs (Kelleher and Pausch, 2005). However, few studies have used Python. Wen et al. (2014) pointed out that Python is suitable for cultivating students’ CT and problem-solving skills. Python has been found to be a means of helping learners develop skills to face real-world problems (Tang et al., 2020). Lee and Cho (2017) also mentioned that Python is a programming language that is of interest to beginners and is easy to learn. It is also used as an intermediate language for connecting modules written in other languages. At the same time, Lee and Cho (2017) found that teaching methods that use LAP learning, collaborative learning, game-based learning, and design-based learning environments are often used to cultivate students’ CT. LAP learning can help learners complete tasks in a limited amount of time, but this choice is a negative factor for students’ creativity. Pair programming is a form of collaborative learning. Studies have shown that pair programming has a positive impact on friend relationships, but has no effect on non-friend relationships (Werner et al., 2013). In game-based learning environments, the motivation to complete a game may adversely affect the learning of a certain level, thereby affecting learning from playing the game (Israel-Fishelson and Hershkovitz, 2019). Unplugged programming activities allow students to participate in computer science practice without using digital equipment, which can solve the limitations of computer hardware equipment and students’ lack of early programming knowledge, but this method is more suitable for the early stages of elementary education (Sun et al., 2021). Design-based learning (DBL) activities, such as interactive web design and digital storytelling, allow students to use multiple technologies to conduct activities, which imposes heavy curriculum burdens on teachers and adversely affects the entire process (Saritepeci, 2020). This study took a Python programming course as the context to construct a problem-oriented learning (POL) model to effectively promote students’ CT skills.
Literature Review
Computational Thinking
Computational thinking is a method of designing systems, solving problems, and understanding human behavior (Wing, 2006). It includes engineering and design thinking (effective solution developing), mathematical thinking to solve various problems, and system thinking (system understanding and modeling). Abstraction, decomposition, algorithms, and debugging are the CT components that most frequently arise in the literature (Shute et al., 2017). The International Society for Technology in Education (ISTE) defined CT as the common skills of algorithmic thinking, creativity, critical thinking, cooperative thinking, problem solving, and communication skills (ISTE, 2015). The goal of developing CT is not as a replacement for creative thinking, critical thinking or other kinds of thinking skills, but rather to increase the skills of using computers and algorithms to solve problems (Wing, 2011; Furber, 2012).
Computational thinking is a basic skill for all people (Wing, 2006), and will be used everywhere (Wing, 2008). Educational researchers have been actively seeking innovative methods and ways to incorporate CT into the curriculum and encourage students to participate in CT (Bower et al., 2017). They have attempted to teach and develop the knowledge and skills of CT in different educational situations by various means including programming (Basu et al., 2017; Bati, 2021), educational robotics (Chevalier et al., 2020; Qu and Fok, 2021), unplugged activities (Kuo and Hsu, 2020; Huang and Looi, 2021), games/simulations (Danial et al., 2021; Hooshyar et al., 2021), storytelling (Soleimani et al., 2019; Parsazadeh et al., 2020), and so forth. These tools become “technical partners in the learning process” (Jonassen et al., 2012), and the rationale for improving CT skills in each of these tools emphasizes various CT components (Shute et al., 2017). In addition, existing research has developed CT interventions in different disciplines such as physics and biology (Sengupta et al., 2013), expository writing and journalism (Wolz et al., 2011), mathematics (Wilkerson-Jerde, 2014), science in general (Weintrop et al., 2016; Basu et al., 2017), and science and arts (Sáez-López et al., 2016).
Previous studies, however, preferred to apply visual programming languages, while paying less attention to text programming languages. Nonetheless, visual programming is not as reliable as text programming, and its functions are not as good as the latter. Deng et al. (2020) pointed out that visual programming requires learners to focus on a great number of grammar rules, and the programming foundation is needed to develop CT. Thus, visual programming alone may not be sufficient for students to understand the true meaning of programming and to master CT, especially for beginners. Text programming can compensate for visual programming based on program functionality (Weintrop and Wilensky, 2015). In the era of artificial intelligence, Python has become the preferred development language for artificial intelligence applications (Okonkow and Ade-Ibijola, 2021). Compared with other text programming languages, Python is closer to human languages as it conforms to people’s thinking habits. It can reduce unnecessary grammar learning, thereby reducing cognitive load and allowing students to focus on solving programming problems (Maria and Tsiatsos, 2018). Kim et al. (2019) developed a data visualization education program, and the sixth-grade students received 6 days and 36 h of training. It was found that Python for data visualization education can effectively improve the CT of sixth graders, including their computational cognition, fluency, originality, and elaboration. Another study used the Python language to develop a learning program and model, and applied it to the 10-h learning of sixth-grade students. It was found that the robot-based Python learning model had a significant effect on improving students’ thinking skills, which confirms the applicability of the text-based programming language to elementary school students. Other studies (e.g., García Monsálvez, 2017; Lee and Cho, 2017; Maria and Tsiatsos, 2018) also found that Python programming education had a positive impact on students’ CT.
At present, there is no common definition of CT. Therefore, the evaluation methods of CT are very diverse. A CT questionnaire based on the five CT factors proposed by ISTE was designed by Durak and Saritepeci (2018), namely algorithmic thinking, creativity, problem solving, cooperation, critical thinking, and communication. Brennan and Resnick (2012) proposed a three-dimensional CT framework for visual programming using Scratch, and pointed out that the framework could be transferred to other programming teaching practices such as Logo programming (Lye and Koh, 2014). The framework consists of three dimensions: CT concepts (the concepts which designers use when they are programming), CT practices (the practices that designers develop while they are programming), and CT perspectives (the perspectives on the world and themselves that designers form).
Learning Models of Programming Teaching
An increasing amount of attention is being paid to programming teaching. Garneli et al. (2015) pointed out that game design, robotics teaching, project-based learning, and collaborative learning are becoming increasingly popular in programming teaching. The teaching intervention of Vihavainen et al. (2014) included collaboration and peer support. Florez et al. (2017) further pointed out the importance of using visualization tools to help students develop programming concepts. Scherer et al. (2020) also found from meta-analysis that visualization had a moderate effect on programming learning, whereas physicality had a large effect.
Many teaching methods have been explored and applied to improve students’ programming skills. For example, Olelewe and Agomuo (2016) discussed the influence of two teaching methods on programming learning. The results showed that the B-learning model (the combination of e-learning and traditional face to face learning) could improve students’ programming language performance more effectively than the traditional face-to-face model. Researchers have found that pair programming can improve personal programming skills, programming efficiency and quality (Zhong et al., 2017), while also increasing self-confidence in learning (Lai and Xin, 2011). Corral et al. (2014) found that game-oriented methods based on interaction with tangible user interfaces could improve students’ motivation and academic performance. Uysal (2014) indicated that problem-solving instructional methods can effectively improve students’ academic performance and problem perception. Early studies adopted the problem-solving approach to give students the detailed steps and sequences of behaviors to solve the problems they encountered while they were coding (Scherer et al., 2020).
Scherer et al. (2020) conducted a meta-analysis of some instructional approaches and found that blended learning (1.023) had the largest intervention effect, followed by game-based learning (0.821) and metacognitive strategies (0.658), and finally collaborative activities (0.560), problem solving instruction (0.518), and feedback strategies (0.436). At the same time, the effectiveness of teaching methods may vary depending on the research content and teaching conditions (Li and Ma, 2010), and an integration of various teaching methods should be more effective for tutoring programming (Vihavainen et al., 2014).
In recent decades, the study of K-12 programming teaching was mostly carried out in high schools. For example, He et al. (2014) studied robotics programming teaching of collaboration with a robotics club for high school students. However, scanty attention has been paid to investigating the model of programming teaching for middle school students. Children should start to learn programming at a much younger age to motivate their learning interest. This study describes a Python programming course for eighth graders. Although there are many innovative teaching methods, they have not been applied to specific programming courses and cannot guide programming teachers’ teaching well. Therefore, teachers often use LAP methods in the programming classroom (Kim and Yun, 2020). Teachers demonstrate the correct steps, and then students imitate to complete the task, which is not conducive to the cultivation of students’ problem-solving ability and computational thinking. Therefore, this study proposed the POL model to promote eighth graders’ CT.
The Problem-Oriented Learning Model
Polya (1957) proposed a four-step process to solve problems; this process is widely used around the world to help people with problem solving: understand the problem, devise a plan, carry out the plan, and look back. In other words, problem solving covers a series of processes. On this basis, Polya, Beichner (2002) developed GOAL-oriented problem solving for physics, which involves collecting information about the problem, figuring out an approach to the problem, analyzing the problem, and learning from one’s efforts. Evidence collected showed that this approach had a positive impact on students’ ability to solve problems. Kalelioğlu et al. (2016) proposed a framework of CT as a process of problem-solving, including identifying the problem; collecting, representing, and analyzing data; generating, choosing, and planning solutions; implementing solutions; and evaluating solutions and continuing for improvement. Kim and Yun (2020) proposed a learning model focused on CT skills, including problem identifying, analyzing, systematizing, and solving. These studies show that problem solving covers a series of processes which differ in specific teaching situations.
In the literature review, many problem-solving instruction approaches have been created in the fields of mathematics (e.g., Suarsana et al., 2019), science (e.g., Akben, 2020), and physics (e.g., Dewi et al., 2019), but few have been applied to Python courses. Most of the research has been to improve problem-solving skills (e.g., Cheng et al., 2018), but the impact on CT is still unclear. In fact, CT, programming skills, and problem solving are closely connected. Kalelioğlu et al. (2016) found that “abstraction,” “problem,” and “solving” are the most commonly used words in the definition of CT. Román-González et al. (2017) developed a scale to measure CT, and the results showed significant correlations with problem-solving skills. Based on the studies of Polya (1957), Beichner (2002), Kalelioğlu et al. (2016), and Kim and Yun (2020), this study formed a POL model including problem decomposition, problem abstraction, algorithmic representation, solution evaluation, and generalization and migration. Compared with previous studies, the framework of the POL model in this study designed five specific phases to solve programming problems, and it was more suitable for Python programming courses. Therefore, this study applied the POL model to Python programming courses to promote the development of students’ CT.
Research Questions
Based on the needs of CT training and its close relationship with programming and problem solving, this study focused on a middle school Python programming course to construct a POL model oriented toward programming problems that effectively cultivates students’ CT. Furthermore, it explored the effectiveness of the POL model for the cultivation of CT through experiments.
The questions this study aimed to address are: Compared with the LAP model, did the POL model significantly improve students’ CT concepts, CT practices, and CT perspectives?
Materials and Methods
Participants
A total of 60 eighth graders from a middle school in China participated in the study. In this school, the learning competence of students in different grade 8 classes is similar, due to the school having implemented the parallel classes method according to the students’ previous academic performance when allocating students to classes. We randomly selected two classes taught by the same teacher to participate in the study. One class was assigned as an experimental group (EG) and another as a control group (CG). There were 30 students with 17 males and 13 females in the EG, who received the guidance of the POL model. The CG was composed of 30 students with 15 males and 15 females, who learned via LAP learning model. Students in both groups had already taken a one-semester Python programming course. The results of the Python final test of the previous semester showed no significant difference between the two groups (t = 1.241, p = 0.220 > 0.05). It could therefore be considered that the two groups had the same initial learning level.
Participants in this study were involved on a voluntary basis and with the approval of their parents. To protect the participants, their personal information was hidden during the study. In addition, they could withdraw from the study at any time.
Instructional Design
Selection of Instructional Content
The course of Python programming covers a wide range of content, including basic input and output, branch selection, loops, and so on, some of which had been taught in the previous semester. This study selected the following three units as the instructional content: list, string and dictionary, and custom function (see Table 1).
The Problem-Oriented Learning Model
Combining the descriptions of the problem-solving process in related research (Kalelioğlu et al., 2016; Kim and Yun, 2020), this study divided the POL model into the following processes, as shown in Figure 1.
Student: First, problem decomposition. Decompose complex problems, extract the key information in the problem, and transform it into a problem that the student knows how to solve. Second, problem abstraction. Use digital language to express text information, abstract actual problems into mathematical problems and model them. Third, algorithmic representation. Express the logic of solving the problem in the Python programming language. Finally, solution evaluation. Use the Python programming language to test the effectiveness of the algorithm; evaluate and optimize the algorithm during continuous debugging. Step 5: generalization and migration. Generalize the solution of the problem and migrate it to other similar programming problem solutions.
Teacher: First, guide students to decompose questions and find out the key information in the problem. Second, guide students to convert text information into mathematical models. Third, provide a programming reference module when students express problem solutions with algorithms, and guide students to debug and optimize the algorithms. Finally, summarize the problem-solving process and give similar problems to promote the development of students’ transfer ability.
The Lecture-and-Practice Learning Model
The CG adopted the LAP learning model, as shown in Figure 2. Teacher: First, teach the problem-solving process, and then give timely guidance when students practice by themselves. Student: Listen carefully and record when the teacher teaches, and then practice according to the teacher’s steps.
The Difference Between the Two Learning Models
The LAP model is often used in programming teaching, whereas this study adopted the POL model in the Python programming course. The two mainly differ in terms of the teaching process and the behavior of teachers and students, as shown in Table 2. Take the “number of daffodils” as an example. In the LAP class, the teacher first analyzes the problem and explains it. For example, what is the number of daffodils? How can we find the number of daffodils? How can we implement it in Python? In the course of the explanation, the code is written gradually (Understand the problem). After the teacher explains, the students refer to the teacher’s code to solve the problem (Practice). When the students are confused, the teacher will guide or explain in groups. In contrast, in the POL class, the teacher first guides students to think about the problem to be solved (Decomposition problem), then guides them to think about the characteristics of daffodil numbers and how to find the ones, tens, and hundreds of a number (Abstract problem). Students try to implement it in Python. If they encounter difficulties, the teacher can prompt them appropriately. For example, the teacher can prompt the students to use the remainder operation (Algorithmic representation). After students finish writing the code, they will debug and think about how to optimize the algorithm (Evaluation solutions). Finally, the teacher gives a similar question, such as finding the number of lightning strikes, to promote students’ summary and migration (Generalization and migration).
Procedure
The experimental school has its own self-developed online program evaluation system. The platform integrates the Python language compiler and the programming question library. In addition, the platform provides online test and evaluation functions. After the students submitted the questions, the platform immediately gave feedback, including the scores obtained and the errors, which provided the students with the opportunity to continuously debug.
The experiment lasted for one semester, from February 2019 to June 2019, a total of 22 weeks, each of which was 40 min. Three units were taught during the semester. Before the instruction, the two groups completed the pre-test of CT perspectives scale. At the end of each of the three units, the corresponding unit test was carried out. During the learning activity, the students in the EG adopted the POL model and the CG adopted the LAP learning model. After completing the learning tasks of the three units, the students took the post-test of the CT perspectives scale, the Bebras test, and the computer-based final test. The experimental procedure is shown in Table 3.
Measurement
Some researchers have incorporated Brennan and Resnick’s (2012) framework into the evaluation of CT (Lye and Koh, 2014; Grover et al., 2015; Kong, 2019; Mouza et al., 2020). Combined with the teaching content and the characteristics of the Python programming course, this study modified their framework and formulated the CT evaluation for this study. Although many efforts have been made regarding CT evaluation (Basawapatna et al., 2011), it is still a challenge to evaluate CT learning in a programming environment. Survey with questionnaires is the most common way to measure CT attitudes or knowledge (e.g., Kim et al., 2013; Jun et al., 2014). Reflection is also often used in studies where students are asked to reflect on their programming experience (e.g., Zimmerman and Tsikalas, 2005; Yang, 2010). Other studies have tested the dimension of subject knowledge in learning achievement after integrating CT with disciplinary teaching (e.g., Sengupta et al., 2013). This study used the tools in Table 4 to measure the three dimensions of CT.
Computational Thinking Concepts
Brennan and Resnick (2012) defined CT concepts as including sequences, loops, parallelism, events, conditionals, operators, and data. Considering that this study was carried out in the context of a Python programming course, the two concepts of sequences and parallelism were not included in the CT concepts. In addition, the concept of events is rarely mentioned in Python programming, but the concept of functions is commonly used. Therefore, functions were used instead of events in this study (see Table 5).
Single choice questions and program comprehension questions in the unit test and computer-based final test were used to analyze CT concepts. Single choice questions and program comprehension questions in the unit test reflect the understanding of CT concepts, and the final test based on the computer reflects the application of CT concepts. This study carried out three unit tests which were prepared by the researcher and the teacher, and were distributed at the end of each unit. Well-designed choice measurements could be applied to further learners’ understanding (Glass and Sinha, 2013) and provided them with feedback and explanations (Black and Wiliam, 1998). The computer-based final test was selected by the researchers and the teacher and was distributed at the end of the semester.
Computational Thinking Practices
Computational thinking practices are iterative and incremental, and include testing and debugging, reusing and remixing, and abstracting and modularizing according to Brennan and Resnick (2012). Considering that this study was set in the context of a Python course and did not involve the production and completion of complex products or huge projects, the CT practices of this study only included testing, debugging, reusing, and remixing. Testing and debugging are to ensure that the program can run automatically and efficiently. They are indispensable practices in programming activities. Testing is to find errors, and debugging is to correct errors. Reusing and remixing are based on the problem solutions given by the samples to construct the solutions that include learners’ own ideas. Reusing and remixing are also a process of summarizing problem solutions and migrating to other related problems.
Program correction questions in the unit test and Bebras test were used to analyze CT practices. Among them, the first was the evaluation of the ability of “testing and debugging” in CT practices, and the latter was the evaluation of the ability of “reusing and remixing” in CT practices. The program correction questions were designed to improve the learners’ ability to read and understand the code (Lopez et al., 2008). The Bebras test, an international challenge in informatics and CT, aims to improve and cultivate the CT ability of primary and secondary school students (about 8–18 years old). Bebras reflects the contestants’ CT ability through real-life problems and some focus issues. The Bebras test questions in this study were selected from the Bebras tests in 2016 and 2017.
Computational Thinking Perspectives
Computational thinking perspectives include expressing (computational thinkers see computation as something they can use to design and self-express), connecting (recognizing the importance of creating with others and the value of creating for others), and questioning (feeling empowered to ask questions about and with technology) based on Brennan and Resnick (2012). ISTE and the Computer Science Teachers Association (CSTA) considered CT as a problem-solving process that includes the following dispositions or attitudes: confidence in dealing with complex task, persistence in solving the difficult task, tolerance for ambiguity, the ability to deal with open-ended problems, and the ability to cooperate with others (ISTE and CSTA, 2011). The framework of the Hong Kong CoolThink@JC Jockey Club’s CT education curriculum includes self-expression, questioning and understanding, connecting with life, digital capability, and computational identity.
The CT perspectives in this study consist of four aspects: creation and expression, communication and cooperation, questioning, and problem solving. According to the ISTE definition of CT, Korkmaz et al. (2017) designed a computational thinking scale (CTS) to evaluate CT in algorithmic thinking, creativity, cooperativity, critical thinking, and problem solving. This study refers to CTS and selected items related to CT perspectives to determine the CT perspectives scale, as shown in Table 6. There are 11 items rated using a 5-point Likert scale (from 1 – strongly disagree, to 5 – strongly agree) in the scale.
To measure the validity and reliability of the CT perspectives scale, two classes were selected for trial testing before the formal experiment, and finally 54 valid data were collected. In the study, the Cronbach’s α value was 0.962, showing acceptable reliability in internal consistency. Kaiser–Meyer–Olkin (KMO, =0.854) and Bartlett’s test (p = 0.00 < 0.05) were calculated to test the validity of the scale.
Data Analysis
In the study, SPSS 22.0 was used to analyze the qualitative and quantitative data collected during the experiment. Single choice questions and program comprehension questions in the unit test and computer-based final test were used to evaluate the CT concepts. Program correction questions in the unit test and the Bebras test were used to evaluate the CT practices. The CT perspectives scale was used to evaluate CT perspectives. For the unit test, Bebras test, and CT perspectives scale, the independent sample t-test were applied. Analyzing the computer-based final test was to extract the application times of CT concepts in the code, and then an independent sample t-test was conducted.
Results
This study conducted statistical analysis of the collected qualitative and quantitative data from CT concepts, CT practices, and CT perspectives. According to the normality test, the results of the Unit Test, Computer-based Final Test, Bebras Test, and CT Perspectives Test all conformed to normal distribution (p = 0.200* > 0.05). Thus, the independent sample t-test was used to test the difference between groups and the pre- and post-test. Cohen’s d, which is widely used for the standardization effect in the t-test (Cuthill et al., 2007), is the difference between the mean values of two groups divided by the standard deviation (equation 1). It is used to calculate the effects by comparing the mean values of two groups (Sullivan and Feinn, 2012). When the value of Cohen’s d is ≥ 0.2 and <0.5, it indicates a small effect. If the value is ≥0.5 and <0.8, it shows a moderate effect. When the value is ≥0.8, it means a large effect (Cohen, 1988). Cohen’s d was used as the auxiliary value of the t-test, where high efficiency should indicate a high experimental effect.
Computational Thinking Concepts
Analysis of Single-Choice and Program Comprehension in the Unit Test
The content of the three units were list, string, and dictionary, and the custom function. The single-choice and program comprehension in the unit test examined the students’ understanding of CT concepts. The total scores of the three units were 18, 26, and 9. The independent sample t-test results showed that the mean scores of single-choice questions and program comprehension questions in the unit tests of the EG were higher than those of the CG (see Table 7). The results showed a significant difference in the “list” unit (t = 3.38, p = 0.001 < 0.01, d = 0.88), the “string and dictionary” unit (t = 2.01, p = 0.049 < 0.05, d = 0.53), and the “custom function” unit (t = 2.50, p = 0.015 < 0.05, d = 0.66). The Cohen’s d was 0.88, 0.52, and 0.66, indicating that the POL model had a large effect on students’ “list” unit learning, and a medium effect on their “string and dictionary” unit and “custom function” unit learning. In terms of the dispersion degree of data, the standard deviation of CG (8.69) was much larger than that of EG (5.75), indicating that the CG sample data had greater volatility and their understanding of CT concepts was more unstable, while EG had a more accurate understanding of the concepts of CT.
Table 7. The independent sample t-test of single-choice and program comprehension in the unit tests of the two groups.
Analysis of the Computer-Based Final Test
By analyzing the code in the computer-based final test, the qualitative code data were converted into quantitative data. The qualitative data were the students’ code, and the quantitative data were the number of code blocks that reflect each CT concept in the code. For example, if “if-else” appeared twice in the student’s code, then the CT concept “Conditionals” was increased twice; if there were two “while loops” and one “for loop” in the student’s code, the CT concept “Loops” would increase three times; if students used a custom function and a built-in function in their code, the CT concept “Functions” was increased twice. The number of applications of each CT concept was extracted from the code. The more application of the CT concepts, the more familiar the students were with this concept. The computer-based final test examined the students’ application of CT concepts. The independent sample t-test results of the computer-based final test are shown in Table 8. The mean application of operators, conditionals, data, loops, and functions in the EG was higher than that in the CG. The results showed a significant difference among the application of CT concepts of operators (t = 2.35, p = 0.029 < 0.05, d = 1.02), conditionals (t = 2.31, p = 0.030 < 0.05, d = 0.94), data (t = 2.08, p = 0.048 < 0.05, d = 0.85), and loops (t = 2.17, p = 0.041 < 0.05, d = 0.88), but there was no significant difference in functions (t = 1.13, p = 0.269 > 0.05, d = 0.46). In general, there was a significant difference in the application of CT concepts between the two groups (t = 4.99, p = 0.000 < 0.001, d = 2.00). The Cohen’s d was greater than 0.8. It proved the POL model had a large effect on students’ application of CT concepts.
Computational Thinking Practices
The CT practices in this study included “testing and debugging” and “reusing and remixing” practices. The program correction questions in the unit test examined the students’ “testing and debugging” abilities, and the total scores of the three units were 3, 5, and 2, respectively. The Bebras test examined the students’ ability to “reuse and remix,” with total scores of 36 and 54 for simple questions and difficult questions, respectively.
Analysis of Program Correction in the Unit Test
The independent sample t-tests were performed (see Table 9). The results showed that the mean scores of the program correction questions of students in the EG were higher than those in the CG in the list (t = 1.29, p = 0.203 > 0.05), string and dictionary (t = 0.63, p = 0.530 > 0.05), and custom functions (t = 0.81, p = 0.422 > 0.05), but there was no significant difference.
Table 9. The independent sample t-test of the two groups’ program correction questions in the unit test.
Analysis of the Bebras Test
The Bebras test can reflect students’ ability to transfer CT, and the premise of transfer is the ability to “reuse and remix.” The higher the ability to “reuse and remix,” the more students can transfer CT to the solution of related practical problems. The independent sample t-test result is shown in Table 10. It was found that the Bebras test score of the EG was higher than that of the CG, especially for difficult questions. No significant difference was found between the two groups for either simple questions (t = 1.14, p = 0.260) or difficult questions (t = 1.40, p = 0.166).
Computational Thinking Perspectives
The CT perspectives in this study consist of four aspects: creation and expression, communication and cooperation, questioning, and problem solving. To perform the independent sample t-test of the pre-test of the CT perspectives scale, the results showed no significant difference in creation and expression (t = 1.55, p = 0.126 > 0.05), communication and cooperation (t = 1.56, p = 0.125 > 0.05), questioning (t = 1.03, p = 0.305 > 0.05), or problem solving (t = 1.12, p = 0.268 > 0.05) between the two groups (see Table 11), indicating that the students in the EG were at the same level as the students in the CG before the experiment.
The result of the independent sample t-test on the post-test of the CT perspectives scale is shown in Table 12. It was found that the mean scores of creation and expression (t = 2.03, p = 0.047 < 0.05), communication and cooperation (t = 2.89, p = 0.005 < 0.01), questioning (t = 2.77, p = 0.008 < 0.01), and problem solving (t = 2.35, p = 0.022 < 0.05) of the EG were higher than those of the CG, and there were significant differences. The Cohen’s d was 0.53, 0.73, 0.70, and 0.60, respectively, indicating that the POL model had a medium effect on students’ creation and expression, communication and cooperation, questioning, and problem solving.
The results of the paired sample t-tests on the two groups’ CT perspectives pre- and post-test are shown in Tables 13, 14. Although it was found that the mean scores of creation and expression (t = 1.31, p = 0.194 > 0.05), communication and cooperation (t = 0.76, p = 0.449 > 0.05), questioning (t = 0.56, p = 0.578 > 0.05), problem solving (t = 0.84, p = 0.405 > 0.05), and total scale (t = 1.06, p = 0.292 > 0.05) of the post-test were higher than those of the pre-test, there were no significant differences. However, it was found in the EG that the mean scores of creation and expression (t = 2.00, p = 0.050 = 0.05), communication and cooperation (t = 2.02, p = 0.049 < 0.05), problem solving (t = 2.28, p = 0.027 < 0.05), and total scale (t = 2.52, p = 0.015 < 0.05) of the post-test were higher than those of the pre-test, and there were significant differences. The Cohen’s d was 0.541, 0.532, 0.588, and 0.650, respectively, indicating that although the Python course cannot significantly improve students’ computational thinking, the Python course taught using the POL model can significantly improve students’ computational thinking.
Discussion
This study combined the process of solving programming problems to construct a POL model, and verified its effectiveness for CT training through the implementation of Python programming teaching.
Computational Thinking Concepts
The evaluation of the CT concepts in this study included two aspects: understanding of the CT concepts and application of the CT concepts. Application of the CT concepts refers to the number of CT concepts included in the students’ programming works. According to the results, the understanding and application of the CT concepts by the students learning with the POL model were better than those of the students learning with the LAP model, and the difference was significant. This implied that the proposed POL model benefited the students’ CT concepts. The applied POL model was a process in which students actively constructed their understanding of programming concepts, as it focused on their understanding and application of CT concepts. The LAP learning model was a process by which teachers instilled programming concepts into students. Students passively accepted the programming concepts, so they did not understand the concepts well and were not proficient in the application. The application of the function to the two groups of students was not significant. Previous studies also found that some concepts are difficult for beginner programmers (Meerbaum-Salant et al., 2013). Grover et al. (2015) developed a course of “Foundations for Advancing Computational Thinking” to promote learners’ understanding of algorithmic concepts, but the mechanics of some constructs were difficult for learners to grasp in the context of text-based languages.
Although students can write and explain simple programs, they have difficulty with programs involving basic programming concepts (Brennan and Resnick, 2012). Students often struggle with algorithmic concepts, especially if teachers do not use appropriate supportive pedagogy to teach these concepts (Grover et al., 2015). In Python programming, the concept of functions includes built-in functions and custom functions. Built-in functions generally only require students to memorize and apply them, while custom functions require students to be able to build function modules by themselves. A function usually contains multiple CT concepts, and students need to be familiar with programming logic. For novice programmers, the application of functions is relatively limited, and the items involved in class are relatively simple. Students can directly write the corresponding code in the program without writing another function.
Computational Thinking Practices
Computational thinking practice in this study included two aspects: “testing and debugging” and “reusing and remixing.” According to the results, the students in the POL mode had better CT practice than the students in the LAP learning mode in terms of “testing and debugging” and “reusing and mixing,” but there was no significant difference. In the POL model classroom, the teacher’s identity was more like a guide. Students designed possible problem solutions and tried them one by one. In this process, they continued to practice testing and debugging. Frequent operations helped develop their ability. In the LAP model classroom, the teacher explained the problem-solving solutions, and the students practiced on this basis, thereby avoiding many errors. Therefore, the students’ testing and debugging practices were lacking in this model.
Testing and debugging are indispensable for any type of problem solving (McCauley et al., 2008). Strict and systematic testing and debugging is an art and science in the field of computing, especially in the field of software development (Grover and Pea, 2013). However, the difference between the “testing and debugging” abilities of students in the two groups was not significant, which is consistent with previous studies. Fessakis et al. (2013) proposed that some students did not show any clear planning but rather tried commands one by one. For novice programmers, it is often difficult to link upper and lower command lines in groups (Robins et al., 2003), and they only analyze the single command line that includes a mistake (Lehrer et al., 1999). Hence, it is difficult to accurately and quickly find errors, correct them, and develop testing and debugging capabilities. Reusing and remixing involve comprehensive migration of problem solutions, and mastering skills in the original environment is essential for migration (Kurland and Pea, 1985). However, under the two learning models, the difference in the ability of students to “reuse and remix” was not significant. Previous research has shown that skill development usually requires sufficient training time (Bers et al., 2014; Atmatzidou and Demetriadis, 2016). On the other hand, most of the computer-based questions in this study were structured programming problems, and there was a lack of unstructured practical problems. Although this is conducive to the solution of the problem, it is not conducive to the development of migration capacity.
Computational Thinking Perspectives
In this study, CT perspectives consist of “creation and expression,” “communication and cooperation,” “questioning,” and “problem solving.” The results showed that the CT perspectives of the students who adopted the POL model were better than those of the students who adopted the LAP model, and the difference was significant, indicating that the POL model was more effective in terms of cultivating students’ CT perspectives than the LAP learning model. Under the POL model, students expressed their own understanding of programming problems through programming to achieve self-creation and expression; in the process of finding solutions, students actively thought and discussed, and improved their communication and collaboration skills. When the solution was wrong or there was a conflict between their own thinking and the ideas of their classmates, the students would have doubts, so as to realize the optimal design of the algorithm. In the LAP model, teachers directly explained the process of problem-solving, while students were involved in the process of absorption, and there was less questioning and less interaction between students.
The result is consistent with previous research findings. Mouza et al. (2020) designed a 9-week after-school computational programming course, and collected long-term records of changes in students’ CT concepts, practices, and perspectives. Through interviews, it was found that the students’ CT perspectives greatly improved. Students were more willing to share programming works with classmates, which was a way to help them build confidence in programming. Moreover, repeated participation in computing courses made students use a number of computing perspectives. Burke (2012) found that middle school students could create their own digital stories through programming tools to express their CT perspectives. Kong and Wang (2020) also found that programming could improve CT perspectives. CT perspectives are connected to the formation of students’ thinking habits and personality, which have a significant influence on shaping teenagers’ cognition and values of the digital society (Deng et al., 2020). The CT perspectives require students to develop an understanding of themselves and relationships with others people and the technological world. When students express themselves in programming, CT perspectives are evident (Lye and Koh, 2014).
Conclusion
This study constructed a POL model oriented to programming problems, and used a quasi-experiment to verify its effect on the cultivation of CT. In all, 60 eighth-grade students from a middle school in China were divided into an EG which adopted the POL model, and a CG which adopted the LAP learning model. The results showed that the students who were instructed with the POL model performed better than those who were instructed using the LAP model in terms of CT concepts, CT practices, and CT perspectives. Significant differences were found in CT concepts and CT perspectives, but no significant difference in CT practices.
Implications
In terms of theory, although there have been studies on the relevance of CT and problem solving, there have been few empirical studies based on this theory. This study explored CT from the perspective of problem solving, and conducted empirical research, which not only enriches the related research on CT, but also provides theoretical references to explore CT in depth with a focus on problem solving. In terms of practice, this study provides a new practical perspective on how to cultivate the CT of middle school students, that is, relying on Python programming courses, applying the POL model, and imperceptibly cultivating students’ computational thinking. The study provides reference cases for computational thinking training and Python programming teaching, and provides an experience reference for teachers to carry out programming teaching.
Limitations and Future Works
It should be noted that this study has some limitations. Firstly, the intervention time was short, just one semester. CT involves the use of computational science concepts and cognitive processes to solve problems creatively and efficiently (Anderson, 2016), and consists of multiple elements. Therefore, the cultivation of CT is not accomplished overnight. This study applied the POL model to a programming course. It is difficult to comprehensively and significantly improve students’ CT through only one semester of study. Further research can be conducted to investigate the effects of long-term use of this model.
Secondly, the fatigue response of the participants may have been a factor in the study. During the Python programming course, experimental participants needed to solve and complete a large number of programming problems and phased tests which could have led to their fatigue response in the later stage of the experiment. In addition, the assessment instrument modality may have had an impact on students’ performance (Atmatzidou and Demetriadis, 2016). CT can be applied in a wider learning environment instead of computational solutions (Kalelioğlu et al., 2016). Therefore, in future research, more CT evaluations that do not require a computer or programming platform should be developed (Tang et al., 2020).
Data Availability Statement
The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.
Author Contributions
All authors contributed equally to the conception of the idea, implementing and analyzing the experimental results, writing the manuscript, and reading and approving the final manuscript.
Funding
This research was funded by the National Social Science Foundation of China (Grant No. BCA190079).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Akben, N. (2020). Effects of the Problem-Posing Approach on Students’ Problem Solving Skills and Metacognitive Awareness in Science Education. Res. Sci. Educ. 50, 1143–1165. doi: 10.1007/s11165-018-9726-7
Ananiadou, K., and Claro, M. (2009). 21st Century skills and competences for new millennium learners in OECD Countries. OECD Education Working Papers. United States: OECD, doi: 10.1787/218525261154
Anderson, N. D. (2016). A call for computational thinking in undergraduate psychology. Psychol. Learn. Teach. 15, 226–234. doi: 10.1177/1475725716659252
Atmatzidou, S., and Demetriadis, S. (2016). Advancing students’ computational thinking skills through educational robotics: a study on age and gender relevant differences. Robot. Auton. Syst. 75, 661–670. doi: 10.1016/j.robot.2015.10.008
Baruch, A. F., and Erstad, O. (2018). Upbringing in a digital world: opportunities and possibilities. Technol. Knowl. Learn. 23, 377–390. doi: 10.1007/s10758-018-9386-8
Basawapatna, A., Koh, K. H., Repenning, A., Webb, D. C., and Marshall, K. S. (2011). “Recognizing computational thinking patterns,” in Proceedings of the 42nd ACM Technical Symposium on Computer Science Education, (New York, NY: ACM), 245–250. doi: 10.1145/1953163.1953241
Basu, S., Biswas, G., and Kinnebrew, J. S. (2017). Learner modeling for adaptive scaffolding in a Computational Thinking-based science learning environment. User Model. User-Adapt. Interact. 27, 5–53. doi: 10.1007/s11257-017-9187-0
Bati, K. (2021). A systematic literature review regarding computational thinking and programming in early childhood education. Educ. Inf. Technol. 26, 1–24. doi: 10.1007/s10639-021-10700-2
Beichner, R. J. (2002). GOAL-Oriented Problem Solving. Available Online at: https://projects.ncsu.edu/PER/archive/GOALPaper.pdf (accessed August 20, 2021).
Bers, M. U., Flannery, L., Kazakoff, E. R., and Sullivan, A. (2014). Computational thinking and tinkering: exploration of an early childhood robotics curriculum. Comput. Educ. 72, 145–157. doi: 10.1016/j.compedu.2013.10.020
Black, P., and Wiliam, D. (1998). Assessment and classroom learning. Assess. Educ. 5, 7–74. doi: 10.1080/0969595980050102
Bower, M., Wood, L. N., Lai, J. M., Howe, C., Lister, R., Mason, R., et al. (2017). Improving the computational thinking pedagogical capabilities of school teachers. Aust. J. Teach. Educ. 42:4. doi: 10.14221/ajte.2017v42n3.4
Brennan, K., and Resnick, M. (2012). “New frameworks for studying and assessing the development of computational thinking,” in 2012 Annual American Educational Research Association Meeting, (Cambridge, Massachusetts: MIT Media Lab).
Bundy, A. (2007). Computational Thinking is Pervasive. Available Online at: http://www.inf.ed.ac.uk/publications/online/1245.pdf (accessed August 20, 2021).
Burke, Q. (2012). The markings of a new pencil: introducing programming-as-writing in the middle school classroom. J. Media Lit. Educ. 4, 121–135.
Cheng, S., She, H., and Huang, L. (2018). The Impact of Problem-Solving Instruction on Middle School Students’ Physical Science Learning: interplays of Knowledge, Reasoning, and Problem Solving. Eurasia J. Math. Sci. Technol. Educ. 14, 731–743. doi: 10.12973/ejmste/80902
Chevalier, M., Giang, C., Piatti, A., and Mondada, F. (2020). Fostering computational thinking through educational robotics: a model for creative computational problem solving. Int. J. STEM. Educ. 7, 938–954. doi: 10.1186/s40594-020-00238-z
Cohen, J. (1988). Statistical Power Analysis for the Behavior Sciences (Revised Edition). New York: Routledge. doi: 10.2307/2529115
Corral, J. M. R., Balcells, A. C., Estevez, A. M., Moreno, G. J., and Ramos, M. J. F. (2014). A game-based approach to the teaching of object-oriented programming languages. Comput. Educ. 73, 83–92. doi: 10.1016/j.compedu.2013.12.013
Cuthill, I., Nakagawa, S., and Cuthill, I. C. (2007). Effect size, confidence intervals and statistical significance: a practical guide for biologists. Biol. Rev. Camb. Philos. Soc. 82, 591–605. doi: 10.1111/j.1469-185X.2007.00027.x
Danial, H., Liina, M., Yeongwook, Y., Margus, P., Minhong, W., and Heuiseok, L. (2021). An adaptive educational computer game: effects on students’ knowledge and learning attitude in computational thinking. Comput. Hum. Behav. 114:106575. doi: 10.1016/j.chb.2020.106575
Deng, W., Pi, Z., Lei, W., Zhou, Q., and Zhang, W. (2020). Pencil Code improves learners’ computational thinking and computer learning attitude. Comput. Appl. Eng. Educ. 28, 90–104. doi: 10.1002/cae.22177
Dewi, G. A. C., Sunarno, W., and Supriyanto, A. (2019). “The needs analysis on module development based on creative problem solving method to improve students’ problem solving ability,” in in 9th International Conference on Physics and Its Applications (ICOPIA), (Bristol: IOP Publishing)
Durak, H. Y., and Saritepeci, M. (2018). Analysis of the relation between computational thinking skills and various variables with the structural equation model. Comput. Educ. 116, 191–202.
Fessakis, G., Gouli, E., and Mavroudi, E. (2013). Problem solving by 5-6 years old kindergarten children in a computer programming environment: a case study. Comput. Educ. 63, 87–97. doi: 10.1016/j.compedu.2012.11.016
Florez, F. B., Casallas, R., Hernandez, M., Reyes, A., Restrepo, S., and Danies, G. (2017). Changing a generation’s way of thinking: teaching computational thinking through programming. Rev. Educ. Res. 87, 834–860. doi: 10.3102/0034654317710096
Furber, S. (2012). Shut Down or Restart: The Way Forward for Computing in UK Schools. Available Online at: https://royalsociety.org/topics-policy/projects/computing-in-schools/report/ (accessed August 20, 2021).
García Monsálvez, J. C. (2017). Python como primer lenguaje de programación textual en la Enseñanza Secundaria. Educ. Knowl. Soc. 18, 147–162. doi: 10.14201/eks2017182147162
Garneli, V., Giannakos, M. N., and Chorianopoulos, K. (2015). “Computing education in K-12 schools: a review of the literature,” in in IEEE global engineering education conference, (Tallinn, Estonia: IEEE), 543–551.
Glass, A. L., and Sinha, N. (2013). Providing the answers does not improve performance on a college final exam. Educ. Psychol. 33, 87–118. doi: 10.1080/01443410.2012.723864
Grover, S., and Pea, R. (2013). Computational thinking in K-12: a review of the state of the field. Educ. Res. 42, 38–43. doi: 10.3102/0013189X12463051
Grover, S., Pea, R., and Cooper, S. (2015). Designing for deeper learning in a blended computer science course for middle school students. Comput. Sci. Educ. 25, 199–237. doi: 10.1080/08993408.2015.1033142
He, S., Maldonado, J., Uquillas, A., and Cetoute, T. (2014). “Teaching K-12 students robotics programming in collaboration with the robotics club,” in 2014 IEEE Integrated STEM Education Conference (ISEC), (Princeton, NJ: IEEE), doi: 10.1109/ISECon.2014.6891023
Hooshyar, D., Pedaste, M., Yang, Y., Malva, L., Hwang, G.-J., Wang, M., et al. (2021). From Gaming to Computational Thinking: an Adaptive Educational Computer Game-Based Learning Approach. J. Educ. Comput. Res. 59, 383–409. doi: 10.1177/0735633120965919
Hsu, T.-C., Chang, S.-C., and Hung, Y.-T. (2018). How to learn and how to teach computational thinking: suggestions based on a review of the literature. Comput. Educ. 126, 296–310. doi: 10.1016/j.compedu.2018.07.004
Huang, W., and Looi, C.-K. (2021). A critical review of literature on “unplugged” pedagogies in K-12 computer science and computational thinking education. Comput. Sci. Educ. 31, 83–111. doi: 10.1080/08993408.2020.1789411
Israel-Fishelson, R., and Hershkovitz, A. (2019). Persistence in a Game-Based Learning Environment: the Case of Elementary School Students Learning Computational Thinking. J. Educ. Comput. Res. 58, 891–918. doi: 10.1177/0735633119887187
ISTE (2015). CT leadership toolkit. Available online at: https://cdn.iste.org/www-root/2020-10/ISTE_CT_Leadership_Toolkit_booklet.pdf (accessed August 5, 2021).
ISTE, and CSTA (2011). Operational Definition of Computational Thinking for K-12 Education. Available online at: https://cdn.iste.org/www-root/ct-documents/computational-thinking-operational-definition-flyer.pdf (accessed August 5, 2021).
Jonassen, D., Howland, J., Marra, R. M., and Crismond, D. (2012). Meaningful learning with technology, 4th Edn. London: Pearson.
Jun, S., Han, S., Kim, H., and Lee, W. (2014). Assessing the computational literacy of elementary students on a national level in Korea. Educ. Assess. Eval. Acc. 26, 319–332. doi: 10.1007/s11092-013-9185-7
Kalelioğlu, F., Gülbahar, Y., and Kukul, V. (2016). A Framework for Computational Thinking Based on a Systematic Research Review. Balt. J. Mod. Comput. 4, 583–596.
Kelleher, C., and Pausch, R. (2005). Lowering the barriers to programming: a taxonomy of programming environments and languages for novice programmers. ACM Comput. Surv. 37, 83–137. doi: 10.1145/1089733.1089734
Kim, B., Kim, T., and Kim, J. (2013). Paper-and-pencil programming strategy toward computational thinking for non-majors: design your solution. J. Educ. Comput. Res. 49, 437–459. doi: 10.2190/EC.49.4.b
Kim, H.-S., and Yun, B.-H. (2020). A Study on Teaching-Learning Model Development for Problem Solving based on Computational Thinking. J. Digit. Contents Soc. 21, 865–873. doi: 10.9728/dcs.2020.21.5.865
Kim, J., Kim, M., Hyejin, Y., Kim, Y., and Kim, J. (2019). Effect of data visualization education with using Python on computational thinking of six grade in elementary school. J. Korean Assoc. Inf. Educ. 23, 197–206. doi: 10.14352/JKAIE.2019.23.3.197
Kong, S.-C. (2019). “Components and Methods of Evaluating Computational Thinking for Fostering Creative Problem-Solvers in Senior Primary School Education,” in Computational Thinking Education eds S. C. Kong and H. Abelson (Singapore: Springer). 119–141. doi: 10.1007/978-981-13-6528-7_8
Kong, S.-C., Lai, M., and Sun, D. (2020). Teacher development in computational thinking: design and learning outcomes of programming concepts, practices and pedagogy. Comput. Educ. 151:103872. doi: 10.1016/j.compedu.2020.103872
Kong, S.-C., and Wang, Y. Q. (2020). Formation of computational identity through computational thinking perspectives development in programming learning: a mediation analysis among primary school students. Comput. Hum. Behav. 106:106230. doi: 10.1016/j.chb.2019.106230
Korkmaz, Ö, Cakir, R., and Özden, M. Y. (2017). A Validity and Reliability Study of the Computational Thinking Scales (CTS). Comput. Hum. Behav. 72, 558–569. doi: 10.1016/j.chb.2017.01.005
Kuo, W.-C., and Hsu, T.-C. (2020). Learning Computational Thinking Without a Computer: how Computational Participation Happens in a Computational Thinking Board Game. Asia Pac. Educ. Res. 29, 67–83. doi: 10.1007/s40299-019-00479-9
Kurland, D. M., and Pea, R. D. (1985). Children’s mental models of recursive logo programs. J. Educ. Comput. Res. 1, 235–243.
Lai, H., and Xin, W. (2011). “An experimental research of the pair programming in Java programming course,” in 2011 International Conference on e-Education, Entertainment and e-Management, (Bali, Indonesia: IEEE), 257–260. doi: 10.1109/ICeEEM.2011.6137800
Lee, Y., and Cho, J. (2017). The Influence of Python Programming Education for Raising Computational Thinking. Int. J. u-e-Serv. Sci. Technol. 10, 59–72. doi: 10.14257/ijunesst.2017.10.8.06
Lehrer, R., Lee, M., and Jeong, A. (1999). Reflective teaching of Logo. J. Learn. Sci. 8, 245–289. doi: 10.1207/s15327809jls0802_3
Li, Q., and Ma, X. (2010). A meta-analysis of the effects of computer technology on school students’ mathematics learning. Educ. Psychol. Rev. 22, 215–243. doi: 10.1007/s10648-010-9125-8
Lopez, M., Whalley, J., Robbins, P., and Lister, R. (2008). “Relationships between reading, tracing and writing skills in introductory programming,” in Proceedings of the fourth international workshop on computing education research, (New York: ACM), 101–112.
Lye, S. Y., and Koh, J. H. L. (2014). Review on teaching and learning of computational thinking through programming: what is next for K-12? Comput. Hum. Behav. 41, 51–61. doi: 10.1016/j.chb.2014.09.012
Maria, S., and Tsiatsos, T. (2018). “Case Study: integrating Computational Thinking into the Introductory Course of Computer Science via the Use of the Programming Language Python,” in Interactive Mobile Communication Technologies and Learning, eds M. E. Auer and T. Tsiatsos (Berlin: Springer), 531–541. doi: 10.1007/978-3-319-75175-7_52
McCauley, R., Fitzgerald, S., Lewandowski, G., Murphy, L., Simon, B., Thomas, L., et al. (2008). Debugging: a review of the literature from an educational perspective. Comput. Sci. Educ. 18, 67–92. doi: 10.1080/08993400802114581
Meerbaum-Salant, O., Armoni, M., and Ben-Ari, M. (2013). Learning computer science concepts with Scratch. Comput. Sci. Educ. 23, 239–264. doi: 10.1080/08993408.2013.832022
Mishra, P., and Yadav, A. (2013). Rethinking technology and creativity in the 21st century. TechTrends 57, 10–14. doi: 10.1007/s11528-013-0655-z
Mouza, C., Pan, Y.-C., Yang, H., and Pollock, L. (2020). A Multiyear Investigation of Student Computational Thinking Concepts, Practices, and Perspectives in an After-School Computing Program. J. Educ. Comput. 58, 1029–1056. doi: 10.1177/0735633120905605
Nouri, J., Zhang, L., Mannila, L., and Noren, E. (2020). Development of computational thinking, digital competence and 21st century skills when learning programming in K-9. Educ. Inq. 11, 1–17. doi: 10.1080/20004508.2019.1627844
Okonkow, C. W., and Ade-Ibijola, A. (2021). Python-Bot: a Chatbot for Teaching Python Programming. Eng. Lett. 29, 25–34.
Olelewe, C. J., and Agomuo, E. E. (2016). Effects of B-learning and F2F learning environments on students’ achievement in QBASIC programming. Comput. Educ. 103, 76–86. doi: 10.1016/j.compedu.2016.09.012
Parsazadeh, N., Cheng, P.-U., Wu, T.-T., and Huang, Y.-M. (2020). Integrating Computational Thinking Concept Into Digital Storytelling to Improve Learners’ Motivation and Performance. J. Educ. Comput. Res. 59, 470–495. doi: 10.1177/0735633120967315
Qu, J. R., and Fok, P. K. (2021). Cultivating students’ computational thinking through student–robot interactions in robotics education. Int. J. Technol. Des. Educ. 31, 1–20. doi: 10.1007/s10798-021-09677-3
Repenning, A., Webb, D. C., Koh, K. H., Nickerson, H., Miller, S. B., Brand, C., et al. (2015). Scalable game design: a strategy to bring systemic computer science education to schools through game design and simulation creation. ACM Trans. Comput. Educ. 15, 1–31. doi: 10.1145/2700517
Rich, P. J., Jones, B., Belikov, O., Yoshikawa, E., and Perkins, M. (2017). Computing and engineering in elementary school: the effect of year-long training on elementary teacher self-efficacy and beliefs about teaching computing and engineering. Int. J. Comput. Sci. Eng. Syst. 1, 1–20. doi: 10.21585/ijcses.v1i1.6
Robins, A., Rountree, J., and Rountree, N. (2003). Learning and teaching programming: a review and discussion. Comput. Sci. Educ. 13, 137–172. doi: 10.1076/csed.13.2.137.14200
Román-González, M., Pérez-González, J.-C., and Jiménez-Fernández, C. (2017). Which cognitive abilities underlie computational thinking? Criterion validity of the computational thinking test. Comput. Hum. Behav. 72, 678–691. doi: 10.1016/j.chb.2016.08.047
Sáez-López, J.-M., Román-González, M., and Vázquez-Cano, E. (2016). Visual programming languages integrated across the curriculum in elementary school. Comput. Educ. 97, 129–141. doi: 10.1016/j.compedu.2016.03.003
Sands, P., Yadav, A., and Good, J. (2018). “Computational thinking in K-12: in-service teacher perceptions of computational thinking: foundations and research highlights,” in Computational thinking in the STEM disciplines, ed. M. S. Khine (Berlin: Springer), 151–164. doi: 10.1007/978-3-319-93566-9_8
Saritepeci, M. (2020). Developing Computational Thinking Skills of High School Students: design-Based Learning Activities and Programming Tasks. Asia Pac. Edu. Res. 29, 35–54. doi: 10.1007/s40299-019-00480-2
Scherer, R., Siddiq, F., and Viveros, B. S. (2020). A meta-analysis of teaching and learning computer programming: effective instructional approaches and conditions. Comput. Hum. Behav. 109:106349. doi: 10.1016/j.chb.2020.106349
Sengupta, P., Kinnebrew, J. S., Basu, S., Biswas, G., and Clark, D. (2013). Integrating computational thinking with K-12 science education using agent-based computation: a theoretical framework. Educ. Inf. Technol. 18, 351–380. doi: 10.1007/s10639-012-9240-x
Shute, V. J., Sun, C., and Asbell-Clarke, J. (2017). Demystifying computational thinking. Educ. Res. Rev. 22, 142–158. doi: 10.1016/j.edurev.2017.09.003
Soleimani, A., Herro, D., and Green, K. E. (2019). CyberPLAYce—A tangible, interactive learning tool fostering children’s computational thinking through storytelling. Int. J. Child Comput. Interact. 20, 9–23. doi: 10.1016/j.ijcci.2019.01.002
Suarsana, I. M., Lestari, I. A. P. D., and Mertasari, N. M. S. (2019). The effect of online problem posing on students’ problem-solving abilitiy in Mathematics. Int. J. Instr. 12, 809–820. doi: 10.29333/iji.2019.12152a
Sullivan, G. M., and Feinn, R. (2012). Using effect size-or why the p value is not enough. J. Grad. Med. Educ. 4, 279–282. doi: 10.4300/JGME-D-12-00156.1
Sun, L., Hu, L., and Zhou, D. (2021). Improving 7th-graders’ computational thinking skills through unplugged programming activities: a study on the influence of multiple factors. Think. Ski. Creat. 42:100926. doi: 10.1016/j.tsc.2021.100926
Tang, X., Yin, Y., Lin, Q., Hadad, R., and Zhai, X. (2020). Assessing computational thinking: a systematic review of empirical studies. Comput. Edu. 148:103798. doi: 10.1016/j.compedu.2019.103798
Uysal, M. P. (2014). Improving first computer programming experiences: the case of adapting a web-supported and well-structured problem-solving method to a traditional course. Contemp. Educ. Technol. 5, 198–217. doi: 10.30935/cedtech/6125
Vihavainen, A., Airaksinen, J., and Watson, J. (2014). “A systematic review of approaches for teaching introductory programming and their influence on success,” in Tenth Annual conference on international computing education research (ICER), (New York: ACM), doi: 10.1145/2632320.2632349
Weintrop, D., Beheshti, E., Horn, M., Orton, K., Jona, K., Trouille, L., et al. (2016). Defining computational thinking for mathematics and science classrooms. J. Sci. Educ. Technol. 25, 127–147. doi: 10.1007/s10956-015-9581-5
Weintrop, D., and Wilensky, U. (2015). “To block or not to block, that is the question: students’ perceptions of blocks-based programming,” in IDC 15 Proc. 14th Int. Conf. Interact. Des. Child. (Boston, Massachusetts-June 21 - 24), (New York, NY: ACM), 199–208.
Wen, X., Yang, Z., Wang, Z., and Zhao, M. (2014). “Exploration on Cultivating Students’ Abilities Based on Python Teaching Practice,” in 2nd International Conference on Teaching and Computational Science (ICTCS 2014), (Paris, France: Atlantis Press), 192–195. doi: 10.2991/ictcs-14.2014.51
Werner, L., Denner, J., Campe, S., Ortiz, E., DeLay, D., Hartl, A. C., et al. (2013). “Pair programming for middle school students: does friendship influence academic outcomes?,” in Proceeding of the 44th ACM technical symposium on Computer science education, (New York: ACM), 421–426.
Wilkerson-Jerde, M. H. (2014). Construction, categorization, and consensus: student generated computational artifacts as a context for disciplinary reflection. Educ. Technol. Res. Dev. 62, 99–121. doi: 10.1007/s11423-013-9327-0
Wing, J. M. (2008). Computational thinking and thinking about computing. Philos. Trans. R. Soc. A 366, 3717–3725. doi: 10.1098/rsta.2008.0118
Wing, J. M. (2011). Computational thinking: what and why. Available Online at: https://www.cs.cmu.edu/link/research-notebook-computational-thinking-what-and-why(accessed August 5, 2021).
Wolz, U., Stone, M., Pearson, K., Pulimood, S. M., and Switzer, M. (2011). Computational thinking and expository writing in the middle school. ACM Trans. Comput. Educ. 11:9. doi: 10.1145/1993069.1993073
Yang, Y. F. (2010). Students’ reflection on online self-correction and peer review to improve writing. Comput. Educ. 55, 1202–1210. doi: 10.1016/j.compedu.2010.05.017
Zhong, B., Wang, Q., Chen, J., and Li, Y. (2017). Investigating the Period of Switching Roles in Pair Programming in a Primary School. Educ. Technol. Soc. 20, 220–233.
Keywords: computational thinking, problem-oriented learning model, programming teaching, Python, middle school students
Citation: Bai H, Wang X and Zhao L (2021) Effects of the Problem-Oriented Learning Model on Middle School Students’ Computational Thinking Skills in a Python Course. Front. Psychol. 12:771221. doi: 10.3389/fpsyg.2021.771221
Received: 06 September 2021; Accepted: 16 November 2021;
Published: 07 December 2021.
Edited by:
Cristina M. Pulido, Universitat Autònoma de Barcelona, SpainReviewed by:
Patcharin Panjaburee, Mahidol University, ThailandWatcharee Ketpichainarong, Mahidol University, Thailand
Copyright © 2021 Bai, Wang and Zhao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Li Zhao, bGkuemhhb0Buam51LmVkdS5jbg==