Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 10 January 2022
Sec. Educational Psychology
This article is part of the Research Topic Deep Learning in Adaptive Learning: Educational Behavior and Strategy View all 55 articles

The Influencing Legal and Factors of Migrant Children’s Educational Integration Based on Convolutional Neural Network

\r\nChi ZhangChi Zhang1Gang Wang*Gang Wang1*Jinfeng ZhouJinfeng Zhou2Zhen ChenZhen Chen3
  • 1School of Marxism, Northeast Forestry University, Harbin, China
  • 2China Biodiversity Conservation and Green Development Foundation, Beijing, China
  • 3School of Marxism, Northeastern University, Shenyang, China

This research aims to analyze the influencing factors of migrant children’s education integration based on the convolutional neural network (CNN) algorithm. The attention mechanism, LSTM, and GRU are introduced based on the CNN algorithm, to establish an ALGCNN model for text classification. Film and television review data set (MR), Stanford sentiment data set (SST), and news opinion data set (MPQA) are used to analyze the classification accuracy, loss value, Hamming loss (HL), precision (Pre), recall (Re), and micro-F1 (F1) of the ALGCNN model. Then, on the big data platform, data in the Comprehensive Management System of Floating Population and Rental Housing, Student Status Information Management System, and Student Information Management System of Beijing city are taken as samples. The ALGCNN model is used to classify and compare related data. It is found that in the MR, STT, and MPQA data sets, the classification accuracy and loss value of the ALGCNN model are better than other algorithms. HL is the lowest (15.2 ± 1.38%), the Pre is second only to the BERT algorithm, and the Re and F1 are both higher than other algorithms. From 2015 to 2019, the number of migrant children in different grades of elementary school shows a gradual increase. Among migrant children, the number of migrant children from other counties in this province is evidently higher than the number of migrant children from other provinces. Among children of migrant workers, the number of immigrants from other counties in this province is also notably higher than the number of immigrants from other provinces. With the gradual increase in the years, the proportion of township-level expenses shows a gradual decrease, whereas the proportion of district and county-level expenses shows a gradual increase. Moreover, the accuracy of the ALGCNN model in migrant children and local children data classification is 98.6 and 98.9%, respectively. The proportion of migrant children in the first and second grades of a primary school in Beijing city is obviously higher than that of local children (p < 0.05). The average final score of local children was greatly higher than that of migrant children (p < 0.05), whereas the scores of migrant children’s listening methods, learning skills, and learning environment adaptability are lower, which shows that an effective text classification model (ALGCNN) is established based on the CNN algorithm. In short, the children’s education costs, listening methods, learning skills, and learning environment adaptability are the main factors affecting migrant children’s educational integration, and this work provides a reference for the analysis of migrant children’s educational integration.

Introduction

Since the reform and opening-up, the scale floating population in China has shown a trend of increasing year by year (Wang et al., 2019). After 1990, “family” migration had developed in China’s migrant population, so that many school-age children followed their parents into cities, and the number increased dramatically (Ajay et al., 2018). Childhood is an important period of education and physical and mental development, and children’s education relates to the harmony and stability of the entire society. To guarantee the education of migrant children, most national and local governments have successively promulgated relevant policies to give more support to migrant children during the compulsory education stage (Hu et al., 2018; Liang et al., 2019). The children migrating to the city have undergone a transformation of their living space to a certain extent. It should focus on whether the school’s learning environment can provide these migrant children with a comfortable and suitable education and living environment (Vogel, 2018), thereby eliminating or reducing the negative impacts on migrant children when they attend a school in the city with disadvantages in their families, schools, and society, and an unequal educational starting point. The level of education of migrant children and their integration with the cities where they flowed into not only have an important impact on China’s social stability and social development, but are also closely related to industrialization, urbanization, and the progression of the future labor market (Flensner et al., 2021). The direction of national education policy formulation changes with the change of social development. As a provider of public education services, the government has introduced various educational policies in the process of overall planning for the education of school-age migrant children, which will inevitably have a certain impact on the social integration of migrant children (Vargas-Valle and Aguilar-Zepeda, 2020). Scholars have been paying attention to the issue of migrant children’s social integration for a long time, but most of them start from the family environment or community level participation of migrant children (Kyereko and Faas, 2021). When it comes to the educational integration of migrant children, most studies are conducted from the perspective of school adaptation and the implementation of educational policies (Wang, 2020; Jørgensen et al., 2021), where those who adopt educational policy as an influencing factor to analyze the educational integration of migrant children are relatively rare.

Using big data to analyze the current situation of migrant children’s education is of great significance for us to correctly understand the current education status and future trends. When faced with big data, choosing an algorithm that can quickly and automatically extract text information is of great significance for improving the mining of effective information in the data. The vectorization result of the text directly determines the performance of the final classification of the text to a large extent (Lima et al., 2020). Text classification methods based on machine learning, such as support vector machine (SVM), decision tree (DT), and naive Bayes (NB), are proved to be suitable for effective classification of text information (Liang and Yi, 2021). Logistic regression and Softmax in the deep learning algorithm can also be applied to the feature extraction and classification of text information, and the classification effect is greatly superior to that of the machine learning algorithm (Griffiths and Boehm, 2019). However, these models face more complex problems due to factors such as human interference and training abnormalities, which leads to their classification results to be further optimized. Convolutional neural network (CNN) obtains local features through the movement of the window and shows the highest usage rate in the form of maximum pooling and average pooling (Barberá et al., 2021). The long–short-term memory (LSTM) model overcomes the gradient disappearance through gate control. Some researchers proposed the gated recurrent unit (GRU) model, which can adaptively capture the dependence of different time scales through the recurrent block, and the attention mechanism is introduced to ensure the different effects of input data on output data (Dashtipour et al., 2020). Basiri et al. (2021) established an ABCNN model based on the CNN and attention mechanism, which improved the effective classification of text. However, different text classification models still have the influence of noise such as human factors during feature selection, which needs to be further optimized.

To sum up, studies adopting educational policy as an influencing factor to analyze the educational integration of migrant children are relatively rare. Therefore, attention mechanism, LSTM, and GRU are introduced based on the CNN algorithm, and an ALGCNN model for text classification is innovatively established. ALGCNN model classification performance is evaluated based on known public data sets, to analyze the big data of migrant children in China and the current situation of migrant children’s education in a primary school in Beijing city. The data in the Comprehensive Management System of Floating Population and Rental Housing, Student Status Information Management System, and Student Information Management System of Beijing city are taken as samples. The ALGCNN model is used to classify and analyze the samples to explore the relevant factors affecting education integration, to provide help for migrant children to establish better education integration, and to provide a basis for improving the education environment and policies of migrant children. This work also aims to provide research reference of relevant issues for the future legislative work, to ensure that the migrant students can get legal and effective guarantee mechanism as protection. It also provides more in-depth and more direct legal provisions for different students to receive fair education after indicating the direction and content.

Methodology

Definitions of Related Concepts

Migrant children: “children” refers to people in the childhood stage, and “migrant children” is a relatively special part of the “children” (Stadelmannsteffen, 2018). With the deepening of China’s urbanization process, more and more rural people choose to work in cities. As they migrate between cities and rural areas, a special group of “migrant children” is evolved (Schrier et al., 2019). Based on the Education Policies for Migrant Children (Provisional) promulgated by the State Education Commission of the People’s Republic of China (PRC) and the Ministry of Public Security of PRC in 1998, “migrant children” are defined as children who can learn in 6–14 years old (or 7–15 years old) and live with parents or other guardians in the inflow place temporarily for more than 6 months. Based on this concept, 6–14 years of age (7–15 years of age) is defined as the selection range of childhood, and the children who migrate to this city with parents or other guardians are included, whereas children who do not migrant with their parents or guardians (i.e., left-behind children) are excluded.

Social integration: social integration is based on pluralism. When the culture of the place of migration is relatively inclusive, new immigrants are more inclined to maintain the original cultural value, and their identities and values are reshaped in the new settlement place, which help the formation of a diversified social and economic order, so that all members of society can enjoy fair and equal rights (Carolin et al., 2018; Moksony and Hegedus, 2018). Social integration covers cultural, economic, social, and psychological integration, and integration emphasizes the suitable construction of different groups with new cultural meanings (Noyens et al., 2018). Based on this, social integration is defined to build a good and harmonious education environment under the cooperation and support of society, school, and family, to promote the education development of migrant children. In addition, social integration is not to assimilate children into urban children but to build an educational environment suitable for migrant children. Therefore, the social integration of migrant children is defined as the action and process for good education and life of migrant children in the school through the mutual influences among education system, teaching curriculum, teachers, parents, students, and migrant children.

Construction of the Text Classification Model Base on Improved Conventional Neural Network

The CNN model can capture local features such as keywords and phrases in sentences, but the most imperative step in the CNN model is how to determine the size of the convolution kernel (Chen and Jahanshahi, 2018). Recurrent neural network (RNN), LSTM, and GRU models can better grasp the features in the text and find the optimal feature (Iso et al., 2018; Jujie and Yaning, 2018). The data of the input model do not affect the output result because of the introduction of attention into the neural network model, and thus, it can optimize the text feature vector and improve the classification effect. Therefore, the LSTM model and the GRU model are combined in parallel based on the attention mechanism in the CNN model, and a new classification model ALGCNN is proposed, whose basic structure is shown in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. Basic structure of ALGCNN model.

The input layer of the ALGCNN model is composed of different words, and the dimension of each word is 286, so each sentence is expressed as a feature map of d*l (l is the length of the sentence).

The LSTM in the network layer of LSTM and GRU is a variant of the RNN model and is mainly used to overcome gradient disappearance in the model. GRU model performs the simplification of LSTM. The basic structures of LSTM and GRU are illustrated in Figure 2 below.

FIGURE 2
www.frontiersin.org

Figure 2. Basic structures of LSTM (A) and GRU (B).

Figure 2A shows that LSTM includes an input gate, a forget gate, an output gate, and a storage unit. Different gate structures control the input, storage, and output of the data stream. In LSTM, if the matrix xt at time t is given, the mathematical expression calculated by the forget gate can be written as below equation:

f t = σ ( ω f [ y t - 1 x t ] + b f ) (1)

In the above equation, σ refers to the Sigmoid function, ωf is the weight matrix of the forget gate, yt–1 represents the output value at the previous moment, ⊕ is the splicing of different matrices, and bf represents the deviation value of the forget gate.

The input gate is mainly responsible for the input of current position of the sequence, and the calculation process through the input gate is expressed as Equation (2) below.

{ u t = σ ( ω u [ y t - 1 x t ] + b u ) v t = δ ( ω v [ y t - 1 x t ] + b v ) c t = f t c t - 1 + u t v t (2)

In the above equation, ut is the first step of the calculation (it must consider how much of the current input information needs to be stored in the storage unit), vt is the second step of the calculation (how much new information in the control input needs to be stored), δ refers to the Tanh activation function, ct is the update of the neuron, and b is the deviation value.

The output value obtained by the output gate can be transmitted to the next moment, and the calculation process of the output gate is given as below:

{ z t = σ ( ω z [ y t - 1 x t ] + b z ) y t = z t δ ( c t ) (3)

Figure 2B suggests that the GRU model does not contain a storage unit compared with the LSTM model. The calculation process of the GRU model is given in Equation (4).

{ R t = σ ( ω R [ y t - 1 x t ] ) T t = σ ( ω T [ y t - 1 x t ] ) y t ¯ = δ ( ω y ¯ [ R t y t - 1 x t ] ) y t = ( 1 - T t ) y t - 1 + T t y t ¯ (4)

In the above equation, Rt is the control unit of reset gate, Tt represents the control unit of update gate, and ⊗ is the multiplication of corresponding elements among the matrices.

Both the LSTM and GRU models in the constructed ALGCNN model are utilized to process the original text information, so the number of hidden neurons in the LSTM and GRU models is set to be the same.

Figure 1 shows that the constructed ALGCNN model contains the attention mechanism layer (Connie et al., 2018). After all elements are compared using the attention matrix A, the model will be overfitting due to too many parameters. Thus, values of all neurons in the model are summed in this article. If the output value of the LSTM and GRU model is Qmodel ∈ ℜns, then the model is ∈(LSTM, GRU). n in the equation refers to the number of hidden neurons in the LSTM and GRU model, and the matrix A can be calculated with below equation:

A i , j = E u c l i d e a n ( Q L S T M [ : , i ] , Q G R U [ : , j ] ) (5)

In the above Equation (5), Euclidean () refers to the Euclidean distance, and it can be calculated with Equation (6).

E u c l i d e a n ( u , v ) = 1 1 + | u - v | (6)

The matrix A can be calculated with Equation (5) above, and then, the attention weight value az,j = ΣA[j,:] of the jth neuron in the LSTM output and the attention weight value a1,j = ΣA[:,j] of the jth neuron in the GRU output can be generated.

After the attention mechanism layer, the ALGCNN model can realize convolution and pooling operations. If n new features cn are calculated from the convolution operation window on text feature xn:n+w–1, the calculation equation can be written as below:

c i = f ( U x n : n + w - 1 + b c ) (7)

In the Equation (7) above, bc refers to the deviation value, f() is a non-linear function, and the word window of the convolution kernel can generate a feature map c = c1c2⊕,⋯,⊕csw + 1.

The constructed ALGCNN model can equalize the pooling operation based on the attention mechanism. The calculation equation of the pooling operation is given as below.

Q i p [ : , j ] = k = j : j + w a i , k c i [ : , k ] (8)

In the above equation, Qipnsi, and i refers to LSTM or GRU model.

The last fully connected layer in the ALGCNN model is obtained by combining the results of LSTM and GRU pooling operations, and the fully connected layer is connected to the Softmax classifier with below equation.

{ p ( y | s ) = S o f t m a x ( ω s N + b s ) y = arg max y p ( y | s ) (9)

In the equation given above, N is the combined value of pooling results of LSTM and GRU, and y is the possible label value through the Softmax classifier.

Loss function of the model is the crossentropy loss function, which can be calculated with below equation:

S ( θ ) = - 1 K k = 1 K y k log ( y k ) + λ || θ || 2 (10)

In the equation above, K is the text category, yk is the true category label, yk is the label estimated by Softmax, and λ is the regularized hyperparameter of L2.

Finally, the model is optimized with the batch gradient descent (Wu et al., 2019) and Adam algorithm (Caltagirone and Frincke, 2005).

Model Evaluation Indicators

In this study, the model is evaluated regarding the Hamming loss (HL), precision (Pre), recall (Re), and micro-F1 (F1). HL is often used as an evaluation indicator for multilabel classification, which indicates the number of misclassifications and misclassifications of a label. The smaller it is, the better the performance of the model. The calculation method was given as follows:

H L = 1 N X O R ( Y i , j , P i , j ) L (11)

In the equation above, N is the number of samples, L refers to the total number of labels, Yi,j is the j-th label in the true label set of the i-th sample, Pi,j represents the j-th label in the predicted label set of the i-th sample, and XOR is an exclusive OR operation.

Pre represents the proportion of correct predictions in samples with positive predictions, and it reflects the precision of the model, which is calculated with below equation:

Pr e = T P T P + F P (12)

The Re indicates the proportion of the positive samples in the sample that are correctly predicted. It reflects the recall rate of the model, whose calculation method is expressed as follows. The greater the Pre and Re, the better the effect of the model.

R e = T P T P + F N (13)

In Equations (12) and (13), TP represents the original number of positive samples predicted to be positive, FP represents the original number of positive samples predicted to be negative, and FN represents the original number of negative samples predicted to be positive.

Micro-F1 is the weighted average of micro-precision and micro-recall. The larger the F1, the better the performance of the model. The calculation method is as follows:

F 1 = 2 × Pr e × R e Pr e + R e (14)

Development of Education Information Management System of Migrant Children

To fully grasp the number of migrant children and education-related information, a migrant children management module is developed based on Shenyang children information management client and big data platform. The data in the Integrated Management System of Migrant Population and Rental Housing of the community grid office, the Student Status Information Management System of the education department, and the Student Information Management System of the primary schools in Shenyang city are integrated for comparison and analysis to develop the Education Information Management System of Migrant Children.

The system includes four functional modules: data comparison, classification management, data statistics, and data export. The data comparison module uses the large integrated management data of the community grid office and the student information of the education department to obtain the data for the children of primary school. After such data are imported into the system, a code is generated automatically; and the automatic matching of information is completed by selecting one or more information personal items (including name, gender, date of birth, father’s name, and mother’s name) and one or more position information items (this region, city, and province). The data information is classified into “migrant children” and “local children” with the ALGCNN model constructed in section “Definitions of Related Concepts” after the data comparison is completed. Data statistics refers to collect the children’s age, regional division, school status, school results, and class teacher evaluations. Finally, children’s information, academic performance tables, and learning assessment without query results are exported in the form of Excel.

Statistical Processing and Analysis

SPSS22.0 software is utilized for data entry and analysis, mainly for descriptive statistical analysis of collected data, independent sample t-test, F-test, and correlation analysis. When p < 0.05, the difference is considered statistically significant.

Results and Discussion

Simulation Verification of ALGCNN Model

Simulation of ALGCNN model is verified using deep learning framework Tensorflow under Google open source, which integrates CNN, RNN, LSTM, and GRU models (Perez, 2019). The operating system for ALGCNN model simulation is Ubuntu 14.05, the development language is Python 3.6, and the framework used is Tensorflow 1.4.0. First, classification of the model is verified with MR, SST, and MPQA data sets. For MR data set, the text category is 2, the average sentence length is 20, the training set contains 9,595 words, and the testing set contains 1,050 words. For the SST data set, the text category is 5, the average sentence length is 18, the training set contains 9,845 words, and the testing set contains 2,010 words. For the MPOA data set, the text category is 2, the average sentence length is 3, the training set contains 9,500 words, and the testing set contains 1,105 words. The classification effects of LSTM (Chen et al., 2020), Bi-LSTM (Ye et al., 2019), Text CNN (Xie et al., 2020), Text RCNN (Huang et al., 2021), attention-based LSTM (ALSTM) (Liu et al., 2018), and attention-based Bi-LSTM (ABi-LSTM) (Song et al., 2018) are compared. The number of hidden layer neurons in the LSTM, Bi-LSTM, GRU, and RCNN model is set to 128, the text batch size is set to 100, the convolution kernel size in the CNN model is set to 3*3, and the number of convolution kernels is set to 128. The dimension of the attention layer is set to 128, the initial learning rate is set to 0.001, and the number of circuit training is 50. The differences in accuracy and loss values of different data sets are compared and analyzed in the text classification by each model, and the results are shown in Figures 35, respectively.

FIGURE 3
www.frontiersin.org

Figure 3. Comparison of classification performance of different models on the MR data set. (A) Is the comparison on classification accuracy; (B) is the comparison of classification loss value.

FIGURE 4
www.frontiersin.org

Figure 4. Comparison of classification performance of different models on the SST data set. (A) Is the comparison on classification accuracy; (B) is the comparison of classification loss value.

FIGURE 5
www.frontiersin.org

Figure 5. Comparison of classification performance of different models on the MPQA data set. (A) Is the comparison on classification accuracy; (B) is the comparison of classification loss value.

Figures 3A,B illustrate that the ALGCNN model shows the highest classification accuracy and the lowest loss value on the MR data set. There is no great difference in the classification accuracy and loss value between the ALSTM model and the ALGCNN model. The LSTM and Bi-LSTM models show the lowest classification accuracy and the largest loss values. The differences in the classification accuracy and the loss value of TextCNN, TextRCNN, and ABi-LSTM models are not obvious. At the same time, the constructed ALGCNN model has the fastest convergence speed.

Figures 4A,B suggest that the constructed ALGCNN model has the highest classification accuracy and the lowest loss value on the SST data set, followed by the ALSTM model, and the third is the TextRCNN model. The classification accuracy is the lowest and the loss is the largest for both LSTM model and Bi-LSTM model. In addition, the ALGCNN model shows the fastest convergence speed and lower probability of overfitting.

In Figures 5A,B, the constructed ALGCNN model shows the highest classification accuracy and the lowest loss value on the MPQA data set, followed by the ALSTM model, and the classification effects of the LSTM model and the ALGCNN model are not different so much. The LSTM and Bi-LSTM models show the lowest classification accuracy and the largest loss. At the same time, the ALGCNN model and the ALSTM model show the fastest convergence speed, and the ALGCNN model reduces the possibility of overfitting.

The difference in classification accuracy of different models on different data sets is quantitatively analyzed, and the results are shown in Table 1. Compared with the LSTM, Bi-LSTM, TextCNN, TextRCNN, ALSTM, and ABi-LSTM models, the classification accuracy of the ALGCNN model constructed on the MR data set is improved by 26.5, 32.3, 6.6, 3.9, 9.0, and 3.5%, respectively; the classification accuracy on the SST data set of the ALGCNN model constructed has increased by 25.0, 29.8, 2.0, 0.0, 14.2, and 8.0%, respectively; and the classification accuracy rate on the MPQA data set has increased by 20.6, 17.3, 6.0, 2.1, 9.2, and 2.9%, respectively.

TABLE 1
www.frontiersin.org

Table 1. Text classification accuracy of different models on different data sets.

It is found that the classification accuracy of the ALSTM and ABi-LSTM models after the attention mechanism is added is significantly improved, which shows that the adding of the attention mechanism can improve the ability of the network model to extract features and then enhance the model’s ability to classify text. The classification performance of the TextCNN model is not good, which may be caused by the fact that the English TextCNN model does not consider the correlation between text word vectors (Banerjee et al., 2019). The ALGCNN model shows the best text classification effect, because the model first adopts LSTM and GRU models for feature screening, then applies the attention mechanism to increase the model’s ability to extract features, and finally obtains the final classification results through merging. Thus, it can improve the accuracy of text classification.

Furthermore, the ALGCNN, ABi-LSTM, TextCNN, TextRCNN, ALSTM, and BERT model are compared regarding the average HL, Pre, Re, and F1s on different data set classifications (Figure 6). In different algorithms, ALGCNN shows the lowest HL (15.2 ± 1.38%), which is 1.1% lower than the BERT model. The Pre and F1 of the ALGCNN model are second only to the BERT model, and those of the Pre and F1 are 0.74 and 0.92% lower than the BERT model. The Re value of the ALGCNN model is higher than the highest among all models. The running time of the ALGCNN model is the shortest, which is 5.02 min shorter than that of the BERT model. Therefore, although the BERT model text classification accuracy rate and micro-F1 are better than the ALGCNN model, the training time of the ALGCNN model I is significantly shorter than that of the BERT model.

FIGURE 6
www.frontiersin.org

Figure 6. Comparison of evaluation index values of different models. (A) Showed the comparison of the HL value; (B) showed the comparison on Pre; (C) illustrated the comparison on the Re value; (D) was the comparison on F1; and (E) showed the comparison on training time.

Big Data Analysis of Migrant Children in Primary School of China

The data analysis results of children of migrant workers and children of migrant workers in primary school released by the Ministry of Education from 2015 to 2019 are shown in Figure 7. Figure 7A illustrates that the number of migrant children in different grades of primary school has shown a gradual increase from 2015 to 2019. In 2019, the number of migrant children in grade 1, 2, 3, 4, 5, and 6 increased by 0.76 × 105, 1.95 × 105, 0.83 × 105, 1.65 × 105, 2.52 × 105, and 2.72 × 105, respectively, compared with 2015. This may be because the younger generation of migrants is more inclined to keep their children around, so that the number of migrant children has shown an increasing trend year by year, which also points out that the number of left-behind children in rural areas is also gradually decreasing (Charania et al., 2019). Figure 7B reveals that the numbers of children of migrant workers in grades 5 and 6 have been increasing year by year, whereas the increase in the number of other grades has not changed much. The number of children of migrant workers in grades 1, 2, 3, 4, 5, and 6 in 2019 had increased by −0.50 × 105, 0.36 × 105, −0.31 × 105, 0.46 × 105, 1.21 × 105, and 1.63 × 105 people, respectively, compared with 2015.

FIGURE 7
www.frontiersin.org

Figure 7. Changes in number of migrant children in primary school in 2015–2019. (A) Shows the changes in number of migrant children; and (B) shows the changes in number of children of migrant workers.

Then, the changes in the number of migrant children with different sources in 2019 are compared. Figure 8A suggests that the number of migrant children from other counties in the province is significantly higher than the number of children from other provinces. With the increase of children’s learning grade, the numbers of children from different sources have shown gradual decrease, and the number of migrant children in grade 6 is significantly lower than that of other grades. Figure 8B indicated that among the children of migrant workers, the number of migrants from other counties in the province is also significantly higher than the number of migrants from other provinces. At the same time, with the increase of children’s learning grade, the number of children from different sources of immigration also discloses a gradual decrease. The number of migrant children in the grade 6 is significantly lower than other grades.

FIGURE 8
www.frontiersin.org

Figure 8. Changes in number of migrant children in primary school with different sources in 2019. (A) Shows the changes in number of migrant children; and (B) shows the changes in number of children of migrant workers.

The educational expenditure data published by the China Financial Yearbook are applied to compare the differences and change trends in the proportion of education expenditures undertaken by governments at all levels in China from 2015 to 2019. Figure 9 indicates that the district- and county-level government undertakes the largest proportion in different years. The central government assumes the smallest proportion. With the gradual increase in years, the proportion undertaken by the township-level government has shown a gradual decrease, whereas the proportion of county-level government has shown a gradual increase. Therefore, it is speculated that the education issue of migrant children is also a fiscal issue. At present, compulsory education still suffers from many problems such as lack of national education funds, low proportion of central fiscal expenditures, uneven distribution of central fiscal transfer payments, unfulfilled “provincial overall planning” policy of compulsory education funds, and insufficient public finance investment in private education.

FIGURE 9
www.frontiersin.org

Figure 9. The proportion of government education expenditures at all levels in 2015–2019.

Classification and Analysis on Information of Migrant Children

The school state and information of a total of 1,579 children in primary school are matched and classified from September 2018 to March 2019 in Shenyang city in the next semester, and the classification results of migrant workers’ children and local children are shown in Table 2. The accuracy of the constructed model is higher than 98% in both the classification of migrant children and local children.

TABLE 2
www.frontiersin.org

Table 2. Classification results based on ALGCNN model.

First, the differences between migrant children and local children in different grades and gender distribution are compared and analyzed in Table 3. There is no significant difference in distribution of migrant children and local children in grades 3, 4, 5, and 6, and it is the same case in gender (p > 0.05). However, the proportion of migrant children in grades 1 and 2 is significantly higher than that of local children (p < 0.05). The high proportions of migrant children in the grades 1 and 2 may be because the children are so young that the parents want to take them and care them themselves (Hu and Wu, 2020).

TABLE 3
www.frontiersin.org

Table 3. Comparison on distribution of migrant children.

Then, the differences in the final examination scores and average scores of migrant children and local children are analyzed and compared, and the distributions are shown in Figure 10. Figure 10A reveals that the final average scores of migrant children are mainly distributed between 40 and 60 points, which may be because that it takes time for children to adapt to the learning environment in new schools. This may be the main reason for the poor performance of migrant children. Figure 10B that the final average scores of local children are mainly distributed between 40 and 90 points, which is in line with the actual situation of children’s score distribution.

FIGURE 10
www.frontiersin.org

Figure 10. Descriptive statistics on final examination scores of children. (A) Illustrates the score distribution of migrant children; and (B) illustrates the score distribution of local children.

Third, the difference in the average final examination scores of migrant children and local children is illustrated in Figure 11. The average final examination score of local children is significantly higher than that of migrant children (p < 0.05).

FIGURE 11
www.frontiersin.org

Figure 11. Comparison on average final examination score of children. Indicates that the difference between two groups is obvious (p < 0.05).

The learning adaptability scores of migrant children and local children are analyzed. Figures 12A,B show that both migrant children and local children have lower scores in listening methods, learning skills, and adaptability to learning and teaching environment, whereas both have higher scores in terms of learning attitude and independent learning ability. After comparison, it is found that the differences are not obvious in each dimension score and total score between migrant children and local children (p > 0.05). Migrant children have a weak ability to adapt to the teacher’s teaching methods and the teaching schedule, which indicates that migrant children need to improve their listening methods during the learning process.

FIGURE 12
www.frontiersin.org

Figure 12. Comparison in children’s learning adaptability scores. (A) The scores of migrant children; (B) the score of local children; A refers to learning attitude; B refers to listening method; C refers to learning skills; D indicates the adaptability to the school teaching environment; and E represents the independent learning ability.

Conclusion

This research aims to explore the impact of China’s existing education policies on the educational integration of migrant children. First, ALGCNN model is constructed for text classification of children’s education information. After analysis of its performance, it is applied to text classification of migrant children and local children, and the influencing factors of migrant children’s education integration are discussed. The results show that ALGCNN model has a high text classification accuracy, and children’s education cost, way of attending classes, learning skills, and adaptability to learning environment are the main influencing factors for educational integration of migrant children. However, there are still some shortcomings in this study. This study only collects big data to conduct a comparative analysis on the education status of migrant children, but does not conduct a correlation regression analysis on the influencing factors of different dimensions of migrant children’s education integration and education policies. In the follow-up, based on the results and many field investigations, further regression analysis will be made on influencing factors of different dimensions of children’s educational integration and education policies, and reasonable suggestions will be given based on existing problems.

In terms of practical issues, it is also necessary to provide provisions and protection in the legislative field, especially for large immigrant provinces such as Hainan, Guangdong, Zhejiang, and so on. It is necessary to share information in the source provinces of local immigrants, establish and perfect local legislative guarantee mechanism, and provide legal support for improving education and fill in legislative vacancies.

In short, the results of this article can provide a theoretical basis for solving the education problems of migrant children and improving the education policies of migrant children.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

The studies involving human participants were reviewed and approved by the Ethics Committee of Northeastern University. The patients/participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author Contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work, and approved it for publication.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fpsyg.2021.762416/full#supplementary-material

References

Ajay, B., Jyoti, H., and James, K. (2018). ‘They had to go’: indian older adults’ experiences of rationalizing and compensating the absence of migrant children. Sustainability 10:1946.

Google Scholar

Banerjee, I., Ling, Y., Chen, M. C., Hasan, S. A., Langlotz, C. P., Moradzadeh, N., et al. (2019). Comparative effectiveness of convolutional neural network (CNN) and recurrent neural network (RNN) architectures for radiology text report classification. Artif. Intell. Med. 97, 79–88. doi: 10.1016/j.artmed.2018.11.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Barberá, P., Boydstun, A. E., Linn, S., McMahon, R., and Nagler, J. (2021). Automated text classification of news articles: a practical guide. Polit. Anal. 29, 19–42. doi: 10.1017/pan.2020.8

CrossRef Full Text | Google Scholar

Basiri, M. E., Nemati, S., Abdar, M., Cambria, E., and Acharya, U. R. (2021). ABCDM: an attention-based bidirectional CNN-RNN deep model for sentiment analysis. Future Gener. Comput. Syst. 115, 279–294. doi: 10.1016/j.future.2020.08.005

CrossRef Full Text | Google Scholar

Caltagirone, S., and Frincke, D. (2005). “ADAM: active defense algorithm and model,” in Aggressive Network Self-Defense, (Amsterdam: Elsevier).

Google Scholar

Carolin, R., Tim, H., Terje, E., and Stathopoulou, T. (2018). Social integration and self-reported health: differences between immigrants and natives in Greece. Eur. J. Public Health 28, 48–53. doi: 10.1093/eurpub/cky206

PubMed Abstract | CrossRef Full Text | Google Scholar

Charania, N. A., Paynter, J., Lee, A. C., Watson, D. G., and Turner, N. M. (2019). Vaccine-preventable disease-associated hospitalisations among migrant and non-migrant children in New Zealand. J. Immigr. Minor. Health 22, 223–231. doi: 10.1007/s10903-019-00888-4

PubMed Abstract | CrossRef Full Text | Google Scholar

Chen, C. W., Tseng, S. P., Kuan, T. W., and Wang, J. F. (2020). Outpatient text classification using attention-based bidirectional LSTM for robot-assisted servicing in hospital. Information 11:106. doi: 10.3390/info11020106

CrossRef Full Text | Google Scholar

Chen, F. C., and Jahanshahi, R. M. R. (2018). NB-CNN: deep learning-based crack detection using convolutional neural network and naïve bayes data fusion. IEEE Trans. Ind. Electron. 65, 4392–4400. doi: 10.1109/tie.2017.2764844

CrossRef Full Text | Google Scholar

Connie, L., Troller-Renfree, S. V., Zeanah, C. H., Nelson, C. A., and Fox, N. A. (2018). Impact of early institutionalization on attention mechanisms underlying the inhibition of a planned action. Neuropsychologia 117, 339–346. doi: 10.1016/j.neuropsychologia.2018.06.008

PubMed Abstract | CrossRef Full Text | Google Scholar

Dashtipour, K., Gogate, M., Li, J., Jiang, F., Kong, B., and Hussain, A. (2020). A hybrid persian sentiment analysis framework: integrating dependency grammar based rules and deep neural networks. Neurocomputing 380, 1–10. doi: 10.1016/j.neucom.2019.10.009

CrossRef Full Text | Google Scholar

Flensner, K. K., Korp, P., and Lindgren, E. C. (2021). Integration into and through sports? sport-activities for migrant children and youths. Eur. J. Sport Soc. 18, 64–81. doi: 10.1080/16138171.2020.1823689

CrossRef Full Text | Google Scholar

Griffiths, D., and Boehm, J. (2019). A review on deep learning techniques for 3d sensed data classification. Remote Sens. 11:1499. doi: 10.1007/s11548-019-02097-8

PubMed Abstract | CrossRef Full Text | Google Scholar

Hu, B., and Wu, W. (2020). Parental support in education and social integration of migrant children in urban public schools in China. Cities 106:102870. doi: 10.1111/cch.12370

PubMed Abstract | CrossRef Full Text | Google Scholar

Hu, H., Gao, J., Jiang, H., and Xing, P. (2018). A comparative study of unintentional injuries among schooling left-behind, migrant and residential children in China. Int. J. Equity Health 17:47. doi: 10.1186/s12939-018-0767-3

PubMed Abstract | CrossRef Full Text | Google Scholar

Huang, W., Su, C., and Wang, Y. (2021). An intelligent work order classification model for government service based on multi-label neural network. Comput. Commun. 172, 19–24. doi: 10.1016/j.comcom.2021.02.020

CrossRef Full Text | Google Scholar

Iso, S., Shiba, S., and Yokoo, S. (2018). Scale-invariant feature extraction of neural network and renormalization group flow. Phys. Rev. E 97:053304. doi: 10.1103/PhysRevE.97.053304

PubMed Abstract | CrossRef Full Text | Google Scholar

Jørgensen, C. R., Dobson, G., and Perry, T. (2021). Migrant children with special educational needs in European schools–a review of current issues and approaches. Eur. J. Special Needs Educ. 36, 438–453. doi: 10.1080/08856257.2020.1762988

CrossRef Full Text | Google Scholar

Jujie, W., and Yaning, L. (2018). Multi-step ahead wind speed prediction based on optimal feature extraction, long short term memory neural network and error correction strategy. Appl. Energy 230, 429–443. doi: 10.1016/j.apenergy.2018.08.114

CrossRef Full Text | Google Scholar

Kyereko, D. O., and Faas, D. (2021). Integrating marginalised students in Ghanaian schools: insights from teachers and principals. Compare J. Comp. Int. Educ. 1–18. doi: 10.1080/03057925.2021.1929073

CrossRef Full Text | Google Scholar

Liang, D., and Yi, B. (2021). Two-stage three-way enhanced technique for ensemble learning in inclusive policy text classification. Inf. Sci. 547, 271–288. doi: 10.1016/j.ins.2020.08.051

CrossRef Full Text | Google Scholar

Liang, Y., Zhou, Y., and Liu, Z. (2019). Traumatic experiences and posttraumatic stress disorder among Chinese rural-to-urban migrant children. J. Affect. Disord. 257, 123–129. doi: 10.1016/j.jad.2019.07.024

PubMed Abstract | CrossRef Full Text | Google Scholar

Lima, M. D. D., Lima, J. D. O. R. E., and Barbosa, R. M. (2020). Medical data set classification using a new feature selection algorithm combined with twin-bounded support vector machine. Med. Biol. Eng. Comput. 58, 519–528. doi: 10.1007/s11517-019-02100-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, J., Wang, G., Duan, L. Y., Abdiyeva, K., and Kot, A. C. (2018). Skeleton-based human action recognition with global context-aware attention LSTM networks. IEEE Trans. Image Process. 27, 1586–1599. doi: 10.1109/TIP.2017.2785279

PubMed Abstract | CrossRef Full Text | Google Scholar

Moksony, F., and Hegedus, R. (2018). Religion and suicide: how culture modifies the effect of social integration. Arch. Suicide Res. 23, 151–162. doi: 10.1080/13811118.2017.1406830

PubMed Abstract | CrossRef Full Text | Google Scholar

Noyens, D., Donche, V., Coertjens, L., Van Daal, T., and Van Petegem, P. (2018). The directional links between students’ academic motivation and social integration during the first year of higher education. Eur. J. Psychol. Educ. 34, 67–86. doi: 10.1007/s10212-017-0365-6

CrossRef Full Text | Google Scholar

Perez, R. O. (2019). Using tensorflow-based neural network to estimate GNSS single frequency ionospheric delay (IONONet). Adv. Space Res. 63, 1607–1618. doi: 10.1016/j.asr.2018.11.011

CrossRef Full Text | Google Scholar

Schrier, L., Wyder, C., Torso, S. D., Stiris, T., and Ritz, N. (2019). Medical care for migrant children in Europe: a practical recommendation for first and follow-up appointments. Eur. J. Pediatr. 178, 1–19. doi: 10.1007/s00431-019-03405-9

PubMed Abstract | CrossRef Full Text | Google Scholar

Song, S., Lan, C., Xing, J., Zeng, W., and Liu, J. (2018). Spatio-temporal attention-based LSTM networks for 3D action recognition and detection. IEEE Trans. Image Process. 27, 3459–3471. doi: 10.1109/TIP.2018.2818328

PubMed Abstract | CrossRef Full Text | Google Scholar

Stadelmannsteffen, I. (2018). Migrant Children in Switzerland - how integration and school policy affects their educational chances. Virology 436, 89–99.

Google Scholar

Vargas-Valle, E. D., and Aguilar-Zepeda, R. (2020). School integration of migrant children from the united states in a border context. Int. Migr. 58, 220–234. doi: 10.1111/imig.12696

CrossRef Full Text | Google Scholar

Vogel, L. (2018). Health professionals decry detention of migrant children in Canada. Can. Med. Assoc. J. 190:E867. doi: 10.1503/cmaj.109-5630

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, T. (2020). Rural migrants in China: barriers to education and citizenship. Intercultural Educ. 31, 578–591. doi: 10.1080/14675986.2020.1794121

CrossRef Full Text | Google Scholar

Wang, X., Shan, T., Pang, S., and Li, S. (2019). Assessment of the genetic relationship between the recently established benthic population and the adjacent floating populations of Sargassum horneri (Phaeophyceae) in Dalian of China by newly developed trinucleotide microsatellite markers. J. Appl. Phycol. 31, 3989–3996. doi: 10.1007/s10811-019-01853-2

CrossRef Full Text | Google Scholar

Wu, H., Li, D., and Cheng, M. (2019). Chinese text classification based on character-level CNN and SVM. Int. J. Intell. Inf. Database Syst. 12:212. doi: 10.1504/ijiids.2019.10024507

CrossRef Full Text | Google Scholar

Xie, J., Hou, Y., Wang, Y., Wang, Q., Li, B., Lv, S., et al. (2020). Chinese text classification based on attention mechanism and feature-enhanced fusion neural network. Computing 102, 683–700. doi: 10.1007/s00607-019-00766-9

CrossRef Full Text | Google Scholar

Ye, H., Cao, B., Peng, Z., Chen, T., Wen, Y., and Liu, J. (2019). Web services classification based on wide & Bi-LSTM model. IEEE Access 7, 43697–43706.

Google Scholar

Keywords: convolutional neural network algorithm, attention mechanism, text classification model, migrant children, educational integration

Citation: Zhang C, Wang G, Zhou J and Chen Z (2022) The Influencing Legal and Factors of Migrant Children’s Educational Integration Based on Convolutional Neural Network. Front. Psychol. 12:762416. doi: 10.3389/fpsyg.2021.762416

Received: 21 August 2021; Accepted: 10 December 2021;
Published: 10 January 2022.

Edited by:

Chia-Chen Chen, National Chung Hsing University, Taiwan

Reviewed by:

Hsin-Te Wu, National Ilan University, Taiwan
Kia Dashtipour, University of Glasgow, United Kingdom

Copyright © 2022 Zhang, Wang, Zhou and Chen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Gang Wang, MTU2NTI5ODcyQHFxLmNvbQ==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.