Skip to main content

ORIGINAL RESEARCH article

Front. Psychol., 22 July 2022
Sec. Educational Psychology

Learning Behavior Evaluation Model and Teaching Strategy Innovation by Social Media Network Following Learning Psychology

\r\nLijuan YuanLijuan Yuan1Hongming LiHongming Li2Shiman FuShiman Fu2Zizai Zhang*Zizai Zhang3*
  • 1College of Journalism and Communications, Zhoukou Normal University, Zhoukou, China
  • 2Southampton Education School, University of Southampton, Southampton, United Kingdom
  • 3Hangzhou Preschool Teachers College, Zhejiang Normal University, Hangzhou, China

With the development of various network technologies and the spread of coronavirus disease 2019, many online learning platforms have been built. However, some of them may negatively impact student learning outcomes. Therefore, this study aims to improve the online learning effect of students by comprehensively evaluating their learning behavior by using deep learning algorithms. On this basis, new teaching strategies are proposed. According to the structured deep network embedding model, a network representation learning algorithm is proposed with the help of auto-encoders under deep learning. This study elaborates the concept and structure of the encoder model and tests its performance. After the node labels and dataset are trained, the applicable parameter λ2 of the model is 0.3. During the teaching process, the model’s reliability in distinguishing users is examined. Therefore, this model can be applied to network teaching, is an innovative teaching strategy, and provides a theoretical basis for improving teaching methods.

Introduction

With the rapid development of network technology, social media networks play an increasingly important role in people’s daily life, work, and entertainment and become indispensable (Zahara et al., 2021). Social media networks can meet people’s communicative needs without meeting the person face to face. The scale of social media networks and the size of users continue to expand every year. These changes bring unprecedented challenges to the use of network media (Hoi et al., 2020). In particular, the scale of social media networks expands sharply and reaches an unprecedented level due to coronavirus disease 2019 (COVID-19) (Adnan and Anwar, 2020). For example, network platforms such as nails and the superstar learning network platform have become the most frequently used online teaching platforms (Kent and Rechavi, 2020).

There are many forms of networks in the real world, such as citation networks, social networks, etc. These networks describe a certain relationship between each object and other objects and have important practical significance for data analysis and practical applications. Traditional research methods usually describe a network abstractly as a discrete adjacency matrix. Among them, each vertex of the matrix represents an object in the network. Each element in the matrix represents the adjacency relationship between objects (Sun et al., 2021). However, this representation not only needs to calculate the relationship between each pair of data, but also is difficult to describe the more complex structural relationship between the data, such as the path relationship between nodes and so on. This makes this class of methods problematic when dealing with large-scale network tasks. On the one hand, the adjacency matrix needs to calculate the adjacency relationship between each pair of samples, which will greatly increase the computational cost (Liang and Markchom, 2022). On the other hand, this simple adjacency relationship representation is difficult to describe more complex relationships between data, which is difficult to accurately describe the information of network structure. Therefore, how to construct a graph that can accurately describe the structural relationship of network vertices and reduce the high computational cost has attracted extensive attention. Network representation learning algorithms embed the graph into a vector space, which preserves the structure and inherent properties of the graph (Chen et al., 2022). Network representation learning is mainly used to learn vector representations of real data in low-dimensional latent spaces. Additionally, the original topology structure of the real network, including the information on vertices and edges, is maintained during the learning process. Thereafter, traditional machine learning algorithms are applied to the learned low-dimensional space for subsequent network analysis tasks. This representation method can not only describe the complex network structure more accurately but also transform the analysis task of the original network into the analysis of the potential low-dimensional network space, which is beneficial to saving the computational cost (Guo et al., 2021; Nguyen et al., 2021).

How to provide effective teaching strategies for students is still an urgent problem to be solved. Since many online platforms only set up the live broadcast function to provide students with the real-time viewing process of online teaching, the functions for students’ after-class course playback, after-class homework, and guidance are relatively lacking. Therefore, the auto-encoder (AE) in deep learning (DL) is used to establish a model similar to the neural network, and a network representation learning technology is proposed. According to the concept of DL, the establishment process of the model is explained, and the accuracy of node classification of the model is analyzed. Then, the classification ability for normal users and unrelated users is analyzed to identify the applicability of the model. In order to gradually narrow the educational gap among regions, urban and rural areas, and schools, Xingnan Primary School has determined the working idea of “taking platform construction as the basis, focusing on innovative application, leading demonstration as the starting point, deeply integrating information technology and education, and promoting joint construction and sharing of high-quality educational resources.” Schools should constantly strengthen the construction of the platform, innovate network teaching and research methods, and improve the effect of network classroom teaching. The innovation is to use DL to classify normal users and unrelated users.

Literature Review

Social media network platforms play different roles in people’s lives. For example, Facebook is for entertainment as a WeiBo platform, Zhihu, and Twitter are forums for people to express their views, and nails are used as learning platforms. In recent years, many research experts have been exploring the relationship between teaching quality and teaching methods. Wu and Chen (2021) introduced a method of designing courses in the case of limited prior knowledge for NPD courses. In their study, potential semantic analysis is used to extract the main research topics from journals, which provides a new basis for the innovation and management of curriculum development (Wu and Chen, 2021), Wu et al. (2017) studied the relationship between the entrepreneurial knowledge system and traditional teaching methods and gave some suggestions on the future theoretical direction of the entrepreneurial knowledge system, improving the acceptance of entrepreneurship and promoting economic development. There are also some studies that focus on the psychological factors in online learning. Rogoza et al. (2018) analyzed the influence of fragile psychology, narcissism, admiration, and competition on the teaching quality in the learning process. The results show that fragile narcissism and admiration are contradictory to shyness and loneliness, and liking competitions reflects that the person has low empathy. Similarly, Wang et al. (2019a) explored the relationship between the driving forces behind the entrepreneurial intention of online teaching and practice and revealed that psychological states had a certain impact on entrepreneurship education and practical education. In addition, many experts discussed the learning effects of different online teaching methods on students’ academic performance. For example, Lai et al. (2019) conducted a study on whether learning media affect students’ academic performance and participation and found that the learners who participated in the teaching activities actively had better academic achievements. Some innovative suggestions for students’ teaching strategies through social media networks will be proposed.

Early research on network representation learning was mainly based on factorization methods. This type of method mainly adopts the idea of graph embedding to explore the manifold structure of the data (Chen et al., 2021). On the premise that the input data are independent and identically distributed, the graph embedding method believes that the data has a manifold structure, so the similarity between the data is calculated first. A graph describing the relationship of the network structure is constructed, and then the relationship graph is embedded in the low-dimensional subspaces (Fu et al., 2021). The local linear embedding method describes the reconstruction relationship between local neighbor nodes with a graph, and maintains this local reconstruction relationship in a low-dimensional space. The Laplacian feature mapping method enables the latent low-dimensional space to still maintain the data similarity relationship in the original space, which is used to preserve the nonlinear structure of the data (Gao et al., 2021). However, when such methods are applied to large-scale networks, the computational complexity is greatly affected by the number of vertices, and the scalability is limited. Shang et al. (2021) proposed a multilayer network representation learning method (NRLM) for drug-target interaction prediction. The method can integrate useful information from different networks, reduce noise in multilayer networks, and learn feature vectors for drugs and targets. Eigenvectors of drugs and targets are put into the drug-target space to predict potential drug-target interactions. The multilayer NRLM and prediction accuracy are improved by adding parameter regularization constraints. Multi-label learning has received extensive attention for its applicability to machine learning problems. In recent years, in order to improve the efficiency of multi-label classification, scholars have proposed several methods based on extreme value learning machine (ELM) or Radial Basis Function (RBF) neural network. Most of the existing multi-label learning algorithms mainly focus on the information of the feature space. Rezaei-Ravari et al. (2021) used two neural network structures, namely multi-label RBF and multi-label multilayer ELM. Then, a regularized multi-label learning method based on feature manifold learning and a regularized multi-label learning method based on two-manifold learning are established to train the two network structures. Experiments show that two-manifold learning is used as a training method for neural classifiers, and using some cutting-edge multi-label techniques can significantly improve classification efficiency.

In short, the NRLM based on matrix decomposition can reconstruct the network, but it is prone to overfitting, and the overall performance is not satisfactory. Although the network representation learning algorithms based on natural language models can use the network structure to learn network representations, they are all shallow models. This also means that it is difficult for them to learn the deeper and more complex features of the network structure. DL has developed rapidly in recent years and has made important progress in many fields. Its essence is to perform deep abstraction on data features and learn the mapping function from high-dimensional data to low-dimensional features. Network representation learning can also be regarded as the process of converting the representation of nodes from the high-dimensional space of the original network to a low-dimensional vector space. The essential problem is to learn the mapping function between two vector spaces. Therefore, AE in DL are used to build models like neural networks, and a network representation learning technique is proposed. The purpose is to study how the relevant technology is used.

Materials and Methods

Auto-Encoder Under Deep Learning

The essence of the network representation learning proposed is to transform each node in the high-dimensional space into a vector in a relatively low-dimensional space, that is, the mapping function between every two vectors is learned and applied. AE is a model similar to a neural network model (Wang and Cha, 2021), and its structure is shown in Figure 1. It regards the hidden layer in the neural network as one of the encoders and decoders to process the data imported into the network model, which need to be encoded and decoded. When the data reach the output layer, it should be confirmed that the output data are consistent with the input prototype data. In other words, the hidden layer is responsible for ensuring that the processed data are consistent with the original data because it can identify the characteristics of the input data and retain these characteristics and features. AE is a neural network model to realize that the data are input. Finding the main factors that represent the prototype data is an important work to identify the characteristics of the input data. A variety of encoders are combined to form a stack auto-encoder (SAE) (Wang et al., 2020).

FIGURE 1
www.frontiersin.org

Figure 1. Structure of auto-encoder.

The performance parameters of the model can be obtained after the model is trained. If a training set has m training samples x1,x2,…,xn, they must meet the condition of xiRN.

The operation process is that xi passes through weight matrix W1RKxN in the input layer and the hidden layer, where K is the number of hidden nodes. Then, the data are encoded by bias vector b1RK, and the characteristics of the K dimension in the hidden layer are obtained. The equation is as follows:

f ( x i ) = σ ( w 1 x i + b 1 ) (1)

σ(ldot) is a nonlinear activation function.

When the data are decoded, the weight matrix W2 ∈ RNxK and offset vector b2 ∈ RN in the hidden layer and output layer can be used to construct the vector with dimensions N and K from the feature data. The equation is:

g ( x i ) = σ ( W 2 * f ( x i ) + b 2 ) (2)

The whole training process is to constantly adjust weight matrices W1 and W2 and bias vectors b1 and b2, and the objective function is minimized by:

L S A E = 1 2 m i = 1 m || g ( f ( x i ) ) - x i || 2
+ τ 2 | + | W 1 | | 2 + τ 2 || W 2 || 2 (3)

||…||2 represents the square of norm L2, and τ is a hyperparameter.

The structure of contractive auto-encoder (CAE) (Aamir et al., 2021) is similar to that of AE. The difference is that a new penalty term for error reconstruction is added, and it is viewed as the activation function for input data. Its main function is to capture the changes in training data samples. The advantage of CAE is that the extracted features are robust to the subtle changes in the training samples. The activation function of the encoder is used to improve the robustness of the data, which is realized by:

|| J ( x ) || F 2 = i j ( f i ( x i ) x i ) 2 (4)

In the above equation, J(x) represents the Jacobian matrix (Yan et al., 2018), j represents the total number of hidden layers, so the total number of objective functions of the CAE is as follows:

L C A E = 1 2 m i = 1 m || g ( f ( x i ) ) - x i || 2 + τ || J ( x ) || F 2 (5)

In Equation 5, the first half on the right side of the equal sign is the minimum reconstruction error, and the second half is to make the model network unchanged locally. The value of the partial derivative should be minimized in the calculation. If the partial derivative is 0, the model is robust to the vibration of the local part.

Concept of the Network Representation Learning Method

Network representation learning method (Zhang et al., 2018) is learning and mastering the node vector in the network. It is a feature learning method, which is the same as learning of an image or text. The code of the assignment group is regarded as the vectors for the network, and it is similar to the node characteristics in the network. The judgment network represents the quality of learning, and the indicators are compared by edge prediction and visual networks. The advantages of node vectors are shown in Table 1.

TABLE 1
www.frontiersin.org

Table 1. Advantages of node vectors.

Social media networks (Üsküplü et al., 2020) are used. If two people have a common interest, they will be more likely to have similar hobbies. In Figure 2, nodes 2 and 3 represent two people, and they are directly connected. However, node 1 and node 2 have a common interest, so are closer they are closer in the low dimension space.

FIGURE 2
www.frontiersin.org

Figure 2. Network proximity.

When the adjacent connection matrix of a network model is widely used to describe a network topology, each row vector or column vector can represent one of the nodes, so these vectors can be used to represent the node, and | V| represents the dimensions of the node, namely the total number of all nodes. This symbolic method has the same problem with coding in natural language processing (NLP), that is, it is difficult to choose an appropriate word in another language system, which means that the representation between any two nodes is not completely equivalent. The ultimate goal of NRLM is to use a low-dimensional thought space to describe the whole network. The decomposition algorithm of the matrix is also to learn the low-dimensional ordered space of the original matrix. The purposes of the two are similar, and different algorithms are applied to different matrix decompositions, like Laplacian matrix.

The Laplacian matrix (Belgacem et al., 2021) demonstrates that each data node can be calculated by the adjacent linear weighted combination, and the local weight matrix of each grounding is obtained by the adjacent matrix processing. In the node space, the same weight matrix is used to construct the adjacent points of the node, and the dimensionality reduction process is transformed into the feature decomposition process. The objective function is as follows:

( Y ) = i | Y i - j W i j Y j | 2 (6)

In the above equation, Yi and Yj represent the representation vector of the node, and Wij is the weighted value of Xj. If it is adjacent to node i, the value of Wij is between 0 and 1. If node i is not adjacent to node j, the value of Wij is 0, indicating jWijYj is the weighted reconstruction process of the node.

Laplace feature mapping algorithm (Acosta-Cabronero et al., 2018) aims at the approximate similarity of adjacent nodes in the original space where the adjacent matrix is located in the implicit feature space. The objective function is calculated by:

m i n i , j || y i - y j || 2 A i j (7)

yi and yj are the representation vectors of nodes i and j in the hidden space, and A is the adjacent matrix of the entire network. After deduction, the following objective function can be obtained by:

m i n i , j || y i - y j || 2 A i j = 2 t r a c e ( Y T L Y ) (8)

In the above equation, L represents the Laplacian matrix of the network.

Network Structure of Network Representation Learning Model

Network representation learning method is based on the shallow network model, which cannot meet the requirements of the new era. The goal of DL and NRLM is to learn the mapping function and vector in the feature space. Therefore, the two are combined, and NRLM based on DL is constructed.

A model called Structural Deep Network Embedding (SDNE) (Wang et al., 2019b) is to obtain high-latitude and nonlinear structural data features in the network to solve the problems in shallow networks. Figure 3 shows the structural schematic of SDNE, which uses SAE to retain adjacent data features of nodes.

FIGURE 3
www.frontiersin.org

Figure 3. Structure diagram of Structural Deep Network Embedding.

After adjacent matrices are input, the adjacent vectors of nodes can be represented by several hidden layers and nonlinear activation functions. The equations for each hidden layer are as follows.

y i ( 1 ) = σ ( W ( 1 ) x i + b ( 1 ) ) (9)
y i ( k ) = σ ( W ( k ) y i k - 1 + b ( k ) ) , k = 2 , , K (10)

In Equations 9, 10, W and b represent the parameters in the hidden layer according to the frame and the bias, and y is the hidden layer. The non-zero value of the input adjacent vector needs to be weighted to reconstruct node x^i of SAE because the network model is sparse. And more penalty weights are given to the value of restoring 1 to 0, and the penalty vector is calculated by:

b i = { b i j } j = 1 2 (11)

The values in the penalty vector are known, so the objective function of the second proximity is achieved by:

L 2 n d = i = 1 n || ( x ^ i - x i ) b i || F 2 (12)

When the objective function is designed, it is necessary to ensure that the implicit vectors in the encoders of the two nodes are closer, so that the topological characteristics of the first proximity are preserved. The designed objective function is as follows:

L 1 s t = i , j = 1 n s i , j || ( y i ( K ) - y i ( K ) ) b i || 2 (13)

In the above equation, si,j represents that whether the two nodes i and j are connected. If they are connected, the value is 1; otherwise, it is 0. Whenever the first proximity is calculated, the Laplace feature mapping is used, and the objective function is obtained by:

L 1 s t = 2 t r ( Y T L Y ) (14)

In Equation 14, tr() is the matrix trace. The final objective function of the network model is obtained by combining the objective function of the first proximity with the objective Korean of the second. The equation is as follows:

L m i x = L 1 s t + α L 2 n d + γ L r e g (15)

Lreg is the normalized term of the parameters in the model, and α and γ are the super-parameters, which is the optimal value that needs to be obtained by adjusting the parameters.

Construction of Network Representation Learning Method Based on Deep Learning

Similarity partition is a basic task to be done in feature computing. Network node partition is realized by using training data, and the similarity between samples are obtained. But the traditional metric learning method (TMLM) just obtains a linear Mahalanobis distance, and it cannot show the nonlinear structure of the sample. In recent years, a kind of depth measurement learning technology is widely used in this field. The objective function is optimized and the measurement space is better divided by establishing a deep neural network (Bianco et al., 2018).

Therefore, the proposed NRLM based on DL has the following characteristics: (1) the network model can maintain the adjacent relationship between the original network nodes after the representation is embedded; (2) embedding the network model (Shi et al., 2018) can make each node of the same type close to each other, and the nodes of different types can be gradually kept away. As shown in Figure 4, NRLM based on DL has three parts, which are the joint training update link (JTUL), the generating node sequence link (GNSL), and the network random walk link (NRWL). Based on the existing algorithms, the similarity of each node is measured to achieve the goal of mutual proximity of the same type of nodes and mutual distance of different types of nodes.

FIGURE 4
www.frontiersin.org

Figure 4. Structure diagram of network representation learning method based on deep learning [(A) network random walk link (NRWL), (B) generating node sequence link (GNSL), and (C) joint training update link (JTUL)].

As for the objective function of the model, if it is known that there are a small number of node labels, the vector initialization distance is shortened by selecting different types of nodes as the initialization nodes and expressing the nodes as vectors before training the model. Whenever different types of nodes are marked, node pairs are built to learn in a specific dimension space.

For each node target vi, the sequence of upper and lower nodes in the sliding window w is wv1 = vi–w, vi–w + 1,…, vi + w. The model can maximize the probability of node walking sequence and obtain the embedding vector of network nodes by training. The embedding loss function of the model is obtained by:

L u = - v i V log p ( v i - w , v i - w + 1 , , v i + w | v i ) (16)

After the above equations are integrated, the simplified equation is obtained:

L u = - v i V k = - w w log p ( v i + k | v i ) (17)

In the above equation,p(vi+k|vi) is the probability of the upper and lower vi + k nodes in the target which can be observed when target nodes vi is known. The calculation equation is as follows:

p ( v i + k | v i ) = exp ( v i + k T v i ) v V exp ( v T v i ) (18)

The gradient algorithm of loss in the above equation uses the construction of the Huffman coding tree (Moffat, 2019) to obtain the updated nodes.

In the process of training the model, k nodes are randomly selected from each type of node to form s. For a pair of different types of nodes in a node vi and s, if the two nodes have the same label, their indicator function is 1. Otherwise, it is 0. The metric learning is to maximize the similarity of the same type of nodes and approach each other during training, while the similarity of different types of nodes is minimized and kept away from each other during training. Figure 5 shows the structure diagram of metric learning. In d dimension, different types of nodes have different distances, and three colors represent three different types of node sets.

FIGURE 5
www.frontiersin.org

Figure 5. Classification accuracy of different nodes for the labels on the same dataset.

The whole training process of the model is to minimize the loss function. In the joint training, the random gradient descent method (RGDM) is used to combine the gradient of embedding loss and measuring loss and then reversely transmit them to the network model. For the problem that the computation of all nodes is complex, the negative sampling method (NSM) is used, and the whole training process is shown in Table 2.

TABLE 2
www.frontiersin.org

Table 2. Training process of the network model.

Training Evaluation Indicators

For the evaluation indicators of NRLM, the commonly used methods are the multi-node label classification method (MNLCM) and visualization or prediction link (VOPL). MNLCM is used to evaluate the model. The calculation equation are as follows:

p r e c i s i o n = T P ( A ) T P ( A ) + F P ( A ) (19)
r e c a l l = T P ( A ) T P ( A ) + F N ( A ) (20)
F 1 ( A ) = 2 p r e c i s i o n * r e c a l l p r e c i s i o n r e c a l l (21)
M a c r o - F 1 = A C F 1 ( A ) | C | (22)

In the above equation, TP (A) is the positive classification case for result A. FP (A) is the negative classification case, and its result is not A1. c is the aggregation of all label nodes, F1(A) is value F1 of label A after calculation. Macro-F1 is the evaluation indicator that treats the weight of various label nodes equally and can comprehensively evaluate the performance of hard node labels.

Results

Classification Accuracy of Different Nodes for Labels on the Same Dataset

The performance of the network model needs to be tested. The same kind of datasets is used and the classification accuracy of different nodes on labels is used to judge the applicable performance of the model. The datasets are Doc2vec (node vector representation), Deepwalk (network structure information), and Node2vec (breadth search strategy). The Doc2vec method is an unsupervised algorithm that learns fixed-length feature representations from variable-length text such as sentences, paragraphs, or documents. It can obtain the vector representation of sentences, paragraphs, and documents and is an extension of Word2Vec. It has some advantages, such as not fixing the sentence length and accepting sentences of different lengths as training samples. Deepak is to solve the problem of node embedding in the graph. Graph Embedding technology expresses the nodes in the graph in the form of low-dimensional dense vectors and requires that the similar nodes in the original graph (different methods have different definitions of similarity) are also close in the low-dimensional expression space. Expression vectors can be used for downstream tasks such as node classification, link prediction, etc. Node2vec is a model used to generate node vectors in the network, and the input is the network structure (which can be weightless). The output is the vector of each node. The main idea is to directly guide the word2vec package, sample through a specific walking method, and generate a corresponding sequence for each point. These sequences are treated as text and imported into the bow or skip-gram model in word2vec, and the vector of each node (corresponding to the vector of each word in word2vec) can be obtained. The test results are shown in Figure 5.

The analysis of the data results in Figure 5 shows that with the increase of the number of training sets (10→90%), the classification accuracy of the model for three node labels is improved, and the accuracy of Doc2vec is 0.7249 when the training set is 90%. The accuracy of Deepwalk is 0.5764 when the training set is 90%; the accuracy of Node2vec is 0.5125 when the training set is 90%. Therefore, it is believed that this network representation learning algorithm can make full use of the existing label information to embed network nodes. The final generated vector has better classification effect. The application of the algorithm in practical work can also reflect the role of label nodes. Whether user classification is necessary depends on whether it is always beneficial to user research goals. In many cases, placed in a specific task scenario, the user’s behavior will not show essential differences in the face of a specific product. It is necessary to understand how “behavior” arises to understand this problem. Any kind of human behavior is the result of the combined action of dozens of internal and external factors. In different scenarios, although there are still so many types of influencing factors, the influence weight of each factor will change. In the process of user interaction with a specific product, the constraints formed by the product’s own characteristics may become the most important factor affecting user behavior, rather than some characteristics of the user itself, such as gender, age, or personality. Since the characteristics of a particular product are unique, people with different intrinsic characteristics may behave similarly to the same product. At this point, the value of user research lies in observing and discovering how users react to these restrictive factors of the product rather than studying how to classify people.

Classification Effect of the Same Dataset in Different Parameter Weights

The dataset used is the Flickr social network dataset. This dataset contains social network relationship data scraped by websites. In this dataset, each node is a user in Flickr, and each edge is a friendly relationship between users. Additionally, each node has a label that identifies the user’s interest group. The model evaluation indicators in part 3.5 are used to judge the effect of the model to compare the classification effects of the same dataset under different parameter weights λ2. The specific test results are shown in Figure 6.

FIGURE 6
www.frontiersin.org

Figure 6. Recognition effect of model Macro-F1 value under different parameter weights.

Figure 6 shows that when the value of λ2 is 0, the learning loss of non-fusion measure is shown. If λ2 is 0.3, the node classification effect of the model is the highest. When λ2 is small, this learning method has no obvious effect on the classification. When λ2 is 0.4, the weight is too high in the classification process, which affects the classification effect of adjacent labels. Therefore, parameter weight λ2 is recommended to be set as 0.3.

In Figure 7, as the parameter weight increases, the recognition effect of the model F1 value is gradually enhanced. When the parameter weight λ2 is 0.1, the F1 value of the model is 30%. When the parameter weight λ2 is 0.4, the F1 value of the model is 64.4%. After that, the F1 value of the model can stabilize at around 64%. The data show that the proposed model can achieve stability under certain parameter weights.

FIGURE 7
www.frontiersin.org

Figure 7. Recognition effect of model F1 value under different parameter weights.

In Figure 8, under different parameter weights, the recognition accuracy of the model will gradually increase. When the parameter weight λ2 is 0.1, the recognition accuracy of the model is only 60%. In addition, after the parameter weight λ2 is increased to 0.3, the average recognition accuracy of the model can reach about 84%.

FIGURE 8
www.frontiersin.org

Figure 8. Change curve of model recognition accuracy under different parameter weights.

In Figure 9, the larger the parameter weight, the better the recognition recall of the model. When the parameter weight λ2 is 0.1, the recognition recall rate of the model is only 74%. With the increase of the parameter weight, after the parameter weight λ2 increases to 0.3, the average recognition recall rate of the model can reach more than 83%.

FIGURE 9
www.frontiersin.org

Figure 9. Recognition effect of model recall under different parameter weights.

Two orientations are used to select the appropriate dimension:

One is insight orientation. This refers to the key dimensions that the researcher finds that affect user behavior based on experience and thinking accumulated in the target product and product field, as well as through qualitative research methods such as observation and interviews to understand users. This method is strong in-depth, and the dimension discovered may be more striking to the essence.

The other is statistical orientation. That is, through quantitative research methods such as questionnaires and log analysis, the key factors affecting user behavior are analyzed in quantitative data. These methods are more scientific and more convincing. But the discovered dimension may only be an intermediate factor, not necessarily an essential factor.

In practice, the statistical approach is more demanding on organizational resources, while the insight approach is more efficient. Therefore, when the resources are sufficient, the insight method and the statistical method are used in combination, and the most likely candidate dimensions are first screened out by the insight method and then verified by the statistical method. When resources are limited, it is recommended to use only the insight method. Of course, whether the insight method works depends on the researcher’s own skills, which not only refers to the mastery of the ideas and methods of user research itself but also refers to the depth of understanding of the target product and its field.

User Identification Test of Network Representation Learning Method

After the parameters suitable for the model are obtained, NRLM is used to identify the user’s types in the network learning process, and judges whether the model can be used in the actual learning process by identifying the user types. The test results are shown in Figure 10.

FIGURE 10
www.frontiersin.org

Figure 10. Distinguishing ability of network representation learning method for users’ types.

The data Figure 10 show that the red curve represents the correlation indicator of related users when using the network, and the blue curve shows the correlation indicator of unrelated users when using the network. The red curve is kept below 0.1. Although the lowest value of the red curve is lower than 0.1, it is maintained at 0.2–0.3. There is no overlap between the two curves, indicating that the model can effectively distinguish the two users, and can also accurately list the correlation indicator of users. Therefore, NRLM can be applied to students’ online education.

Discussion

This study proposes a network representation learning algorithm based on a structured deep network embedding model with the help of AE under DL. The concept and structure of the model are elaborated. Its performance is tested. Node labels and datasets are trained and found that with the increase in the number of training sets, the model’s classification accuracy for the three-node labels improved. The accuracy of Doc2vec is 0.7249 when the training set is 90%. When the training set is 90%, the accuracy of Deepwalk is 0.5764. When the training set is 90%, the accuracy of Node2vec is 0.5125. Therefore, the proposed network representation learning algorithm can fully utilize the existing label information to embed network nodes. The final generated vector has better classification performance. The application of the algorithm in practical work can also reflect the role of label nodes. In different scenarios, although there are still many types of influencing factors, the influence weight of each factor will change. In the process of user interaction with a specific product, the constraints formed by the product’s own characteristics may become the most important factor affecting user behavior, rather than some characteristics of the user itself, such as gender, age, or personality. Since the characteristics of a particular product are unique, people with different intrinsic characteristics may behave similarly to the same product. The proposed model can effectively distinguish two users and can also accurately list the relevant indicators of users. Therefore, NRLM is applied to students’ online education, which is an innovative teaching strategy and provides a theoretical basis for improving teaching methods. The disadvantage is that the research on the model is not very deep. In future research, factors such as proximity and structural features will be added to test the model’s performance, and different images, audio, and video will be used to study the recognition ability of the model and optimize it.

Conclusion

Online learning based on networks is proved to be an important way for modern people to live and work. Machine learning provides important support in online learning. As a new learning method, DL plays a huge role in NLP, analyzing media images and recommendation systems. The study aims to improve the learning effect of students by separating irrelevant users from the people who need learning resources and using AE in DL. Combined with the online learning network model, NRLM is proposed. After the establishment of the model, the parameters of the model are analyzed. The applicable parameter λ2 of the model is 0.3 by using the dataset and label nodes. It is also proved that the model can effectively distinguish relevant users and unrelated users and list the relevant indicators. It is believed that NRLM can guarantee users’ learning effect in students’ online education. Innovative teaching strategies provide a theoretical basis for improving teaching methods. The disadvantage is that the research on the model is not very deep. In future studies, factors such as proximity and structural characteristics will be added to test the performance of the model. Different images, audio, and video are used to study the recognition ability of the model and optimize it.

Data Availability Statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics Statement

The studies involving human participants were reviewed and approved by the Zhoukou Normal University Ethics Committee. The patients/participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author Contributions

All authors listed have made a substantial, direct, and intellectual contribution to the work, and approved it for publication.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Aamir, M., Nawi, N. M., Wahid, F., Zada, M. S. H., Rehman, M. Z., and Zulqarnain, M. (2021). Hybrid contractive auto-encoder with restricted boltzmann machine for multiclass classification. Arab. J. Sci. Eng. 46, 9237–9251. doi: 10.1007/s13369-021-05674-9

CrossRef Full Text | Google Scholar

Acosta-Cabronero, J., Milovic, C., Mattern, H., Tejos, C., Speck, O., and Callaghan, M. F. (2018). A robust multi-scale approach to quantitative susceptibility mapping. Neuroimage 183, 7–24. doi: 10.1016/j.neuroimage.2018.07.065

PubMed Abstract | CrossRef Full Text | Google Scholar

Adnan, M., and Anwar, K. (2020). Online learning amid the COVID-19 Pandemic: students’ perspectives. Online Submission J. Pedagog. Soc. Psychol. 2, 45–51. doi: 10.33902/JPSP.2020261309

CrossRef Full Text | Google Scholar

Belgacem, R., Bokhari, A., Djilali, S., and Kumar, S. (2021). Numerical inversion method for the laplace transform based on Boubaker polynomials operational matrix. Int. J. Model. Simul. Sci. Comput. 13:2550010. doi: 10.1142/S1793962322500106

CrossRef Full Text | Google Scholar

Bianco, S., Cadene, R., Celona, L., and Napoletano, P. (2018). Benchmark analysis of representative deep neural network architectures. IEEE Access 6, 64270–64277. doi: 10.1109/ACCESS.2018.2877890

CrossRef Full Text | Google Scholar

Chen, D., Nie, M., Yan, J., Meng, J., and Wang, D. (2022). Network representation learning algorithm based on community folding. J. Internet Technol. 23, 215–223. doi: 10.53106/160792642022032302020

CrossRef Full Text | Google Scholar

Chen, L., Fan, G., and Guo, K. (2021). A QoS prediction approach based on fusion of network representation learning and dynamic collaborative filtering for cloud service. Int. J. Intell. Internet Things Comput. 1, 184–199. doi: 10.1504/IJIITC.2021.115708

PubMed Abstract | CrossRef Full Text | Google Scholar

Fu, S., Wang, G., and Xu, J. (2021). hier2vec: interpretable multi-granular representation learning for hierarchy in social networks. Int. J. Mach. Learn. Cybern. 12, 2543–2557. doi: 10.1007/s13042-021-01338-0

CrossRef Full Text | Google Scholar

Gao, X., Wang, Y., Cheng, J., Xu, M., and Wang, M. (2021). Meta-learning based relation and representation learning networks for single-image deraining. Pattern Recogn. 120:108124. doi: 10.1016/j.patcog.2021.108124

CrossRef Full Text | Google Scholar

Guo, L., Yin, H., Chen, T., Zhang, X., and Zheng, K. (2021). Hierarchical hyperedge embedding-based representation learning for group recommendation. ACM Trans. Inf. Syst. (TOIS) 40, 1–27. doi: 10.1145/3457949

CrossRef Full Text | Google Scholar

Hoi, T. M., Guo, S., and Luo, S. (2020). When blockchain meets social-media: will the result benefit social media analytics for supply chain operations management? Transp. Res. Part E Logist. Transp. Rev. 135:101860. doi: 10.1016/j.tre.2020.101860

CrossRef Full Text | Google Scholar

Kent, C., and Rechavi, A. (2020). Deconstructing online social learning: network analysis of the creation, consumption and organization types of interactions. Int. J. Res. Method Educ. 43, 16–37. doi: 10.1080/1743727X.2018.1524867

CrossRef Full Text | Google Scholar

Lai, C. H., Lin, H. W., Lin, R. M., and Tho, P. D. (2019). Effect of peer interaction among online learning community on learning engagement and achievement. Int. J. Distance Educ. Technol. 17, 66–77. doi: 10.1186/s12913-016-1423-5

PubMed Abstract | CrossRef Full Text | Google Scholar

Liang, H., and Markchom, T. (2022). TNE: a general time-aware network representation learning framework for temporal applications. Knowl. Based Syst. 240:108050. doi: 10.1016/j.knosys.2021.108050

CrossRef Full Text | Google Scholar

Moffat, A. (2019). Huffman coding. ACM Comput. Surveys (CSUR) 52, 1–35. doi: 10.1145/3342555

CrossRef Full Text | Google Scholar

Nguyen, T. T., Pham, M. T., Nguyen, T. T., Huynh, T. T., Tong, V. V., Nguyen, Q. V. H., et al. (2021). Structural representation learning for network alignment with self-supervised anchor links. Expert Syst. Appl. 165:113857. doi: 10.1016/j.eswa.2020.113857

CrossRef Full Text | Google Scholar

Rezaei-Ravari, M., Eftekhari, M., and Saberi-Movahed, F. (2021). Regularizing extreme learning machine by dual locally linear embedding manifold learning for training multi-label neural network classifiers. Eng. Appl. Artif. Intell. 97:104062. doi: 10.1016/j.engappai.2020.104062

CrossRef Full Text | Google Scholar

Rogoza, R., żemojtel-Piotrowska, M., Kwiatkowska, M. M., and Kwiatkowska, K. (2018). The bright, the dark, and the blue face of narcissism: the spectrum of narcissism in its relations to the metatraits of personality, self-esteem, and the nomological network of shyness, loneliness, and empathy. Front. Psychol. 9:343. doi: 10.3389/fpsyg.2018.00343

PubMed Abstract | CrossRef Full Text | Google Scholar

Shang, Y., Gao, L., Zou, Q., and Yu, L. (2021). Prediction of drug-target interactions based on multilayer network representation learning. Neurocomputing 434, 80–89. doi: 10.1016/j.neucom.2020.12.068

CrossRef Full Text | Google Scholar

Shi, C., Hu, B., Zhao, W. Z., and Yu, P. S. (2018). Heterogeneous information network embedding for recommendation. IEEE Trans. Knowl. Data Eng. 31, 357–370. doi: 10.1109/TKDE.2018.2833443

CrossRef Full Text | Google Scholar

Sun, H., Jie, W., Loo, J., Chen, L., Wang, Z., Ma, S., et al. (2021). Network representation learning enhanced by partial community information that is found using game theory. Information 12:186. doi: 10.3390/info12050186

CrossRef Full Text | Google Scholar

Üsküplü, T., Terzi, F., and Kartal, H. (2020). Discovering activity patterns in the city by social media network data: a case study of Istanbul. Appl. Spat. Anal. Policy 13, 945–958. doi: 10.1007/s12061-020-09336-5

CrossRef Full Text | Google Scholar

Wang, H., Wu, W., Zheng, C., and Wu, Y. J. (2019a). Effect of narcissism, psychopathy, and machiavellianism on entrepreneurial intention—the mediating of entrepreneurial self-efficacy. Front. Psychol. 10:360.

Google Scholar

Wang, H., Li, S., Song, L., Cui, L., and Wang, P. (2019b). An enhanced intelligent diagnosis method based on multi-sensor image fusion via improved deep learning network. IEEE Trans. Instrum. Meas. 69, 2648–2657.

Google Scholar

Wang, Y., Yang, H., Yuan, X., Shardt, Y. A. W., Yang, C., and Gui, W. (2020). Deep learning for fault-relevant feature extraction and fault classification with stacked supervised auto-encoder. J. Process Control 92, 79–89. doi: 10.1016/j.jprocont.2020.05.015

CrossRef Full Text | Google Scholar

Wang, Z., and Cha, Y. J. (2021). Unsupervised deep learning approach using a deep auto-encoder with a one-class support vector machine to detect damage. Struct. Health Monit. 20, 406–425. doi: 10.1177/1475921720934051

CrossRef Full Text | Google Scholar

Wu, Y. J., and Chen, J. C. (2021). Stimulating innovation with an innovative curriculum: a curriculum design for a course on new product development. Int. J. Manag. Educ. 19:100561. doi: 10.1016/j.ijme.2021.100561

CrossRef Full Text | Google Scholar

Wu, Y., Wu, T., and Wu, Y. C. (2017). A decade of entrepreneurship education in the Asia Pacific for future directions in theory and practice. Manag. Decis. 55, 1333–1350. doi: 10.1108/MD-05-2017-0518

CrossRef Full Text | Google Scholar

Yan, L., Yuan, H., Xu, W., Hu, Z., and Liang, B. (2018). Generalized relative Jacobian matrix of space robot for dual-arm coordinated capture. J. Guid. Control Dyn. 41, 1202–1208. doi: 10.2514/1.G003237

CrossRef Full Text | Google Scholar

Zahara, Z., Rombe, E., Ngatimun, N., and Suharsono, J. (2021). The effect of e-service quality, consumer trust and social media marketing on intention to use online transportation services. Int. J. Data Netw. Sci. 5, 471–478. doi: 10.5267/j.ijdns.2021.4.001

CrossRef Full Text | Google Scholar

Zhang, D., Yin, J., Zhu, X., and Zhang, C. (2018). Network representation learning: a survey. IEEE Trans. Big Data 6, 3–28. doi: 10.1109/TBDATA.2018.2850013

CrossRef Full Text | Google Scholar

Keywords: online learning, deep learning, encoder, network embedding model, teaching strategies

Citation: Yuan L, Li H, Fu S and Zhang Z (2022) Learning Behavior Evaluation Model and Teaching Strategy Innovation by Social Media Network Following Learning Psychology. Front. Psychol. 13:843428. doi: 10.3389/fpsyg.2022.843428

Received: 25 December 2021; Accepted: 13 June 2022;
Published: 22 July 2022.

Edited by:

Chin-Feng Lai, National Cheng Kung University, Taiwan

Reviewed by:

Ruey-Shun Chen, National Chiao Tung University, Taiwan
Jia-Lang Xu, National Chung Hsing University, Taiwan

Copyright © 2022 Yuan, Li, Fu and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Zizai Zhang, zhangzizai@zjnu.edu.cn

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.