- 1School of Foreign Languages, Changshu Institute of Technology, Suzhou, China
- 2The Institute for Sustainable Development, Macau University of Science and Technology, Macau, China
Literary therapy theory provides a new field for contemporary literature research and is of great significance for maintaining the physical and mental health of modern people. The quantitative evaluation of psychotherapy effects in Japanese healing literature is a hot research topic at present. In this study, a text convolutional neural network (Text-CNN) was selected to extract psychological therapy features with different levels of granularity by using multiple convolutional kernels of different sizes. Bidirectional threshold regression neural network (BiGRU) can characterize the relationship between literature research and the psychotherapy effect. On the basis of the CNN-BilSTM model, a parallel hybrid network integrated with the attention mechanism was constructed to analyze the correlation between literature and psychotherapy. Through experimental verification, the model in this study further improves the accuracy of correlation classification and has strong adaptability.
Introduction
Ryan Idell, a famous contemporary American literary psychologist and critic, particularly analyzed the development of Western literature and psychology from the late nineteenth century to the early twentieth century. After the First World War, the “inward-turning” movement in western literature gradually erased the frontiers between literature and psychology, and the two tended to combine at the edge of the new century. In view of psychologists and psychotherapists, verbal narration and art related to literary narration can play therapeutic roles. Wang Meng, a writer, said, “The rich spiritual life of man [sic] depends not only on psychology but also on literature.” From the perspective of psychoanalysis and humanistic psychology, the psychological principles of literary therapy can be relied on Goh et al. (2018).
The correlation between literature and psychology can be quantitatively evaluated by the attention mechanism. The attention mechanism originated from the study of human vision, and its primary function is to pay attention to the information that is more critical to the current task. Dabek and Caban (2015) introduced the attention mechanism based on RNN when conducting image processing psychotherapy (Chorowski et al., 2015). Chorowski et al. (2015) used the attention-based neuromachine method for the first time to establish a mental state prediction model (Dabek and Caban, 2015). Huang and Sun (2017) used the attention mechanism to dynamically calculate the source language context solved the problem of capturing long-distance relevance in the source language sentences. Bai et al. (2018) proposed a model integrating double-level attention to improve the performance of emotion analysis. Ravi et al. (2020) put forward an interactive attention mechanism for capturing network faults in text classification key characteristics, focusing on the extraction and classification of key characteristics using the BiLSTM and the CNN, as well as the formation of global and local features: the use of attention mechanism failure and failure mode of the key information that is used to fuse the two form classification features. While literature shows that the study of attention mechanisms is mature in long-distance property correlation, there remains a gap in the correlation analysis between literature and psychology.
This study attempted to analyze the correlation between Japanese literature and psychotherapy using the diagnostic equation algorithm model. In the second part of this study, aiming at the CNN's inability to fully and comprehensively extract psychological therapy feature information, the Text-CNN was used to extract psychological therapy features with different levels of granularity by using multiple convolutional kernels of different sizes and to capture the relationship between sentences in Japanese literary works and readers' psychological emotions. BiGRU was used instead of the BiLSTM to capture the emotionally dependent relationship with a large time step distance more accurately and completely. This principle can explore deep semantic information in Japanese literature: simplify model structure and reduce training time. On the basis of the CNN-BilSTM model, we further optimized the attention mechanism by combining the Text-CNN and the BiGRU in parallel so that the model pays close attention to the features of important words, which cause great psychological fluctuations. Finally, a method of correlation analysis between literature and psychology based on the parallel hybrid network and attention mechanism is proposed. In the third part, this study has a detailed introduction of the application of a parallel hybrid network into the attention mechanism model of the whole process, and it completes the relevant parameter selection and comparative experiments. The experimental results verify the effectiveness of the improvement after the classification model of the parallel hybrid network into the attention mechanism, can effectively improve the precision of correlation analysis, and has a strong ability to adapt.
Research Status of Literary Therapy
The study of literary therapy has been discussed and applied in many fields and more dimensions.
Research at home and abroad is mainly carried out from three aspects: theoretical discussion, literary therapeutic criticism, and social practice.
Theoretical Discussion Level
Tang studied the therapeutic relationship between witchcraft and myth and witchcraft and literature, respectively (Shao and Cai, 2018). More attention is paid to the exploration of the narrative therapy principle. According to Tang W, narrative therapy externalizes problems and becomes a means of self-recognition to be noticed. Self-healing of creators also provides therapeutic possibilities for recipients, who can get cured by looking at themselves through the eyes of others in literary studies (Ghamisi et al., 2018). Gu added the isomorphism of the literary narrative structure and the “prior structure” of readers' minds (Chen et al., 2020). The dialogue and blending of individual and narrative discourses can help the individual mind achieve balance and surpass the power of generating and solving problems so as to realize the role of literary narrative in constructing the mind. Dai Y. H. also expanded the narrative theory of literary therapy ritual in his doctoral thesis. Fry, Puop, and Ye Shuxian have a common understanding of the common language of narrative and can use plain words to express surging psychological activities (Zhang et al., 2018). In the discussion of literature and literary therapy for the disease, Zhang M. L. thinks that literary therapy is a kind of medical means and original function (Wang et al., 2020). He talks not just about physical illness but also about the whole, holistic experience of illness, subdivided into physical symptoms, metaphorical illness, psychogenic illness, and doctor–patient interpretations of illness. Literature and therapy are separated from each other in the representation of discipline, but confluence occurs in the actual process of therapy. In daily life, through literature sharing or stimulating instinct, we can get recognition, feel better by venting, get self-affirmation, and achieve a balance between physical, psychological, and social functions. In addition, socio-cultural metaphorical therapy is widely used (Kim, 2017). Literary therapy takes responsibility for the process of correspondence between literature and society and also stimulates the courage and vitality of literature. In addition, Zeng H. W. applied literary therapy to the study of online literature, paying particular attention to the comprehensive effect of all media (Chen et al., 2018).
The Study of Literary Therapeutic Criticism
The metaphorical sociological function of literature therapy and the medical practice functions of physiology and psychology can be interpreted in the practice of literary study. Literature therapy not only has the largest number of papers applied to the practice of literary criticism, including a new perspective on the study of the works of Haruki Murakami and other relevant writers so that the recipients get a new understanding, but it also provides new ideas for the study of other writers and works (Cheung et al., 2017).
A Fusion of Psychology and Literature and Its Particular Therapeutic Features
At the level of Japanese literature, the research area has further expanded, involving the sacred ritual of new myth and ancient epic, language magic therapy, psychoanalysis, traumatic memory, ecological therapy, feminist language, literary language therapy, among others. These research results fully demonstrate the objective existence and relatively universal effectiveness of the therapeutic function of literature. It also shows that the literary therapy theory has attracted the attention of scholars in various fields of literature research in China. It is worth studying to distinguish the effect of different literary styles (Chen et al., 2019).
Application Level of Social Practice
The applied research and exploration of literary therapy have made progress. Studies on the application of literary therapy can be divided into the following categories: The first is the social and cultural construction promoted by the government, and the second is that the education department has a great role to play in the prevention and treatment of psychological diseases among students (Nanjun et al., 2019). Huang Mingfang believes that literary therapy also plays a good role in the prevention of psychological diseases caused by social employment and other pressures on college students. Third, literature therapy can be combined with network media culture (Liu et al., 2020). In particular, the whole media has outstanding leap-forward and integration, and the audience scope and treatment scale of literary therapy have been increased to unprecedented levels. Fourth, literary therapy is an important auxiliary reconstruction measure in various disaster events.
To sum up, all kinds of psychological awareness and ability are relatively weak. Literature therapy in mainland China is relatively lacking in sociological therapy, especially in the research direction of combining with computer deep learning theory, which makes it difficult to directly respond to the current social and cultural development situation.
Theory of Equation Diagnosis Algorithm
Attention Mechanism
A traditional machine translation mainly consists of an encoder-decoder architecture, which results in partial information loss and an inability to selectively focus on relevant input sequences as each output information is generated. The encoder-decoder architecture with an attention mechanism can introduce attention weight on the input sequence to give priority to the location set with relevant information to generate the next output information (Xu et al., 2018). By separating the attention mechanism from the encoder-decoder architecture, the essential idea of the attention mechanism can be understood more intuitively, as shown in Figure 1.
As shown in Figure 1, the Source is composed of a series of <Key, Value> data pairs. Given a certain element, Query, the correlation between Query and Keyi is calculated to obtain the corresponding weight coefficient and the weighted sum to obtain the Attention Value. Assuming the length of the Source is L, the essential idea of the attention mechanism can be expressed by Formula (1):
The specific calculation process of Attention Value can be abstracted into three stages, as shown in Figure 2.
As shown in Figure 2, Keyi is the keyword, Query is the sequence Query, F (Q, K) is the similarity function, Valuei is the weight, Si is the correlation calculation Value between Query and Keyi, ai is the weight coefficient, and Attention Value is the Attention Value finally calculated. In the first stage, different similarity functions and calculation mechanisms are used to calculate the correlation between Query and Keyi. Commonly used F (Q, K) methods include the dot product, matrix multiplication, cos similarity, series mode, and multilayer perceptron, as follows:
(1) Dot product
(2) Matrix multiplication
(3) COS similarity
(4) Series mode
(5) Multilayer perceptron
The second stage is normalization, highlighting the weight of important elements to obtain the corresponding weight coefficient ai:
In the third stage, the weighted sum yields the Attention Value:
The Parallel Hybrid Network Into the Attention Mechanism Model Structure
The Text-CNN (Cheng et al., 2019) is an improvement of CNN and a feedforward neural network. Text-CNN extracts emotional features with different levels of granularity and the relationships within and between sentences by using multiple convolution kernels of different sizes. The Bidirectional Gated Recurrent Unit (BiGRU) is mainly composed of GRUs in both positive and negative directions, and the output results are jointly determined by these two GRUs. Not only can BiGRU process the contextual semantic feature information with a large time step distance but the network model structure can also become relatively simple, training parameters are reduced, and the time cost is low. The attention mechanism gives different weights to words with different degrees of importance so that some words get more attention and improve the classification accuracy.
First, the comment text was vectorized by the GloVe method, and then the Text-CNN and the BiGRU were connected in parallel. Finally, an attention mechanism was introduced to build a new parallel hybrid network and integrate it into the attention mechanism model to solve the related problems of emotion classification, as shown in Figure 3.
The Word Vector Represents Text
The word vector text is mainly expressed by mapping the input comment text into the vector form (Dong et al., 2018). The GloVe method is used to map each word in each sentence into the corresponding word vector xi, and then a sentence matrix S is formed:
where and xiare the word vectors of the i word, and k is the dimension of the word vector. S ∈ Rn×k, and n is the number of words. In this study, the GloVe method is used to vectorize the text. Since the GloVe model includes two training methods, namely, global statistical information from matrix decomposition and the local window context of Word2vec. The word vector obtained contains more associated semantic and grammatical information and is superior to the Word2vec word vector method in overall performance.
Text—The CNN Layer
The sentence matrix S is taken as the input of the Text-CNN layer. By using multiple convolution kernels of different sizes, Text-CNN extracts emotional features with different levels of granularity and the relationship between sentences so as to obtain more comprehensive feature information (Titos et al., 2018).
At the convolution layer, r × k convolution is used to perform a convolution operation on the sentence matrix S to obtain local semantic information. After passing through the convolution layer, the local feature obtained by the i neuron is expressed as Ci, w is the convolution kernel, b is the bias term, xi:i+r−1 is a total r row vector from i to i + r-1 in the sentence matrix, and the calculation of ci is shown in formula (10):
The pooling layer reduces the dimension of the local semantic features obtained by the convolution operation to capture the optimal solution of the local features. In this chapter, the maximum pooling method is used to extract the maximum feature from the local feature set to replace the whole local feature, as shown in Formula (11):
In this chapter, multiple convolution kernels of sizes 3, 4, and 5 are used for the convolution operation, and the obtained characteristic information is input into the pooling layer to obtain three maximum pooling results. The three pooling results are spliced together, and the final output of Text-CNN represents hT as the input of the attention mechanism layer.
The BiGRU Layer
The Bi-lstm is composed of two lstms with opposite directions. It can not only obtain the above information but also consider the impact of the following information on the current word. If the time step distance is large, the BiLSTM cannot fully obtain the context information of the word. In this paper, the BiGRU is used to solve this problem; it consists of a one-way forward propagation GRU and a one-way backward propagation GRU, and the output results are jointly determined by the GRU in these two directions. The BiGRU can not only fully extract the above information but can also take into account the influence of the following information on the current word, capture the word dependence relationship with a large time step distance, and mine the deep semantic information in the text. Moreover, the network model structure becomes relatively simple, the training parameters are reduced, and the time cost is relatively low. The sentence matrix S is used as the input of the BiGRU layer to extract contextual semantic feature information with a large time step distance. The specific processing process of the BiGRU is shown in Formulas (13)–(15):
where is the output of GRU propagating forward and backward, respectively; yB is the output of BiGRU; W is the weight matrix; b is the bias vector; and σ is the sigmoid function.
Attention Mechanism Layer
The application of the attention mechanism in text sentiment analysis can calculate the probability weights of different words by their weight distribution so that some words can get more attention, highlight important words, and improve the accuracy of classification. The input of the attention mechanism layer in this chapter is the result of feature fusion extracted by the Text-CNN and the BiGRU, and the weight configuration of the attention distribution of input information to the emotion vector generated by the attention mechanism is calculated. The model structure of the attention mechanism is shown in Figure 4.
(1) The features extracted by the Text-CNN and the BiGRU are fused as the input of the attention mechanism layer, denoted as hA, as shown in Formula (16):
(2) We calculated the target attention weight , with Wu as the weight matrix and bu as the bias vector, as shown in Formula (17):
(3) Normalized by the softmax function, the weight coefficient is obtained by the probability of attention weight so as to highlight the weight of important words, as shown in Formula (18):
(4) Attention weight configuration mainly sum the fused feature hA and weight coefficient to obtain the text vector ct of the important information of each word in the text so that the attention weight generated by the model plays an important role, as shown in Formula (19):
Emotion Classification Layer
The emotion classification layer takes the text feature of the important information of each word obtained by the attention mechanism layer as input and obtains the final emotion classification result through the sigmoid function.
where pc is the probability of a certain class, Wc is the weight matrix, and bc is the bias vector.
Experimental Study
Experimental Data Set and Environment
In order to verify the performance of the parallel hybrid network-integrated attention mechanism model in the correlation analysis of the influence of Japanese literature and psychotherapy, the following two English data sets were mainly selected for relevant experiments, as shown in Table 1.
Setting and Optimization of Experimental Parameters
Experimental parameter settings directly affect the classification effect of the model. After each iteration, the set hyperparameters are adjusted according to the evaluation index of the experiment. After several iterations, the specific parameter settings are shown in Table 2.
(1) Word vector dimension selection
Select word vectors of 50, 100, 200, and 300 dimensions to explore the influence of word vectors of different dimensions on correlation analysis. Table 3 shows the specific experimental results.
By analyzing Table 3, it can be seen that, when the dimension of the word vector is 50, the accuracy, accuracy rate, recall rate, and F1 of the model on IMDB and SST-2 are the lowest values, indicating that the performance of the model is poor. With the increase of the word vector dimension, the overall performance of the model is improved gradually. When the dimension of the word vector is 200, the accuracy, recall rate, and F1 of the model on IMDB and SST-2 reach their highest values, which are 91.46, 90.45, and 91.36% and 89.48, 90.81, and 89.65%, respectively. When the word vector dimension continues to increase, all the evaluation indexes of the model begin to show a downward trend except for the accuracy. Therefore, in the experiment of this chapter, the dimension of the word vector is set as 200 on IMDB and SST-2.
(2) The selection of the node number of the BiGRU hidden layer
In the experiment of selecting the number of nodes in the BiGRU hidden layer, the number of nodes in the BiGRU hidden layer was set to 32, 64, 96, 128, 160, and 192, respectively, to explore the influence of different numbers of nodes on the results of correlation analysis. Table 4 lists the specific experimental results.
Table 4 shows that, when the number of nodes in the BiGRU hidden layer is set to 32, the performance of the model is relatively poor in IMDB and SST-2 because the correlation information cannot be completely extracted as the number of nodes is too small. As the number of nodes continues to increase, the evaluation indexes of the model generally show a trend of first increasing and then decreasing. When the number of nodes in the hidden layer is increased to 96, the accuracy, recall rate, and F1 of the model on the two datasets reach their best values. Therefore, the number of nodes in the BiGRU hidden layer is set to 96 on IMDB and SST-2 in this study.
(3) Selection of the Dropout value
The Dropout values were set to 0.1, 0.2, 0.3, 0.4, 0.5, 0.6, and 0.7 to explore the impact of different Dropout values on correlation analysis results. Table 5 lists the specific experimental results.
By analyzing Table 5, it can be seen that, when the Dropout value is 0.1, the accuracy, accuracy rate, recall rate, and F1 of the model on IMDB and SST-2 data sets are all the lowest, which may be due to the small number of randomly discarded neurons and too many parameters in the model. The Dropout values continue to increase, and the values of each evaluation index of the model tend to increase. When the Dropout value reaches 0.5, the model's four evaluation indexes on the two datasets reach optimal values. As we continued to increase the Dropout value, the model's learning ability became worse due to a large number of discarded neurons, and the model's performance began to gradually decline, so we set the Dropout value to 0.5 on both datasets.
(4) Batch_size value selection
In the batch sample size selection experiment, this section will set the batch_size values to 16, 32, 64, 128, and 256, respectively, to explore different batches, the influence of size value on the results of correlation analysis. Table 6 lists the specific experimental results.
As can be seen from Table 6, with the continuous increase of batch_size, the accuracy of the model generally increases first and then decreases. When batch_size is 32, the model not only achieves the highest accuracy, recall rate, and F1 value on the IMDB and SST-2 but also achieves a better convergence effect in shorter training time and smaller memory resource occupancy. Therefore, the batch_size value is set to 32 on both IMDB and SST-2.
(5) Selection of the optimizer
For the comparison experiment of optimizer selection, five commonly used optimizers were selected: SGD, Adagrad, Adadelta, RMSProp, and Adam, in order to explore the influence of different optimizers on correlation analysis results. Table 7 lists the specific experimental results.
As can be seen from the experimental results in Table 7, the Adam optimizer has the best experimental results, and the overall performance of the model is the best, mainly because Adam can make an adaptive adjustment to the learning rate parameters, and the selection of super-parameters is quite robust, so the optimizer selects Adam.
Following the experimental exploration and analysis of the above important parameters, it is concluded that the final hyperparameter settings of the parallel hybrid network integrating attention mechanism model proposed in this study are as follows: the word vector dimension is 200; the node number of the BiGRU hidden layer is 96; the Dropout value is 0.5; batch_size 32; and the optimizer selects Adam. Other parameters are the same as those shown in Table 2.
Experimental Results and Analysis
Contrast Experimental Settings
The correlation analysis method of the influence of Japanese literature and psychotherapy based on a parallel hybrid network and attention mechanism proposed in this chapter will be compared with the following commonly used models to verify the validity of the model presented in this study. In order to ensure the accuracy and comparability of the comparison experiment results, all parameter settings of the comparison experiment are consistent with those of the model experiment proposed in this study.
(1) CNN: A feedforward neural network with a deep structure and convolution computation in the deep learning model is used to extract text phrase features and semantic information.
(2) Text-CNN: An improvement on the CNN, which uses convolution kernels of different sizes to extract emotional features with different levels of granularity and the relationships within and between sentences.
(3) BiLSTM: It is composed of two LSTM in opposite directions, which takes into account the influence of psychological therapy information on current words while obtaining Japanese literature information.
(4) BiGRU: The BiGRU is mainly composed of GRUs in both positive and negative directions. It can not only process literature and psychological-characteristic information with a large time step distance but also has a relatively simple network model structure, reduced training parameters, and a low time cost.
(5) Text-CNN-BiGRU: In critical literature, first, the GloVe method was used to obtain the representation of literary features, and then the Text-CNN network was used to extract the characteristic information of psychotherapy with different levels of granularity, which was used as the input of the BiGRU network model to obtain the information of Rien literature and psychotherapy with a large time step distance through BiGRU. Finally, the sigmoid function is used to obtain the final correlation classification results.
(6) Text-CNN + BiGRU: In critical literature, the GloVe method was first used to obtain the representation of literary features, which were input into Text-CNN and BiGRU networks, respectively, to extract different feature information. Then, the obtained features were segmented and input into the psychotherapy classification layer to obtain the final correlation classification result through the sigmoid function.
(7) Text-CNN + BiGRU + Attention: The correlation analysis method of the parallel hybrid network integrating attention mechanism was proposed in this chapter. First, the GloVe method was used to map critical literature into word vectors and input them into the Text-CNN to extract psychological therapy features with different levels of granularity as well as internal and interrelationships between them, and the feature information of Japanese literature and psychotherapy was extracted from the BiGRU. Then, the features obtained by the Text-CNN and the BiGRU are combined and input into the attention mechanism to assign different weights so that the important words get more attention. Finally, the sigmoid function is used to obtain the final correlation category label.
Experimental Results and Analysis
Table 8 shows the experimental results of the model in this chapter and other comparison models on the same data set.
Table 8 shows seven experimental results. In this chapter, the accuracy, accuracy rate, and F1 values of the Text-CNN + BiGRU + Attention model in the IMDB data set are 91.46, 92.32, and 91.36%, respectively, which are superior to the other six models. The accuracy rate, recall rate, and F1 value of the SST-2 data set reached 89.48, 90.81, and 89.65%, respectively, which were higher than those of the other six models. From groups 1 and 2, it can be seen that the effect of the Text-CNN with multiple convolutional kernels of different sizes is higher than that of single convolutional kernels. From groups 2 and 3, it can be seen that the effect of extracting emotion feature information of different levels of granularity using only the Text-CNN is worse than that of extracting Japanese literature feature information by the BiLSM alone. From groups 3 and 4, it can be seen that the BiGRU with an improved BiLSM structure not only greatly reduces the training time but also improves the model accuracy, recall rate, and F1 to some extent, indicating that BiGRU is more suitable for dealing with correlation tasks than BiLSM. It can be seen from groups 5, 6, 2, 3, and 4 that any combination of the Text-CNN and BiGRU network models is better than a single Text-CNN, BiLSTM, and BiGRU models in the task of correlation analysis. The main reason is that Text-CNN convolution kernels of different sizes can be used to extract text correlation information of different levels of granularity. The BiGRU can extract the learning ability of serialized features, literature, and psychotherapy. Therefore, it has a positive effect on the reprocessing of features extracted by the Text-CNN in a model combination. The chain fusion in group 5 mainly comes from the chain fusion in the Text-CNN and the BiGRU. It can see from Table 5 and Group 6 above that the parallel fusion method proposed in this chapter is better than the chain method in dealing with correlation analysis tasks. The results show that the attentional mechanism can effectively improve the overall performance of the fusion model. The attention mechanism calculates the probability weight of different words through weight distribution, which is conducive to the model grasping essential features and quickly improving classification accuracy.
In this paper, seven commonly used models are used on the IMDB and SST-2 data sets to obtain the changes in accuracy, loss rate, and iteration times of validation sets on different data sets by means of parameter training, as shown in Figures 5, 6, respectively.
As can be seen from Figure 5, with the increase in iteration times, the accuracy of all models gradually increases. In the second iteration, the accuracy of all models on IMDB and SST-2 exceeds 86 and 75%, respectively. Compared with the other six curves, the curve of the model in this study has relatively stable changes and small fluctuations, and the accuracy of the verification set is in a higher position, especially after the seventh iteration, which reaches more than 93% on IMDB and more than 90% on SST-2. It is concluded that the model proposed in this chapter is more stable in the task of extracting the characteristic correlation between literary factors and psychotherapy.
Conclusion
This chapter proposes a correlation analysis method for Japanese literature and psychotherapy based on the parallel hybrid network integrating the attention mechanism. First, the concrete structure of the parallel hybrid network integrating the attention mechanism model is introduced, which mainly includes the word vector representation literature layer, the Text-CNN layer, the BiGRU layer, the attention mechanism layer, and the correlation classification layer. The network structure and specific operating principles of each layer are further introduced in detail. Then, we selected five groups of important parameters, namely word vector dimension, the node number of the BiGRU hidden layer, the Dropout value, batch_size value, and optimizer, to explore the influence of different parameter values on correlation classification in the IMDB and SST-2 data sets so as to select the experimental parameter values most suitable for the model in this study. Finally, the correlation analysis method between Japanese literature and psychotherapy based on a parallel hybrid network and attention mechanism proposed in this study is compared with several commonly used models on the IMDB and SST-2 data sets. By analyzing the accuracy, accuracy rate, recall rate, and F1 of different models, it is found that the model proposed in this study has the best comprehensive performance in correlation classification and then proves that the parallel hybrid network-integrated attention mechanism model proposed in this study is more suitable for the task of multifactor correlation analysis. The experimental results verifired that the improved parallel hybrid network classification model incorporating an attention mechanism can effectively improve the classification accuracy of the model and has strong application ability.
Data Availability Statement
Publicly available datasets were analyzed in this study. This data can be found here: https://library.stanford.edu/research-data-services.
Ethics Statement
The studies involving human participants were reviewed and approved by the Ethics Committee of Changshu Institute of Technology. Written informed consent was obtained from all participants for their participation in this study.
Author Contributions
JS developed the concept, prepared the figures, and wrote the manuscript. LJ read, revised, and approved the manuscript. Both authors contributed to the article and approved the proof for publication.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
We would like to thank all the reviewers.
References
Bai, J., Li, F., and Ji, D. H. (2018). Bi LSTM-CNN position detection model based on attention. Comp. Appl. Software 35, 266–274. doi: 10.3969/j.issn.1000-386x.2018.03.051
Chen, D., Lv, J., and Zhang, Y. (2018). Graph regularized restricted boltzmann machine. IEEE Trans. Neural Network. Learn. Syst. 29, 2651–2659. doi: 10.1109/TNNLS.2017.2692773
Chen, K., Yao, L., and Zhang, D. A. S. (2020). Attention model for human activity recognition. IEEE Trans. Neural Networks Learn. Syst. 31, 1747–1756. doi: 10.1109/TNNLS.2019.2927224
Chen, Z., Jiang, J., Zhou, C., Jiang, X., Fu, S., and Zhihua, C. (2019). Trilateral smooth filtering for hyperspectral image feature extraction. IEEE Geosci. Remote Sens. Lett. 16, 781–785. doi: 10.1109/LGRS.2018.2881704
Cheng, J., Park, J. H., Cao, J., and Qi, W. (2019). Hidden markov model-based nonfragile state estimation of switched neural network with probabilistic quantized outputs. IEEE Trans. Cybemet. 50, 1900–1909. doi: 10.1109/TCYB.2019.2909748
Cheung, R., Aue, A., and Lee, T. (2017). Consistent estimation for partition-wise regression and classification models. IEEE Trans. Signal Proc. 65, 3662–3674. doi: 10.1109/TSP.2017.2698407
Chorowski, J., Bahdanau, D., Serdyuk, D., Cho, K., and Bengio, Y. (2015). Attention based models for speech recognition. arXiv arXiv:1506.07503v1. doi: 10.48550/arXiv.1506.07503
Dabek, F., and Caban, J. J. (2015). “A Neural Network Based Model for Predicting Psychological Conditions,” in Brain Informatics and Health. BIH 2015. Lecture Notes in Computer Science, Vol. 9250, eds Y. Guo, K. Friston, F. Aldo, S. Hill, and H. Peng (Cham: Springer), 252–261. doi: 10.1007/978-3-319-23344-4_25
Dong, S., Wu, Z. G., Shi, P., Su, H., and Huang, T. (2018). Quantized control of markov jump nonlinear systems based on fuzzy hidden Markov model. IEEE Trans. Cybern. 49, 2420–2430. doi: 10.1109/TCYB.2018.2813279
Ghamisi, P., Rash, B., and Benediktsson, J. A. (2018). Multisensor composite kernels based on extreme learning machines. IEEE Geosci. Remote Sens. Lett. 16, 196–200. doi: 10.1109/LGRS.2018.2869888
Goh, S. K., Abbass, H. A., Tan, K. C., Al-Mamun, A., Thakor, N., Bezerianos, A, et al. (2018). Spatio-spectral representation learning for electroencephalographic gait-pattern classification. IEEE Trans. Neural Syst. Rehabil. Eng. 26, 1858–1867. doi: 10.1109/TNSRE.2018.2864119
Huang, W. M., and Sun, Y. Q. (2017). Sentiment analysis of chinese short texts based on maximum entropy. Comput. Eng. Design. 38, 138–143. doi: 10.16208/j.issn1000-7024.2017.01.026
Kim, L.-W. (2017). DeepX: deep learning accelerator for restricted boltzmann machine artificial neural networks. IEEE Trans. Neural Networks Learn. Syst. 29, 1441–1453. doi: 10.1109/TNNLS.2017.2665555
Liu, H., Lin, J., Xu, S., Bi, T., and Lao, Y. (2020). A resampling method based on filter designed by window function considering frequency aliasing. IEEE Trans. Circ. Syst. I Regular Papers 67, 5018–5030. doi: 10.1109/TCSI.2020.3016736
Nanjun, H., Paoletti, M., Haut, J. M., Fang, L., Li, S., Plaza, A., et al. (2019). Feature extraction with multiscale covariance maps for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 57, 755–769. doi: 10.1109/TGRS.2018.2860464
Ravi, A., Beni, N. H., Manuel, J., and Jiang, J. (2020). Comparing user-dependent and user-independent training of CNN for SSVEP BCI. J. Neural Eng. 17, 026028. doi: 10.1088/1741-2552/ab6a67
Shao, Z., and Cai, J. (2018). Remote sensing image fusion with deep convolutional neural network. IEEE J. Select. Topics Appl. Earth Observ. Remote Sens. 11, 1656–1669. doi: 10.1109/JSTARS.2018.2805923
Titos, M., Bueno, A., Garcia, L., and Benitez, C. (2018). A deep neural networks approach to automatic recognition systems for volcano-seismic events. IEEE J. Select. Topics Appl. Earth Observ. Remote Sens. 11, 1533–1544. doi: 10.1109/JSTARS.2018.2803198
Wang, T., Li, C., and Wu, C. (2020). A gait assessment framework for depression detection using kinect sensors. IEEE Sens. J. 21, 3260–3270. doi: 10.1109/JSEN.2020.3022374
Xu, H., Wang, J., Zhu, L., Huang, F., and Wu, W. (2018). Design of a dual-mode balun bandpass filter with high selectivity. IEEE Microwave Wireless Compon. Lett. 28, 22–24. doi: 10.1109/LMWC.2017.2773040
Keywords: equation diagnostic algorithms, attention mechanism, Japanese literature, psychotherapy, Text-CNN, BiGRU
Citation: Shen J and Jiang L (2022) Correlation Analysis Between Japanese Literature and Psychotherapy Based on Diagnostic Equation Algorithm. Front. Psychol. 13:906952. doi: 10.3389/fpsyg.2022.906952
Received: 29 March 2022; Accepted: 20 April 2022;
Published: 30 May 2022.
Edited by:
Maozhen Li, Brunel University London, United KingdomReviewed by:
Ahmad Fathan Hidayatullah, Islamic University of Indonesia, IndonesiaZhiming Shi, Huaqiao University, China
Lianhui Li, North Minzu University, China
Yan Cheng, Jiangxi Normal University, China
Copyright © 2022 Shen and Jiang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Leping Jiang, jiangleping2006@126.com