Skip to main content

METHODS article

Front. Med., 01 July 2024
Sec. Precision Medicine
This article is part of the Research Topic Artificial Intelligence in Imaging, Pathology, and Genetic Analysis of Brain Tumor in the Era of Precision Medicine View all articles

Classification of pain expression images in elderly with hip fractures based on improved ResNet50 network

Yang Shuang&#x;Yang Shuang1Gong Liangbo&#x;Gong Liangbo2Zhao Huiwen&#x;Zhao Huiwen3Liu JingLiu Jing4Chen XiaoyingChen Xiaoying5Shen SiyiShen Siyi1Zhu XiaoyaZhu Xiaoya4Luo Wen
Luo Wen3*
  • 1The 2nd Ward of Hip Joint Surgery, Tianjin Hospital, Tianjin, China
  • 2College of Information Technology and Engineering, Tianjin University of Technology and Education, Tianjin, China
  • 3The 2nd Ward of Joint Surgery, Tianjin Hospital, Tianjin, China
  • 4Traumatic Orthopedics Department, The 3rd Ward of Hip Joint Surgery, Tianjin Hospital, Tianjin, China
  • 5The 2nd Ward of Knee Trauma Department, Tianjin Hospital, Tianjin, China

The aim of this study is designed an improved ResNet 50 network to achieve automatic classification model for pain expressions by elderly patients with hip fractures. This study built a dataset by combining the advantages of deep learning in image recognition, using a hybrid of the Multi-Task Cascaded Convolutional Neural Networks (MTCNN). Based on ResNet50 network framework utilized transfer learning to implement model function. This study performed the hyperparameters by Bayesian optimization in the learning process. This study calculated intraclass correlation between visual analog scale scores provided by clinicians independently and those provided by pain expression evaluation assistant(PEEA). The automatic pain expression recognition model in elderly patients with hip fractures, which constructed using the algorithm. The accuracy achieved 99.6% on the training set, 98.7% on the validation set, and 98.2% on the test set. The substantial kappa coefficient of 0.683 confirmed the efficacy of PEEA in clinic. This study demonstrates that the improved ResNet50 network can be used to construct an automatic pain expression recognition model for elderly patients with hip fractures, which has higher accuracy.

1 Introduction

Pain is often accompanied by changes in behavior, including facial expressions that have received significant attention (1). Similar to expressions of emotion (2), facial expressions during pain play a critical role in communicating information about the experience (3). Hip fracture is a common orthopedic injury that exponentially increases in incidence with age (4). Intense preoperative pain after hip fractures is common and can lead to long-term effects, such as residual pain and poor joint function (5). Accurate detection and estimation of pain is essential for pain management, such as whether and at what level analgesics are administered to the patient. Sometimes the pain on the patient’s face is much worse than it really is (6), they are less accurately detected than the six basic emotions and are often mistaken for negative emotions characterized by similar facial muscle movements, such as disgust (7).

Facial expression of pain is one potential method for improving the detection and estimation of pain. It is the most salient form of pain behavior and communicates information about the sensory and affective components of the multidimensional experience of pain (8). Evidence suggests that the core expression of pain is characterized by four facial muscle movements, including brow-lowering, orbit tightening, levator-contraction, and to a lesser extent, eye closure (9). Facial expressions of pain are associated with different processes than self-report and provide complementary information that may be more valid in some circumstances, such as among individuals with cognitive or expressive impairments (10).

Face recognition is the process by which a computer determines an individual’s identification based on their facial features. Traditional face recognition technology first calculates the feature description factors describing each individual’s identity based on the distribution of each pixel in the face image (11). The field of face recognition has advanced significantly with the development of deep learning, which has increased the efficiency and precision of face recognition compared to traditional techniques. The primary benefit of deep learning is its ability to train on enormous amounts of data, gradually adapt to various scenarios, and discover the optimal features to represent the data (12). Convolutional neural networks (CNN) and their variant networks have excellent effects on image processing, laying the foundation for the application of deep learning in the recognition of pain expression images of elderly individuals with hip fractures (13).

This study selected the ResNet50 network in this study based on the performance of different networks in image recognition projects through literature review, especially compared with VGNet19 and DenseNet121, which has best training time and memory performance (14). ResNet has been known by winning the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) in 2015, which perfectly solves the problem of gradient vanishing in deep neural networks. To improve the detection, recognition, and classification results, the ResNet50 network model has been improved by MTCNN network and Bayesian Optimization, reducing manual facial feature extraction steps, to minimize dependencies for manual experience and expert knowledge.

To improve the efficiency and accuracy of pain assessment, we intend to construct a facial recognition model of pain expression for elderly patients with hip fractures through deep learning technology. To address the issues of determining network structure, excessive number of training runs, and excessive time for processing facial expression classification in convolutional neural networks, we will design an improved ResNet50 network based on Bayesian optimization, and use this algorithm to construct an automatic classification model of pain expressions in elderly patients with hip fractures.

2 Methods

2.1 Technical route

The first phase proposed an automatic image classification model for pain expression based on the ResNet50 network. This approach addressed the shortcomings of traditional machine learning in facial expression image recognition, the high complexity of manual feature extraction and the limitations of conventional deep learning, such as network degradation with increasing network depth. The model flow chart depicted in Figure 1. Face detection was achieved by video using Multi-Task Cascaded Convolutional Neural Networks (MTCNN). Detected pre-processed images were inputted into the ResNet50 network, which was trained and optimized for pain grade classification using Bayesian optimization. Once the requirements were met, voice playback and data recording functions has been implemented. The Pain Expression Evaluation Assistant (PEEA) is an automated software system that analyzes facial pain expressions for the detection and classification of features relevant to pain assessment.

Figure 1
www.frontiersin.org

Figure 1. Flow chart of pain expression recognition method.

2.2 Multi-task cascaded convolutional neural networks

Traditional facial detection techniques used Adaboost facial detection algorithm and Active Appearance Model (AAM). AAM added facial texture information to the shape information of facial images by using 66 facial feature markers to detect the global information of facial images. However, AAM required manual labeling of facial feature points, making the model construction process cumbersome and complex, which has poor generalization performance. The Adaboost facial detection algorithm combined a large number of weak classifiers with average classification ability to form a strong classifier (15). A weak classifier compared by different Haar Like features to select the classifier with better features. Adaboost weighted and combined these weak classifiers to obtain a strong classifier. The Adaboost algorithm obtained weak global features, which has poor generalization ability, especially for specific populations, leading to low recognition accuracy.

In engineering practice, the MTCNN algorithm is renowned for its high detection speed and accuracy. MTCNN employs a three-layer cascade architecture to accomplish face detection and key point localization in images (16). MTCNN comprised three networks: P-Net, R-Net and O-Net.

The three stages of MTCNN can be simply explained as follows: The first network layer: P-Net, which scaled the image at multiple levels and performed sliding detection on each image using a 12 × 12 sliding window with a step size of 2 for each scale. Small pictures can detected large faces, while large pictures can detected small faces. All detected face frames undergo NMS (Non-Maximum Suppression) to obtain the face frames, which were then converted to the original size, and the short side was filled to convert them into squares of 24 × 24. The second network layer: R-Net, which processed multiple 24 × 24 faces from the previous step using the CNN network to obtain more precise face frames, which then undergone NMS and ultimately re-sized to a 48 × 48 square. The third network layer: O-Net, which take multiple 48 × 48 faces from the previous step as input to obtain multiple more accurate boxes, five face position points and confidence scores. NMS is then applied to the face frames to obtain the final required face frame.

The input image of MTCNN has not limited by size. By reducing the size of the convolution kernel, the computational complexity and weight parameters have reduced, and by optimizing the activation function, the network performance has significantly improved. Therefore, this study choose the MTCNN facial detection algorithm to extract facial expressions.

2.3 ResNet 50 convolutional neural network

A transfer learning approach utilizing ResNet50 has been employed in this study. The method involved pre-training a convolutional neural network (CNN) on a large existing dataset, and then transferring the pre-trained CNN to a target dataset for fine-tuning. The proposed model has been initially pre-trained on the ImageNet dataset (17). All layers, except the Softmax layer, are initialized with the pre-trained model parameters, as opposed to traditional random initialization (18). The Softmax layer is added to process the dataset used in this study. This transfer learning method offered several advantages, including superior model generalization performance, significant depth, high accuracy and good convergence (19).

The ResNet 50 model introduced residual blocks in deep neural networks, perfectly solving the problem of “gradient vanishing.” The basic idea of this network was “Identity Mapping,” which has mean that the output of each layer was the same, and residual mapping has easier to optimize than the original mapping, reducing the computational burden of the network (18). In the residual block, assuming x was the input of the model, H(x) was the output of the residual network, and the output after convolution operation was F(x), then H(x) = F(x) + x. As long as F(x) = 0, which formed the identity mapping function H(x) = x mentioned earlier, transforming the problem into an easily fitting residual function F(x) = H(x)−x (Figure 2).

Figure 2
www.frontiersin.org

Figure 2. Residual block structure.

The ResNet50 network consisted of 49 convolutional layers and 1 fully connected layer, with a total of 5 convolutional operation stages. The image entered the first stage of convolution, which has been batch regularization, activation function, and max pooling operation. After that, the network entered the second stage of convolution. From the second stage to the fifth stage of convolution, residual blocks with dimensions has been added, and each residual block contained three convolutional layers. After all convolution operations, the final input has into the fully connected layer and the corresponding classification has output through the softmax layer, as shown in Figure 3.

Figure 3
www.frontiersin.org

Figure 3. ResNet50 structure.

2.4 Bayesian optimization

Bayesian optimization is an algorithm, which is used Bayesian probability to search for the optimal value of an objective function, which the probabilistic surrogate model and acquisition function were the key components (20). The most widely used probabilistic surrogate model was the Gaussian process, and the acquisition function was based on the posterior probability of the objective function. The goal of the Bayesian optimization algorithm was to minimize the total loss r, and which is achieved by selecting the evaluation point xi using the acquisition function, which is formulated as follows Equation (1):

xi+1=maxxXλxD1:iri=yyi    (1)

where X is the decision space, λ(x,D1:i) is the acquisition function, and y* is the optimal solution (21).

The implementation of Bayesian optimization followed the procedure outlined below (22):

Determining the maximum number of iterations N.

Using the acquisition function to obtain the evaluation point xi.

Evaluating the acquisition function yi at the evaluation point yi.

Updating the probabilistic agent model by integrating the data Dt.

If the previous number of iterations n is less than the maximum number of iterations N, output xi; otherwise, return and continue iterating, until output xi.

3 Experiment

3.1 Database construction

In order to construct the pain expression database, this study followed the existing methods and schemes. Firstly, the researchers recorded videos of elderly patients with hip fractures expressing different degrees of pain. Secondly, key frames capturing the required expressions were extracted from the videos. Experienced orthopedic specialist clinicians evaluated and classified these key frames. Thirdly, consistent images were selected and included in the database.

Patients with hip fractures attending the inpatient facility of a hospital’s hip joint trauma department participated in the video collection. The Hospital Research Ethics Committee approved the study protocol, and patients provided the informed consent before participation. The research objective, subjects’ rights and investigators’ obligations were explained to the respondents using uniform language. The privacy of the participants was ensured throughout the study.

For image or video detection, a clinician used a camera to capture the patient’s facial expressions. The camera was positioned 1–1.5 meters away from the patient’s face, and the video acquisition time was 20–25 s. During this period, the patient did not need to hide their state.

The inclusion criteria for the study were: (1) the elderly has been diagnosed hip fractures according to radio-logical diagnostic criteria (23); (2) patients were entirely awake who have understanding independently; (3) age ≥ 65; (4) patient agrees to participate in this study. The exclusion criteria were: (1) patient has neurological diseases, such as Alzheimer’s disease; (2) patient has deafness, aphasia and other symptoms that hindered communication.

The method involved sequentially recording videos of elderly individuals expressing different degrees of pain. The required key expression frames were extracted from these videos, and experienced clinicians evaluated and classified them. The acquired images were labeled according to the pain grade of mild, obvious, severe, and intense. To ensure the reliability of the database, images with high consistency in multiple evaluators’ scores have mainly selected and included in the database, which summarized in the following aspects: turning over one the bed, moving from flat car to bed, lower limb flexion and extension training, straight leg raising training and bedside standing.

To focus on facial expression recognition algorithm research and minimize the influence of other factors, the original images were being normalized. The method involved three steps: rotation correction, image cropping and scale normalization. The purpose was to correct for background interference and angle offsets caused by the shooting environment and changes in face pose. This ensured the face of the patient in the image was upright and both eyes were in a horizontal state. The method removed redundant background information as much as possible, retaining only the effective facial region containing expressions. The center points of the two eyes were manually marked, and the image was rotated using the axis connecting the center points of the eyes as the reference. The center points of the eyes were adjusted to the same horizontal line to eliminate angle deviation. Next the patient’s facial region was manually cropped from the corrected image. Researchers performed pain classification on facial expression images after data augmentation using the Prkachin & Solomon pain severity method. They discovered and confirmed that the four movements of eyebrow depression and convergence (AU4), eye socket tightening (AU6 and AU7), levator muscle contraction (AU9 and AU10), and eye closure (AU43), which can carry a lot information about pain. The pain grade can be evaluated by evaluating the severity of AU, which is the Prkacin & Solomon pain grade, the expressed as Equation (2):

PSPI=AU4+maxAU6,AU7+maxAU9,AU10+AU43    (2)

The total pain grade is the sum from AU4, the maximum intensity of AU6 and AU7, the maximum intensity of AU9 and AU10, and AU43 (taken as 0 or 1) (24). The PSPI pain measurement standard is the only measured method in image frames, which can be used to evaluate the degree of pain currently. The PSPI scoring system used the Likert 5-level scoring method (a-e), which the score range of each item has from 0 to 4. Higher score means the patient has stronger pain symptoms. The core facial movements of pain expression has been shown in Figure 4 (25).

Figure 4
www.frontiersin.org

Figure 4. The core facial movements of pain expression.

The score of Figure 4 as follow Equation (3):

PSPI=4e+max6d,7e+max9c,10d+43b=4+max32+max43+1=4+3+4+1=12    (3)

The evaluation team involved an anesthesiologist, an orthopedic surgeon and two clinical nurses, who assessed the collected facial pain expression images by PSPI. All staff members had over 10 years experience in clinical work. After evaluating and selecting from the video key frames, a database of facial pain expressions of elderly patients has been established in accordance with guidelines for pain management (1). The dataset consisted of 4,538 images, which included 2,247 images of mild pain (VAS: 1–3), 735 images of obvious pain (VAS: 4–6), 729 images of severe pain (VAS: 7–8) and 827 images of intense pain (VAS: 9–10).

The entire dataset has been randomizing, which was partitioned into three sets: the training set (60%), validation set (20%), and test set (20%). The training dataset underwent rotating, cropping and translating to increase the number of training samples, which improved the robustness and generalization performance of model.

In the field of computer vision, the model’s performance depended on not only the data and the quality of its structure but also the optimizer, loss function and data enhancement methods. The effectiveness of the model’s training strategy, such as the optimizer, data augmentation and regularization technique, which impacted its performance importantly and significantly. This study utilized data enhancement techniques (Table 1) to improve the accuracy of model training.

Table 1
www.frontiersin.org

Table 1. Comparison of model accuracy before and after data enhancement.

Furthermore, Bayesian optimization was employed to determine the optimal hyperparameters, including the learning rate with a maximum of 60 iterations, which was expected to enhance the acquisition function. Table 2 showed the decision space for the optimization process.

Table 2
www.frontiersin.org

Table 2. Super parameters and their ranges of improved ResNet50.

In this study, the classification and generalization ability of the model for recognizing pain expressions in elderly patients with hip fracture were assessed based on the prediction accuracy and model training time (26).The formula used for the prediction accuracy P1 as follow Equation (4):

p1=mTAnT×100%    (4)

Where nT is the number of test sets used to validate the model, and mTA is the number of accurately classified samples in the sample test set.

3.2 Pain grade prediction

Inputting the normalized image into the trained improved Resnet50 model, using the Softmax classifier to receive the feature matrix input by the fully connected layer, and outputting the probability value of each category corresponding to the input object, assuming that there are N input objects XiYii1k, the label for each object is yi12k, which is the number of the model output categoriesk2(20), 4 grade classification (1, 2, 3, 4),k is performed for the pain expression, and maximum value is 4. Input Xi, using the assumption function fθXito estimate the probability value of its corresponding class j: Py=j/Xi, and the function is Equation (5):

fθxi=Pyi=1|xi;θPyi=2|xi;θ···Pyi=k|xi;θ=1j1keθjTxieθ1Txieθ2Txi···eθkTxi    (5)

The loss function for the Softmax classifier is (19) Equation (6):

Jxyθ=1Ni1Nj1k1yi=jlog2eθjTxij1keθjTxi    (6)

The pain grade of the facial expression is determined by taking the label category with the highest probability of the Softmax output. To avoid false detections and improve system stability, the detected label must reach a stable and continuous number of frames before it can be used as the output result of the pain grade determination. The collected video data is organized into a folder and transmitted to a dedicated computer, which is carried out the pain expression classification automatically in video detection mode. The pain grade report has been generated with the longest duration of pain expression in the video (Figures 5, 6).

Figure 5
www.frontiersin.org

Figure 5. Patients’ video pain assessment result output.

Figure 6
www.frontiersin.org

Figure 6. Patients’ pain assessment record sheet.

3.3 Measure

The study involved 15 clinicians, each with more than 5 years of clinical pain assessment experience, who underwent a training session. The session covered the structure of the Pain Evaluation and Entitlement Act (PEEA) report and included three videos to demonstrate the process. The trainer, who was familiar with the outputs of PEEA, provided guidance on where to find relevant information in the graphical outputs, but did not interpret any videos or images to allow the clinicians to utilize their medical expertise.

During the session, the clinicians were instructed to rate the pain grade based on their visual inspection of the patients’ condition solely. To ensure accurate assessment and avoid reader fatigue, the clinicians were allowed to take an unlimited amount of time, until completing the assessments. And then the researchers assessed pain grades of the same videos by the PEEA system.

Agreement rates between clinicians and PEEA were assessed by intraclass correlation (ICC), which was assuming random effects for 15 clinicians. 95% confidence intervals was calculated according to the original derivations by Shrout and Fleiss (27). Standard errors of the mean for ICCs were estimated by resampling the observations with replacement (bootstrap) 1,000 times. The two-way of the same raters for all subjects has been selected in this study.

4 Results

4.1 PEEA performance

Following Bayesian hyperparameter optimization, the learning rate, momentum, and L2 regularization were been determined, which were 0.1679, 0.8437, and 2.456 × 10–5, respectively. The optimizer utilized the SGDM optimization algorithm, which updated model parameters using a momentum factor while selecting training data samples. Comparing to the RMSProp and Adam optimizers, this method has better training speed and prediction accuracy (Figure 7).

Figure 7
www.frontiersin.org

Figure 7. Minimum target value and number of function calculations.

The trained model has been tested in a clinical setting, and the results of the statistical data analysis have been presented in Table 3. The experimental analysis demonstrated that improved network has higher accuracy in both the training and validation sets, with better performance than the ResNet50 network for classification and recognition of pain expressions in elderly patients with hip fractures, which has higher training efficiency.

Table 3
www.frontiersin.org

Table 3. Comparison of pain expression classification models in elderly hip fracture patients.

To assess the effectiveness of the MTCNN model for face detection in this study, we evaluated the detection precision and speed in face testing by 32 patients. As shown in Figure 8, the detection precision was 0.97, and the detection speed was 0.023 s/image when using a processor of Core (TM) i5-12400, indicating a high detection speed and precision for the proposed face detection algorithm. The model converged at 1063 iterations, and the result of final validation set was up to 98.7% approximately.

Figure 8
www.frontiersin.org

Figure 8. Schematic diagram of detection precision of the MTCNN model in pain expression recognition method for elderly individuals with hip fractures.

4.2 Validation

For the validation analysis, we selected 300 pain videos of elderly patients with hip fractures, with 28 excluded due to dim lighting, resulting in 272 expression videos for 182 patients with hip features for analysis. The pain assessment and video capture are both conducted 24 and 48 h after the patient’s injury. Some participants collected two videos. The demographic and clinical characteristics of the enrolled patients are shown in Table 4. The pain scores between clinicians and PEEA were been compared. And the results demonstrated that PEEA has good reliability in pain assessment, with an average intraclass correlation measure of 0.9257 (Tables 5, 6) and a substantial kappa coefficient of 0.683 (Table 7; Figure 9).

Table 4
www.frontiersin.org

Table 4. The participants characteristics (N = 182).

Table 5
www.frontiersin.org

Table 5. The pain grade of videos between clinician and PEEA.

Table 6
www.frontiersin.org

Table 6. Intraclass correlation coefficient between clinician and PEEA.

Table 7
www.frontiersin.org

Table 7. Symmetry measure of the pain videos.

Figure 9
www.frontiersin.org

Figure 9. Individual agreement between clinician independently and PEEA for measure the pain videos.

5 Discussion

The effects of data enhancement on model performance were evaluated by training the original and enhanced datasets under the same parameters and network structure. The results showed that the testing accuracy of the enhanced dataset was higher than that of the unexpanded original dataset, indicating that data enhancement can increase data diversity and prevent over-fitting during training. The increased training data also led to more detailed feature types, resulting in more stable training accuracy and reduced fluctuations, which improved the model’s robustness effectively.

The method for recognizing pain expressions in elderly patients with hip fractures involved acquiring a face image, using MTCNN for face detection and pre-processing, building and training a improved Resnet50 network model, and inputting images for pain grade prediction. By integrating MTCNN and Resnet50 network and performing transfer learning on a self-built dataset, the proposed model achieved accurate classification of pain expressions in elderly patients with hip fractures.

In this study, we constructed a deep learning network model for recognizing and classifying pain expressions in elderly patients with hip fractures. The residual module in the model addressed the problem of network degradation effectively, ensuring that performance did not decline with deepened network depth. To tackle the problem of small publicly labeled image datasets on pain expression, transfer learning has been used to avoid over-fitting.

The results showed that PEEA had high accuracy and efficiency, indicating that they tend to decrease nondeterminacy. This is important because nondeterminacy can lead to unnecessary interventions or examinations, costing time, money, and causing discomfort and anxiety for patients. The technology could allow clinicians to recognize pain grades more quickly. However, because our study only included 15 clinicians, we used the type of ICC that considers the clinician as a random effect to quantify agreement rates. This allows us to generalize the results to a broader population. This study strictly adhered to ethical principles during implementation, fulfilled the obligation of disclosure before the experiment, collected videos with the patient’s consent. This study mainly evaluated the dynamic pain grade of patients. The pain situation in the resting state was not be included in this study. Functional exercise and pain assessment are performed on patients in an independent space relatively to reduce external interference factors and protect patient privacy. At the same time, the database of the pain automatic classification systems stored in the specific computer folder. Protective measures has been set up to prevent arbitrary copying and leakage of patient information.

Artificial intelligence has the potential to revolutionize nursing and medicine. Our study highlighted that these software systems are meant to support and enhance clinicians’ performance in clinical practice rather than replace them. Automated assessment systems, such as PEEA, which could be used to assess and grade a large number of pain videos or images quickly, improving the reliability of measurements by decreasing interobserver variability.

The limitations of this study are that: (1) The results may lack effective persuasion because the number of samples is limited, especially in model validation. (2) The evidence of criterion-related validity was limited and sensitivity has not been measured. Although the positioning accuracy of model must be sufficient, but this model has targeted clearly, so the number of samples can be reduced. (3) Pain is a subjective feeling that is influenced by many factors, such as personality, social environment, previous pain experiences, etc. At present this study cannot exclude all influencing factors to accurately evaluate pain grades. In future research, specific character models can be established to improve pain expression recognition results. (4) This study is based on facial expression recognition of pain grades in elderly patients, with limited information. In future research, we should pay more attention to the means of multi information fusion to further improve the accuracy of pain expression recognition.

6 Conclusion

Traditional convolutional neural network (CNN) has been used for recognizing and classifying pain expression in elderly patients with hip fractures suffers from long training times and poor accuracy. In this study, we improved the hyperparameters of the ResNet50 network using Bayesian optimization. The results confirmed that this network can construct an automatic pain expression recognition and classification model for elderly patients with hip fractures. Comparing experimental analysis, the study led to the following conclusions: The automatic pain expression recognition model for elderly patients with hip fractures has been constructed based on the algorithm of this study which does not require manual extraction of facial features, reducing reliance on experience and expertise. The improved network has higher accuracy, fewer training runs and shorter training time.

In conclusion, our study suggests that computer-assisted detection systems, such as PEEA, which can improve both the speed and efficiency of pain grade assessments. This technology can be extended to the clinical evaluation of elderly people aged 65 and above, especially those with traumatic fractures. However, due to limitations in database data, the effect of this method in young patients with fractures still needs further research.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by Tianjin Hospital Medical Ethics Committee, Tianjin Hospital. The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individuals for the publication of any potentially identifiable images or data included in this article.

Author contributions

YS: Formal analysis, Supervision, Writing – review & editing. GL: Methodology, Validation, Writing – original draft. ZH: Funding acquisition, Investigation, Writing – original draft. LJ: Data curation, Formal analysis, Investigation, Project administration, Writing – review & editing. CX: Data curation, Formal analysis, Investigation, Writing – original draft. SS: Investigation, Writing – original draft. ZX: Investigation, Writing – original draft. LW: Methodology, Supervision, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported in part by China Tianjin Health Information Society under Grant (no. TJHIA-2023-013, PI:Huiwen Zhao).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Chou, R, Gordon, DB, LOA, d, Rosenberg, JM, Bickler, S, Brennan, T, et al. Management of postoperative pain: a clinical practice guideline from the American pain society, the American Society of Regional Anesthesia and Pain Medicine, and the American Society of Anesthesiologists' committee on regional anesthesia, executive committee, and administrative council. J Pain. (2016) 17:131–57. doi: 10.1016/j.jpain.2015.12.008

PubMed Abstract | Crossref Full Text | Google Scholar

2. Rahul, M, Shukla, R, Goyal, P K, Siddiqui, Z. A., and Yadav, V. Gabor filter and ICA-based facial expression recognition using two-layered hidden Markov model//advances in computational intelligence and communication technology. Springer, Singapore, (2021): 511–518.

Google Scholar

3. Haque, MA, Bautista, RB, Noroozi, F, Kulkarni, K, Laursen, CB, Irani, R, et al. Deep multimodal pain recognition: a database and comparison of spatio-temporal visual modalities, In: 2018 13th IEEE international conference on Automatic Face & Gesture Recognition (FG 2018), IEEE, (2018):250–257.

Google Scholar

4. Samara, A, Galway, L, Bond, R, and Wang, H. Affective state detection via facial expression analysis within a human–computer interaction context. J Ambient Intell Humaniz Comput. (2019) 10:2175–84. doi: 10.1007/s12652-017-0636-8

Crossref Full Text | Google Scholar

5. Tan, M, Pang, R, and Le, Q V. Efficientdet: scalable and efficient object detection// Proceedings of the IEEE/CVF conference on computer vision and pattern recognition. (2020): 10781–10790.

Google Scholar

6. Lim, H, Kim, B, Noh, G-J, and Yoo, SK. A deep neural network-based pain classififier using a photoplethysmography signal. Sensors. (2019) 19:384. doi: 10.3390/s19020384

PubMed Abstract | Crossref Full Text | Google Scholar

7. Gan, TJ. Poorly controlled postoperative pain: prevalence, consequences, and prevention. J Pain Res. (2017) 10:2287–98. doi: 10.2147/JPR.S144066

Crossref Full Text | Google Scholar

8. Masi, I, Wu, Y, Hassner, T, and Natarajan, P. Deep Face Recognition: A Survey. 31st SIBGRAPI Conference on Graphics, Patterns and Images, 29 October-1 November 2018. Parana, Brazil: IEEE, (2018) 471–8. doi: 10.1109/SIBGRAPI.2018.00067

Crossref Full Text | Google Scholar

9. Kächele, M, Thiam, P, Amirian, M, Schwenker, F, and Palm, G. Methods for person-centered continuous pain intensity assessment from bio-physiological channels. IEEE J Select Top Signal Process. (2016) 10:854–64. doi: 10.1109/JSTSP.2016.2535962

Crossref Full Text | Google Scholar

10. Liu, X, Cheng, X, and Lee, K. Ga-svm-based facial emotion recognition using facial geometric features. IEEE Sensors J. (2020) 21:11532–42. doi: 10.1109/JSEN.2020.3028075

Crossref Full Text | Google Scholar

11. Tiwari, P, Rathod, H, Thakkar, S, and Darji, AD. Multimodal emotion recognition using SDA-LDA algorithm in video clips. J Ambient Intell Humaniz Comput. (2021) 14:6585–602. doi: 10.1007/s12652-021-03529-7

Crossref Full Text | Google Scholar

12. Seok, HS, Choi, B, Noh, G, and Shin, H. Postoperative pain assessment model based on pulse contour characteristics analysis. IEEE J Biomed Health Inform. (2019) 23:2317–24. doi: 10.1109/JBHI.2018.2890482

PubMed Abstract | Crossref Full Text | Google Scholar

13. Henriques, JF, Caseiro, R, Martins, P, and Batista, J. High-speed tracking with kernelized correlation filters. IEEE Trans Pattern Anal Mach Intell. (2014) 37:583–96. doi: 10.1109/TPAMI.2014.2345390

Crossref Full Text | Google Scholar

14. Mandal, B, Okeukwu, A, and Theis, Y. Masked face recognition using resnet-50. (Computer Vision and Pattern Recognition) (2021).

Google Scholar

15. Priyadharsini, GR, and Krishnaveni, K. An analysis of Adaboost algorithm for face detection. Indian J Sci Technol. (2016) 9:628–35. doi: 10.17485/ijst/2016/v9i19/93855

Crossref Full Text | Google Scholar

16. Zhang, K, Zhang, Z, Li, Z, and Qiao, Y. Joint face detection and alignment using multitask cascaded convolutional networks. IEEE Signal Process Letters. (2016) 23:1499–503. doi: 10.1109/LSP.2016.2603342

Crossref Full Text | Google Scholar

17. Xin, M, and Wang, Y. Research on image classification model based on deep convolution neural network. Eurasip J Image Video Process. (2019) 2019:1–11. doi: 10.1186/s13640-019-0417-8

Crossref Full Text | Google Scholar

18. He, KM, Zhang, XY, and Ren, SQ, Deep residual learning for image recognition// proceedings of the IEEE conference on computer vision and pattern recognition. Las Vagas: IEEE, (2016):770–778.

Google Scholar

19. Shin, HC, Roth, HR, Gao, M, Lu, L, Xu, Z, Nogues, I, et al. Deep convolutional neural networks for computer-aided detection: CNN architectures, dataset characteristics and transfer learning. IEEE Trans Med Imaging. (2016) 35:1285–98. doi: 10.1109/TMI.2016.2528162

PubMed Abstract | Crossref Full Text | Google Scholar

20. Cui, JX, and Yang, B. Survey on Bayesian optimization methodology and application. J Software. (2018) 29:3068–90. doi: 10.13328/j.cnki.jos.005607

Crossref Full Text | Google Scholar

21. Mahmood, M, Jalal, A, and Evans, H A. Facial expression recognition in image sequences using 1D transform and gabor wavelet transform// 2018 international conference on Applied and Engineering Mathematics (ICAEM). IEEE, (2018): 1–6.

Google Scholar

22. Arora, M, and Kumar, M. Auto FER: PCA and PSO based automatic facial emotion recognition. Multimed Tools Appl. (2021) 80:3039–49. doi: 10.1007/s11042-020-09726-4

Crossref Full Text | Google Scholar

23. Reiman, MP, and Thorborg, K. Clinical examination and physical assessment of hip joint-related pain in athletes. Int J Sports Phys Ther. (2014) 9:737–55.

PubMed Abstract | Google Scholar

24. Prkachin, KM, and Solomon, PE. The structure,reliability and validity of pain expression:evidence from patients with shoulder pain. Pain. (2008) 139:267–74. doi: 10.1016/j.pain.2008.04.010

PubMed Abstract | Crossref Full Text | Google Scholar

25. Hammal, Z, and Cohn, JF. Automatic detection of pain intensity. The 14th ACM International Conference on Multimodal Interaction (ICMI),Santa Monica, USA: ACM, (2012):47–52.

Google Scholar

26. Anagnostopoulos, CN, Iliou, T, and Giannoukos, I. Features and classifiers for emotion recognition from speech: a survey from 2000 to 2011. Artif Intell Rev. (2015) 43:155–77. doi: 10.1007/s10462-012-9368-5

Crossref Full Text | Google Scholar

27. Xiao, B, Wu, H, and Wei, Y. Simple baselines for human pose estimation and tracking// Proceedings of the European conference on computer vision (ECCV). (2018): 466–481.

Google Scholar

Keywords: expression recognition, Resnet50 network, MTCNN face detection, Bayesian optimization, pain assessment

Citation: Shuang Y, Liangbo G, Huiwen Z, Jing L, Xiaoying C, Siyi S, Xiaoya Z and Wen L (2024) Classification of pain expression images in elderly with hip fractures based on improved ResNet50 network. Front. Med. 11:1421800. doi: 10.3389/fmed.2024.1421800

Received: 23 April 2024; Accepted: 18 June 2024;
Published: 01 July 2024.

Edited by:

Zekuan Yu, Fudan University, China

Reviewed by:

Luigi Vernaglione, Azienda Sanitaria Locale di Brindisi, Italy
Vlatka Sotošek, University of Rijeka, Croatia

Copyright © 2024 Shuang, Liangbo, Huiwen, Jing, Xiaoying, Siyi, Xiaoya and Wen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Luo Wen, wl1984@yahoo.com

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.