Skip to main content

ORIGINAL RESEARCH article

Front. Neurosci., 11 July 2023
Sec. Neuroprosthetics
This article is part of the Research Topic Smart Wearable Devices in Healthcare—Methodologies, Applications, and Algorithms View all 11 articles

Improved spatial–temporal graph convolutional networks for upper limb rehabilitation assessment based on precise posture measurement

Jing Bai,
Jing Bai1,2*Zhixian WangZhixian Wang3Xuanming Lu,Xuanming Lu1,2Xiulan WenXiulan Wen3
  • 1Industrial Technology Research Institute of Intelligent Equipment, Nanjing Institute of Technology, Nanjing, China
  • 2Jiangsu Provincial Engineering Laboratory of Intelligent Manufacturing Equipment, Nanjing, China
  • 3‘School of Automation, Nanjing Institute of Technology, Nanjing, China

After regular rehabilitation training, paralysis sequelae can be significantly reduced in patients with limb movement disorders caused by stroke. Rehabilitation assessment is the basis for the formulation of rehabilitation training programs and the objective standard for evaluating the effectiveness of training. However, the quantitative rehabilitation assessment is still in the experimental stage and has not been put into clinical practice. In this work, we propose improved spatial-temporal graph convolutional networks based on precise posture measurement for upper limb rehabilitation assessment. Two Azure Kinect are used to enlarge the angle range of the visual field. The rigid body model of the upper limb with multiple degrees of freedom is established. And the inverse kinematics is optimized based on the hybrid particle swarm optimization algorithm. The self-attention mechanism map is calculated to analyze the role of each upper limb joint in rehabilitation assessment, to improve the spatial-temporal graph convolution neural network model. Long short-term memory is built to explore the sequence dependence in spatial-temporal feature vectors. An exercise protocol for detecting the distal reachable workspace and proximal self-care ability of the upper limb is designed, and a virtual environment is built. The experimental results indicate that the proposed posture measurement method can reduce position jumps caused by occlusion, improve measurement accuracy and stability, and increase Signal Noise Ratio. By comparing with other models, our rehabilitation assessment model achieved the lowest mean absolute deviation, root mean square error, and mean absolute percentage error. The proposed method can effectively quantitatively evaluate the upper limb motor function of stroke patients.

1. Introduction

Stroke is the second leading cause of death in the world, and its incidence rate is on the rise in recent years (Paul and Candelario-Jalil, 2021). The disability rate of this disease is high, and more than 50% of survivors will leave varying degrees of disability, which seriously affecting the daily life of patients, causing great pain to themselves, and adding a heavy economic burden to families and society. The World Stroke Organization (WSO) estimates that the global cost of stroke is over $721 billion (Feigin et al., 2022). Therefore, there is a great demand for rehabilitation training and assessment in patients with motor dysfunction.

Rehabilitation assessment is not only the basis of making a rehabilitation treatment plan but also the objective standard to observe its treatment effect. It plays an important role in rehabilitation treatment, evaluation of treatment effect, and prediction of functional recovery (Liao et al., 2020). At present, the commonly used assessment method is carried out by experienced rehabilitation physicians using the evaluation scale. The popular clinical evaluation tools are the Brunnstrom evaluation method, Fugl-Meyer Assessment (FMA), Barthel index, and so on. However, these methods are subjective assessment methods of rehabilitation physicians, with inconsistent judgment standards and inability to distinguish between compensation and true recovery (Li et al., 2022; Rahman et al., 2023). The main defect of the subjective scale is that it has a ‘ceiling effect’ on patients with mild injury. In addition, completing assessment tests is time consuming, complex, and labor intensive.

Scholars have carried out related research on rehabilitation assessment to solve the problems above. It is proposed to use an inertial measurement unit, accelerometer, VICON, infrared camera, and so on to capture human posture data (Fang et al., 2019; Hussain et al., 2019; Ai et al., 2021). The manual features are extracted from human posture data to represent human motion (Cai et al., 2019; Hamaguchi et al., 2020). Mahalanobis distance, and dynamic time warping (DTW) algorithm is used to quantify the correctness of rehabilitation exercise, support vector machine, logistic regression, and neural network are also used to grade the rehabilitation assessment (Houmanfar et al., 2016; Fang et al., 2019; Li et al., 2021). These methods rely on the results of sub-problems such as preprocessing and feature extraction, but the optimal solution of the sub-problem is not necessarily the global optimal solution and lacks end-to-end learning intuition.

Because wearable measuring equipment is very cumbersome to use, the acceptance of patients is not high, markers may be moved due to soft tissue effects, and motion capture systems such as VICON are too expensive. As an unmarked tool, Kinect is increasingly being applied to human posture tracking (Zelai and Begonya, 2016; Bawa et al., 2021). Kinect-based joint data contains a variety of information, including spatial information between joint nodes and their adjacent nodes, as well as time-domain information between frames. It has been widely used in motion recognition (Wang et al., 2020), gesture recognition (Ma et al., 2021), somatosensory interaction (Qiao et al., 2022), and also has applications in rehabilitation assessment (Agami et al., 2022) proposed a method for generating accurate skeleton data based on the offline fusion of a Kinect 3D video sensor and an electronic goniometer. This method is difficult to measure the patient’s joint angles with the electronic goniometer (Lee et al., 2018) used Kinect v2 and force sensing resistor sensors based on Fugl-Meyer assessment for evaluating upper extremity motor function (Bai and Song, 2019) conducted a preliminary rehabilitation assessment using the first-generation Kinect to measure the joint data of stroke patients, ignoring the drawbacks of a single camera.

However, there is an issue of inaccurate joint position recognition using a single Kinect. This type of erroneous recognition is prone to occur in situations of self-occlusion, when the subject is not facing the camera, or when moving at high speeds (Han et al., 2016; Wang et al., 2016). This is because although the connections of the bones obtained during recognition are biologically consistent, the length of the limbs and the limitations of the joints are not limited, resulting in unrealistic and distorted movements. Adding additional manual measurements or wearable sensors can be time-consuming and reduce patient comfort. The accuracy of tracking data for human motion posture seriously affects the correctness of rehabilitation assessment results, therefore, the accuracy of human motion tracking should be improved. How to improve the accuracy of patient pose recognition using only visual sensors is a complex problem.

An approach to improve the accuracy of human motion tracking is to combine a rigid body model with the depth camera (Matthew et al., 2019) used this approach in the sit-to-stand movement and the upper limb motion. Due to the lack of hand modeling and occlusion, the estimation of joint position is incorrect. In the Proximal Function test, the system error is introduced, and the accuracy of the overall pose estimate is reduced (Matthew et al., 2020). Using one Kinect for rehabilitation assessment (Liu et al., 2016), the body information is particularly prone to occlusion, in some specific evaluation movements such as touching the back of the head, touching the lumbar vertebrae, and so on. The occlusion problem should be solved in order to improve the accuracy of rehabilitation assessment. So, in our work, we have added an Azure Kinect and optimized the rigid body model.

Neural networks and deep learning have been used in quantitative rehabilitation assessment research (Kipf and Welling, 2017; Williams et al., 2019). Graph convolutional neural networks have been widely used in traffic prediction based on historical traffic speeds and route maps (Guo et al., 2019). It is also possible to realize action recognition and gesture recognition based on human skeleton data (Ahmad et al., 2021). According to current research, spatial–temporal graph convolutional networks (STGCN) have been used to achieve motion recognition based on dynamic bones (Yan et al., 2018). However, the application of STGCN in upper limb rehabilitation assessment is relatively limited. This study proposes to use an improved STGCN based on precise posture measurement to assess the motor function of hemiplegic upper limbs.

In this work, we proposed an innovative method as follows: two Azure Kinects is used combined with a comprehensive rigid body model to improve the biological feasibility of the skeleton. A hybrid particle swarm optimization algorithm is used to optimize inverse kinematics. A rich movement protocol is proposed to test the movement of the patient’s upper limbs from the reachable workspace and proximal function. A modified STGCN model with LSTM is proposed to assess the upper limb motor function.

2. Methods

We proposed an upper limb rehabilitation assessment method based on posture measurement, as shown in Figure 1. The upper limb rigid body model is established to increase the constraints of biological behavior and improve the accuracy of human posture data collection. The motion protocol for upper limb motion assessment is proposed, and the extended STGCN is adopted to achieve continuous upper limb rehabilitation assessment.

FIGURE 1
www.frontiersin.org

Figure 1. The upper limb rehabilitation assessment method based on posture measurement.

2.1. Two-camera synchronization

Azure Kinect can extract the position of 32 human bone skeletons, and there is occlusion when the upper limb moves to the back of the body. Using two Azure Kinect can effectively fill the occluded area and increase the spatial coverage of the camera. Therefore, we use two Azure Kinect to collect the motion of patients’ limbs in this study. During the use of two Azure Kinect, synchronization is necessary to ensure that each frame of data captured by the two cameras is the scene at the same time. One camera is set as master and the other as subordinate. The two cameras are connected via a 3.5 mm synchronization port attached to the device. This study adopts a daisy chain configuration, with the master’s synchronization port connected to the output synchronization port of the slave device through a cable. Then calibrate the two devices using the black and white checkerboard calibration method to obtain the internal and external parameters of the devices, and fuse joint data from different perspectives into the same perspective.

2.2. Models

Taking the torso as the base frame, the upper limbs can be modeled as two branches of the torso. The kinematic model of the right arm is as follows. The Kinect captured joints information, the Torso can be defined by the spine-chest and spine-naval markers. Anatomically, the shoulder is a complex composed of the glenohumeral joint, sternoclavicular joint, acromioclavicular joint and, the scapulothoracic joint. The glenohumeral joint can realize flexion/extension, adduction/abduction, and adduction/abduction. The sternoclavicular joint allows retraction/protraction, elevation/depression and backward of the glenohumeral joint. The elbow allows two movements for flexion/extension and pronation/supination. To simplify the human upper limb mechanism model, this paper singles out 2-DOF at the sternoclavicular joint, 3-DOF at the shoulder, 1-DOF at the elbow, 2-DOF at the wrist, and 1-DOF at the hand. Thus, the equivalent mechanism model of human upper limbs can be established as a 9-DOF series motion model, as shown in Figure 2. LSCAP is the initial length of the upper limb girdle, LUA is the length of the upper arm, LLA is the length of the forearm, LH is the length from palm to the wrist, LT is the length of the hand tip. The position of the hand, wrist, elbow, shoulder, clavicle, neck, and spine chest can be obtained by Azure Kinect. The base frame is fixed at the neck and the hand position is the palm position. Both left and right hands are modeled, and the right hand is taken as an example to illustrate the modeling process.

FIGURE 2
www.frontiersin.org

Figure 2. The rigid model of upper limb.

2.2.1. Rigid model

The upper limb rigid model consists of 10 segments connected by 11 joint markers. The human torso is modeled as the base of the rigid body, the neck joint of the torso is set as the origin, and the two scapulae rotate at the origin. The rigid body model is divided into two continuous chains of the left arm and the right arm. The motion of the torso (T) in the world coordinate system (W) is modeled as a system with associated homogeneous transformations:

TW,To=[RXRYRZt01]    (1)

Where R represents rotation, each rotation is determined by the angle θ, and t represents translation. Then model the scapula (SC), upper arm (UA), forearm (FA), hand (HA) and fingertip (TIP) as two branches of the trunk. The right arm is modeled as:

TTO,RSC=[RZRY001]    (2)
TRSC,RUA=[RXRYRZqRSJC01]    (3)
TRUA,RFA=[RXqREJC01]    (4)
TRFA,RHA=[RXRZqRHJC01]    (5)
TRHA,RTIP=[RXqRTJC01]    (6)

The left arm is modeled in a similar way, but the direction of rotation is opposite. The forward kinematics model of the rigid body can be obtained by multiplying the coordinate changes of each segment in turn. The positions of each joint can be written as:

ptor=TW,TO(t,θ)qtorpRSHO=TW,RSC(t,θ)qRSHOpRELB=TW,RUA(t,θ)qRELBpRWRI=TW,RFA(t,θ)qRWRIpRHAN=TW,RHA(t,θ)qRHANpRTIP=TW,RTIP(t,θ)qRTIP    (7)

Where q is the local position of each joint. The position of each joint p is solved according to the forward kinematics, and the mapping relationship is established. The forward kinematic map is:

F(xi)=[pTOR,pRSHO,pRELB,pRWRI,pRHAN,pRTIP,pLSHO,pLELB,pLWRI,pLHAN,pLTIP]    (8)

Scapulohumeral rhythm is present during arm abduction (Klopčar and Lenarčič, 2006) calculated the scapulohumeral rhythm of the generalized shoulder joint movement of the upper limb on four lifting planes with angles of 0°, 45°, 90° and 135° through experiments. The functional relationship between the lift angle β and the forward/backward extension angle θfb and the upward/downward angle θud of the SC joint is as follows.

θfb={0.35ββ<0°0°0°β70°0.22β+15.4°β>70°    (9)
θud={0.3ββ<0°0°0°β30°0.36β10.8°β>30°    (10)

2.2.2. Inverse kinematics optimized by crossbreed particle swarm optimization

The inverse kinematics of the rigid body model is a nonlinear problem. Solving the joint posture through the upper limb end posture is a one-to-many mapping relationship. The inverse kinematics is optimized based on a hybrid particle swarm optimization algorithm. The classical particle swarm optimization (PSO) algorithm belongs to a global random optimization algorithm with the advantages of few parameters required, simple algorithm structure, fast operation speed, etc. (Zhou et al., 2011). Suppose a D dimension search space has N particles, the position, and velocity of a particle in a group is,

Xi=[xi1,xi2,,xiD]TVi=[vi1,vi2,viD]T    (11)

The evolution of particles at each iteration consists of three parts: inheritance of the previous velocity, self-memory, and information exchange of the population. Therefore, the kth iteration process can be expressed as:

vij(k+1)=ωvij(k)+c1r1[pbest(k)xij(k)]+c2r2[gbest(k)xij(k)]    (12)
xij(k+1)=xij(k)+vij(k+1)    (13)

Where ω is inertia weight coefficient, c1 and c2 are two different learning factors, r1 and r2 are two randomly generated numbers in [0,1], pbest represents the personal best solution of the particle, gbest represents the global best solution of the swarm.

Due to the drawbacks of premature convergence and poor local optimization ability in PSO. Crossbreed Particle Swarm Optimization (CBPSO) is used to increase the fitness of the offspring population through the natural evolution of the population, thus jumping out of the local extreme value in the search process and converging to the global optimal solution. During the iteration process, the formula for updating the position and velocity of the offspring particles is as follows:

{child(x)=Pc×parent1(x)+(1Pc)×parent2(x)child(v)=parent1(v)+parent2(v)|parent1(v)+parent2(v)||parent1(v)|    (14)

where child(x) and child(v) represent the position and velocity of the child particle respectively, parent(x) and parent(v) represent position and the velocity of the parent particle, respectively. When two particles trapped in different local optimums are hybridized, they can often escape from the local optimality, and the introduction of a hybrid algorithm can enhance the global optimization ability of the population.

Our goal is to make the “distance” between the current end effector position F(xi) and yk the shortest. So inverse kinematics is transformed into an optimization problem.yk is the joint point collected by Azure Kinect. The fitness function is as follows:

fitness(X)=minykF(xi)2    (15)

The specific steps of the algorithm are: First initialize the particle swarm and parameter settings, and then iterate the algorithm to calculate the fitness function value of each particle, compare the current fitness of each particle with the size of the individual extreme value, update the individual extreme value, and judge whether the hybridization condition is met. If not, return to the continuous update. Finally, select the particles corresponding to the global extremum as the optimal solution for the population.

2.3. Extended STGCN

Human skeleton motion is a string of time series, with spatial features at each time point and time features between frames. In the process of evaluating the motor function of the upper limb, different joints play different roles. For example, in the movement of touching the nose with the right upper limb, the joints on the left side of the body participate less and show less importance, and the degree of participation of the joints on the right side is different. Self-attention mechanisms can select more critical information from a lot of information. The self-attention mechanism is adopted to extract the spatial relationship of each joint and distinguish their important degree, in order to guide patients to strengthen the rehabilitation training of important joints and obtain higher evaluation scores. The extended graph network structure is shown in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3. The extended STGCN for rehabilitation exercise assessment.

ConvLSTM can extract the characteristics of spatial and temporal features on time series data simultaneously (Deb et al., 2022). The STGCN is improved by the self-attention mechanism graph Sk calculating form ConvLSTM. The skeleton sequence is initially processed by temporal convolution with kernel Γμ.

Z=X(ΓuV)    (16)

Where, V is the skeleton sequence, let Qk=ZWk

ikt=σ(WiQkt+Uihkt1+bi)fkt=σ(WfQkt+Ufhkt1+bf)okt=σ(WoQkt+Uohkt1+bo)gkt=tanh(WcQkt+Uchkt1+bc)ckt=fktckt1+iktgkthkt=okttanh(ckt)    (17)

Where is convolution, σ is the sigmoid function, W is weight, b is bias, Sk=hk. The GCN improved with the self-attention map is as follows,

Gk=σ(ϕ(A^kSk)ZWk)    (18)
A^k=Ak+I    (19)

Where, A+I=kAk, ARN×N is the adjacency matrix, A0=I and A1=A, D is the degree matrix, Wk is the weight matrix. ϕ is normalization, σ is an activation function.

Then three Temporal convolutional (TC) layers with different kernels Γ1l, Γ2l and Γ3l is adopted to extract time features and concatenate them. Multiple STGCN layers are stacked to obtain more complex features of different lengths, and LSTM is used to extract the time dependence of the series. Finally, continuous rehabilitation assessment scores are obtained by the full connection layer.

2.4. Exercise protocols base on VR

The exercise protocols were designed according to the anatomical position, clinical evaluation methods such as the Fugl-Meyer scale, Barthel index, range of motion, and some related articles. The measurement of upper limb motor function mainly includes distal reachable workspace measurement and proximal function measurement, the specific movement methods are shown in Table 1. The reachable workspace measurement was used to evaluate the motion range of the upper limb, and the proximal functional measurement was used to evaluate the subjects’ ability to self-care in daily life, such as eating, combing hair, and so on.

TABLE 1
www.frontiersin.org

Table 1. The exercise protocols.

Vivid virtual reality (VR) scene modeling can improve the enthusiasm and initiative of patients to participate in rehabilitation assessment. In this manuscript, a virtual scene of motion assessment was built, in which the therapist demonstrates the action, and the subjects follow the therapist to carry out the same action. The subject’s avatar was designed and the visual feedback is applied to facilitate the subjects to observe whether their movements are completed or not. Auditory feedback was used to guide related movements with a variety of sensory stimuli, to increase the feasibility of the virtual environment demonstration. The rehabilitation training game is shown in Figure 4, when the position of the subject’s hand coincides with the minion, the animation of the minion jumping with the sound effect is played. The patient’s participation increased, and the patient’s tension and anxiety were relieved.

FIGURE 4
www.frontiersin.org

Figure 4. Virtual scene of motion assessment. In the rehabilitation training game, when the position of the subject’s hand coincides with the minion, the animation of the minion jumping with the sound effect is played. (A) The hand has not move to the minion position (B) Hand and minion position coincide (C) After the position of the hand and the minion overlap, the minion jumps and accompanies the sound effect.

3. Experimental validation

The experimental setup is as follows, two Azure Kinect depth cameras were placed on the tripod with a spacing of 2 m and an angle of 90°, and adjusted horizontally using a spirit level, as shown in Figure 5. During the experiment, the subjects were asked to perform the designed movements in front of the camera and could not rotate their bodies. To reduce the impact of accidental factors, explanations and related training were provided to the subjects before the experiment. The participants simulated the coach’s actions by watching pre-recorded videos on the display screen, enabling them to proficiently apply the assessment method before conducting relevant experiments.

FIGURE 5
www.frontiersin.org

Figure 5. The experimental setup.

This experiment recruited a total of 20 subjects, including 10 healthy individuals and 10 stroke patients. Among them, two rehabilitation physicians from the rehabilitation department of Nanjing Tongren Hospital voluntarily participated in the experiment. The exclusion criteria for participants in the experiment are cognitive impairment or inability to cooperate in the experiment. This work is approved by the local science and ethics committee.

4. Results

The data collected by Azure Kinect needs to be preprocessed to reduce individual differences, and eliminate migration and expansion during the experimental process. Median filtering can effectively eliminate isolated noise points. First, median filtering is performed on the data, and then the 6th-order low-pass Butterworth filter with a cut-off frequency of 30 Hz is used to filter again.

4.1. Model optimization result

According to the reachable workspace and proximal measurement in the exercise protocols, the validity of the optimized rigid body model is verified through the data of 10 healthy people. Taking the right upper limb as an example, the reachable workspace and its area of the upper limb was calculated by reference (Bai and Song, 2019). The brief description is as follows: the center of the upper limb workspace is at the shoulder joint, the motion trajectory is fitted using the least squares method, coordinate transformation is performed, Alpha Shape algorithm is used to locate the maximum boundary of the trajectory, Catmull-Rom spline interpolation is used to smooth the boundary, coordinate transformation is performed again, surface blocks are selected, the surface area is calculated, and normalization is performed. The reachable workspace is divided into four quadrants, with the first quadrant (blue) located on the inner side above the shoulder, quadrant 2 (pink) located on the inner side below the shoulder, quadrant 3 (red) located on the outer side above the shoulder, and quadrant 4 (green) located on the outer side below the shoulder. Figure 6 shows the reachable workspace results, A is the result measured by Azure Kinect, and B is the result optimized using the method we proposed.

FIGURE 6
www.frontiersin.org

Figure 6. Reachable workspace results. (A) Raw data. (B) Data after rigid model.

In the Figure 6A the solid lines represent the original trajectory information, the dashed lines represent the preprocessed results. The red ellipse marks show some points away from the track, or even skipping points. These points are not consistent with the biological characteristics of human movement. These spots can be caused by the arms facing the camera, moving too fast, or being blocked by the torso when extended backwards.

Due to the fan-shaped measurement range of the camera, occlusion can easily occur when the arm moves between the body and the camera, or the arm extends to the back of the body. At this point, a single camera cannot detect the position information of the occluded joint. Error signals will be detected at these occlusion positions, as shown in the ellipse inside Figure 6. Occlusion positions are points with obvious jumps and abrupt changes, which can easily lead to the phenomenon of unclosed fitting boundaries in the reachable workspace.

In this study, a part of the occlusion problems can be solved by using two cameras. The other part of the occlusion problem can be optimized by adding a rigid body model. After model optimization, the number of singularities was significantly reduced, the occurrence of non-biological motion was reduced and the accuracy and stability of hand joint motion measurement was improved.

Figure 7 shows the proximal function results, A is the result measured by Azure Kinect, and B is the result optimized using the method we proposed. From the comparison of Figures 7A,B, without the addition of a rigid body model, during the process of the upper limb touching the ear and the hand touching the lumbar spine, the trajectory did not reach the position of the ear/lumbar spine. After adding the rigid body model, the measurement results were improved, and the hand motion trajectory could reach the corresponding position.

FIGURE 7
www.frontiersin.org

Figure 7. Proximal function results. (A) Raw data. (B) Data after rigid model.

Table 2 shows the Signal Noise Ratio (SNR) of the motion trajectory of raw Kinect, two Azure Kinect, and two Azure Kinect with the rigid model. The raw Kinect trajectory exhibit low SNRs, especially in the Y and Z directions of the chest joint and the Z direction of the ipsilateral shoulder joint, the signal-to-noise ratio is below 10. The SNR of the final motion trajectory measured by two Kinects has increased, but there is still an SNR of less than 10. This study applied two Azure Kinect combined with a rigid body model, the measurement results show that all directions of each joint are greater than 10, and the SNRs are greater than 20 in the elbow joint, wrist joint, and hand joint. This table indicates that the method used in this study can improve the SNR of collected signals from each joint and increase the accuracy of upper limb posture recognition.

TABLE 2
www.frontiersin.org

Table 2. Signal to Noise ratio.

4.2. Assessment result

In rehabilitation assessment experiments, each subject underwent 5 exercises, with 30 groups tested each time and 10 groups resting for 10 min. A total of 3,000 sets of data were collected. Each action data in the dataset consists of a series of skeletal action frames. Each frame contains up to two skeletons, each with 11 skeletal nodes of the upper limbs. The data includes distal and proximal actions, with a total of 16 action categories, and each bone node has corresponding three-dimensional spatial coordinate data.

Three STGCN blocks are used. The optimization strategy of the model is the Adam optimizer, with a learning rate of 0.1, a batch size of 4, and an output space dimension of 80,40,40,80 for the LSTM layer. The model shares four LSTM layers with a dropout of 0.25.

The accuracy of the assessment model is measured using Mean Absolute Deviation (MAD), Root Mean Square Error (RMSE), and Mean Absolute Percentage Error (MAPE). The lower the error, the higher the accuracy of the model. The model was trained and predicted 10 times, and the obtained MAD, RMSE, and MAPE are recorded simultaneously. Finally, the average of the 10 results is taken to ensure the reliability of the results.

Each joint plays a different role in limb movement. Attention-guided graph convolution is used to extract spatial information, and each joint is processed differently based on the spatiotemporal frames, increasing the impact of different joints on the evaluation results. STGCN based on an attention mechanism makes the evaluation results more accurate, and can also provide guidance for rehabilitation and strengthen the training of important joints.

Figure 8 shows the impact of each joint on different assessed movements. As can be seen from the figure, in the movement of touching the nose with the right upper limb, the left joints of the body participate less in the movement and show lower importance. The degree of participation of the right joints varies, such as higher participation of the elbow joint and hand joint, and lower participation of the sternoclavicular joint. The wrist joint plays a crucial role in measuring the entire reachable workspace. The importance of different joints in different movements varies.

FIGURE 8
www.frontiersin.org

Figure 8. The role of joints in different movements. (The importance increases sequentially from 0 to 1. The vertical axis represents 16 evaluation actions, A * and E * represent the reachable workspace actions respectively, S1-H8 represents the proximal actions. The horizontal axis is an abbreviation for the names of each joint, from left hand to neck and then to the fingertip of the right hand).

Table 3 compares the performance of our proposed model, a single Azure Kinect and two Azure Kinect algorithms combined with STGCN. MAD, RMS, and MAPE are analyzed. It is obvious from the table that our proposed model finally gets the lowest evaluation error.

TABLE 3
www.frontiersin.org

Table 3. Assessment results.

5. Discussion

(1) Improve posture recognition accuracy

Accurate recognition of posture is key to the rehabilitation assessment of upper limb motor function using posture. However, human posture recognition is very complex and the accuracy is difficult to be guaranteed. Occlusion is easy to occur when using a single Kinect (Han et al., 2016) only used a one-generation Kinect to collect the reachable workspace of the upper limb in Duchenne muscular dystrophy, without proposing a method to solve the occlusion problem, resulting in low accuracy in human pose recognition (Matthew et al., 2020) also used only one-generation Kinect, with an improvement of adding a model. The model had fewer degrees of freedom and does not include the degrees of freedom of the wrist and hand. The accuracy of human body recognition measurement was not high, and there was a significant error in proximal upper limb movement.

Therefore, this study proposes to use two Azure Kinects and increase the constraint of the rigid body model at the same time, in order to reduce the inconsistency of human bone connection in biology, and then improve the accuracy of posture recognition. In the experimental results, Figure.6 contains the action of touching the lumbar vertebrae by hand. It can be clearly seen that the occlusion phenomenon is very obvious in the absence of a rigid body model, especially when the upper limb moves to coincide with other joints, the trajectory of the occluded part is messy and irregular, and the joint motion trajectory does not comply with human biology. The results of different test methods in the figure can obviously show the effectiveness of the proposed posture recognition method. Due to the inability of the camera to test the rotational motion of the arm, a more abundant human rigid body model is proposed to measure human posture from both attitude and position, achieving accurate posture recognition.

(2) Improve the accuracy of the assessment model

The effective and accurate assessment of upper limb motor function can provide the scientific basis for rehabilitation training, but the existing rehabilitation assessment methods lack universality, robustness, and practical relevance. Convolutional neural networks can be used to design scientifically reasonable quantitative assessment methods, but the accuracy of the assessment results still needs further verification. On the basis of improving the accuracy of human body recognition, this study conducts rehabilitation assessment tasks to increase the accuracy of assessment model recognition. Due to the varying degrees of participation of each joint in different movements, the importance of each joint is increased based on attention mechanisms. When the right hand is active, the participation of the left joint is lower, while when the left hand is active, the participation of the right joint is lower. At the same time, the importance of each joint is calculated for both the expert therapist and the patient. The difference in joint function between the patient and the therapist is significant. By comparing the difference in joint function with the average expert therapist, it can be determined which joints can be trained more effectively to improve the patient’s rehabilitation assessment score based on the size of the difference. The difference in joint function can provide a reasonable direction for rehabilitation training for patients. Adding joint participation to a rehabilitation assessment model can achieve continuous assessment and improve the accuracy of rehabilitation evaluation.

6. Conclusion

This study addresses the issue of non-quantification in rehabilitation assessment, and proposes an improved STGCN based on precise upper limb posture recognition to achieve continuous quantitative rehabilitation assessment. Two Azure Kinects were used to expand the field of view, a multi-degree of freedom upper limb motion rigid body model was proposed, making the upper limb posture measurement in line with normal human biological movement. The accuracy of upper limb posture recognition is increased, and the signal-to-noise ratio of measurement is improved. By identifying the participation degree of each joint in different movements based on the self-attention mechanism, the STGCN algorithm was improved to achieve continuous quantitative rehabilitation assessment. The experimental comparison results show that the upper limb posture recognition algorithm proposed in this study can effectively reduce incorrect joint coordinates, and the rehabilitation assessment model based on improved STGCN can effectively reduce the assessed MAD and RMS and MAPE. This study provides a new approach for the quantitative rehabilitation assessment of stroke patients. In the future work, we will continue to optimize the rigid body model and improve the rehabilitation assessment method.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving human participants were reviewed and approved by Nanjing Tongren Hospital science and ethics committee. The patients/participants provided their written informed consent to participate in this study.

Author contributions

JB was responsible for the study scheme design, data analysis, and manuscript writing. ZW was responsible for the data acquisition. XL was responsible for the experiment. XW was responsible for the revision of the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This work was supported by the Natural Science Foundation of Jiangsu Province (Grant no. BK20210930), Natural Science Research of Jiangsu higher education institutions of China (Grant no. 21KJB510039), and the Scientific Research Foundation of Nanjing Institute of Technology (Grant no. YKJ2019113).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Agami, S., Riemer, R., and Berman, S. (2022). Enhancing motion tracking accuracy of a low-cost 3D video sensor using a biomechanical model, sensor fusion, and deep learning. Front. Rehabil. Sci. 3:956381. doi: 10.3389/fresc.2022.956381

PubMed Abstract | CrossRef Full Text | Google Scholar

Ahmad, T., Jin, L., Zhang, X., Lin, L., and Tang, G. (2021). Graph convolutional neural network for human action recognition: a comprehensive survey. IEEE Trans. Artificial Intell. 2, 128–145. doi: 10.1109/TAI.2021.3076974

CrossRef Full Text | Google Scholar

Ai, Q. S., Liu, Z., Meng, W., Liu, Q., and Xie, S. Q. (2021). Machine learning in robot assisted upper limb rehabilitation: a focused review. IEEE Trans. Cog. Dev. Syst. :1. doi: 10.1109/TCDS.2021.3098350

CrossRef Full Text | Google Scholar

Bai, J., and Song, A. G. (2019). Development of a novel home based multi-scene upper limb rehabilitation training and evaluation system for post-stroke patients. IEEE Access. 7, 9667–9677. doi: 10.1109/ACCESS.2019.2891606

CrossRef Full Text | Google Scholar

Bawa, A., Banitsas, K., and Abbod, M. (2021). A review on the use of Microsoft Kinect for gait abnormality and postural disorder assessment. Hindawi Limited 2021:60122. doi: 10.1155/2021/4360122

PubMed Abstract | CrossRef Full Text | Google Scholar

Cai, S., Li, G., Zhang, X., Huang, S., and Xie, L. (2019). Detecting compensatory movements of stroke survivors using pressure distribution data and machine learning algorithms. J. Neuroeng. Rehabil. 16, 131–111. doi: 10.1186/s12984-019-0609-6

PubMed Abstract | CrossRef Full Text | Google Scholar

Deb, S., Islam, M. F., Rahman, S., and Rahman, S. (2022). Graph convolutional networks for assessment of physical rehabilitation exercises. IEEE Trans. Neural Syst. Rehabil. Eng. 30, 410–419. doi: 10.1109/TNSRE.2022.3150392

PubMed Abstract | CrossRef Full Text | Google Scholar

Fang, Q., Mahmoud, S. S., Gu, X., and Fu, J. (2019). A novel multi-standard compliant hand function assessment method using an infrared imaging device. IEEE J. Biomed. Health Inform. 23, 758–765. doi: 10.1109/JBHI.2018.2837380

PubMed Abstract | CrossRef Full Text | Google Scholar

Feigin, V. L., Brainin, M., Norrving, B., Martins, S., Sacco, R. L., Hacke, W., et al. (2022). World stroke organization (WSO): global stroke fact sheet 2022. Int. J. Stroke 17, 18–29. doi: 10.1177/17474930211065917

PubMed Abstract | CrossRef Full Text | Google Scholar

Guo, S., Lin, Y., Feng, N., Song, C., and Wan, H. (2019). Attention based spatial-temporal graph convolutional networks for traffic flow forecasting. Proc. AAAI Con. Artificial Intell. 33, 922–929. doi: 10.1609/aaai.v33i01.3301922

CrossRef Full Text | Google Scholar

Hamaguchi, T., Saito, T., Suzuki, M., Ishioka, T., Tomisawa, Y., Nakaya, N., et al. (2020). Support vector machine-based classifier for the assessment of finger movement of stroke patients undergoing rehabilitation. J. Med. Biol. Eng. 40, 91–100. doi: 10.1007/s40846-019-00491-w

CrossRef Full Text | Google Scholar

Han, J. J., Bie, E. D., Nicorici, A., Abresch, R. T., Anthonisen, C., Bajcsy, R., et al. (2016). Reachable workspace and performance of upper limb (PUL) in Duchenne muscular dystrophy. Muscle Nerve 53, 545–554. doi: 10.1002/mus.24894

PubMed Abstract | CrossRef Full Text | Google Scholar

Houmanfar, R., Karg, M., and Kulic, D. (2016). Movement analysis of rehabilitation exercises: distance metrics for measuring patient progress. IEEE Syst. J. 10, 1014–1025. doi: 10.1109/JSYST.2014.2327792

CrossRef Full Text | Google Scholar

Hussain, T., Maqbool, H. F., Iqbal, N., Khan, M., and Sanij, A. A. D. (2019). Computational model for the recognition of lower limb movement using wearable gyroscope sensor. Int. J. Sensor Networks. 30, 35–45. doi: 10.1504/IJSNET.2019.10020697

CrossRef Full Text | Google Scholar

Kipf, T. N., and Welling, M. (2017). Semi-supervised classification with graph convolutional networks. ICLR. doi: 10.48550/arXiv.1609.02907

CrossRef Full Text | Google Scholar

Klopčar, N., and Lenarčič, J. (2006). Bilateral and unilateral shoulder girdle kinematics during humeral elevation. Clin. Biomech. 21, s20–s26. doi: 10.1016/j.clinbiomech.2005.09.009

PubMed Abstract | CrossRef Full Text | Google Scholar

Lee, S., Lee, Y. S., and Kim, J. (2018). Automated evaluation of upper-limb motor function impairment using Fugl-Meyer assessment. IEEE Trans. Neural Syst. Rehabil. Eng. 26, 125–134. doi: 10.1109/TNSRE.2017.2755667

PubMed Abstract | CrossRef Full Text | Google Scholar

Li, C., Cheng, L., Yang, H., Zou, Y., and Huang, F. (2021). An automatic rehabilitation assessment system for hand function based on leap motion and ensemble learning. Cybern. Syst. 52, 3–25. doi: 10.1080/01969722.2020.1827798

CrossRef Full Text | Google Scholar

Li, C., Yang, H., Cheng, L., Huang, F., Zhao, S., Li, D., et al. (2022). Quantitative assessment of hand motor function for post-stroke rehabilitation based on HAGCN and multimodality fusion. IEEE Trans. Neural Syst. Rehabil. Eng. 30, 2032–2041. doi: 10.1109/TNSRE.2022.3192479

PubMed Abstract | CrossRef Full Text | Google Scholar

Liao, Y., Vakanski, A., and Xian, M. (2020). A deep learning framework for assessing physical rehabilitation exercises. IEEE Trans. Neural Syst. Rehabil. Eng. 28, 468–477. doi: 10.1109/TNSRE.2020.2966249

PubMed Abstract | CrossRef Full Text | Google Scholar

Liu, Z., Zhou, L., Leung, H., and Shum, H. P. H. (2016). Kinect posture reconstruction based on a local mixture of Gaussian process models. IEEE Trans. Vis. Comput. Graph. 22, 2437–2450. doi: 10.1109/TVCG.2015.2510000

PubMed Abstract | CrossRef Full Text | Google Scholar

Ma, C., Liu, Q., and Dang, Y. (2021). Multimodal art pose recognition and interaction with human intelligence enhancement. Front. Psychol. 12:769509. doi: 10.3389/fpsyg.2021.769509

PubMed Abstract | CrossRef Full Text | Google Scholar

Matthew, R. P., Seko, S., Bajcsy, R., and Lotz, J. (2019). Kinematic and kinetic validation of an improved depth camera motion assessment system using rigid bodies. IEEE J. Biomed. Health Inform. 23, 1784–1793. doi: 10.1109/JBHI.2018.2872834

PubMed Abstract | CrossRef Full Text | Google Scholar

Matthew, R. P., Seko, S., Kurillo, G., Bajcsy, R., Cheng, L., Han, J. J., et al. (2020). Reachable workspace and proximal function measures for quantifying upper limb motion. IEEE J. Biomed. Health Inform. 24, 3285–3294. doi: 10.1109/JBHI.2020.2989722

PubMed Abstract | CrossRef Full Text | Google Scholar

Paul, S., and Candelario-Jalil, E. (2021). Emerging neuroprotective strategies for the treatment of ischemic stroke: an overview of clinical and preclinical studies. Exp. Neurol. 335:113518. doi: 10.1016/j.expneurol.2020.113518

PubMed Abstract | CrossRef Full Text | Google Scholar

Qiao, C., Shen, B., Zhang, N., Jia, J., and Meng, L. (2022). Research on human-computer interaction following motion based on Kinect somatosensory sensor. In 2022 IEEE 6th Information Technology and Mechatronics Engineering Conference (ITOEC), 273–279

Google Scholar

Rahman, S., Sarker, S., Haque, A. K. M. N., Uttsha, M. M., Islam, M. F., and Deb, S. (2023). AI-driven stroke rehabilitation systems and assessment: a systematic review. IEEE Trans. Neural Syst. Rehabil. Eng. 31, 192–207. doi: 10.1109/TNSRE.2022.3219085

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, L., Huynh, D. Q., and Koniusz, P. (2020). A comparative review of recent Kinect-based action recognition algorithms. IEEE Trans. Image Process. 29, 15–28. doi: 10.1109/TIP.2019.2925285

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang, Y., Zhang, J., Liu, Z., Wu, Q., Chou, P. A., Zhang, Z., et al. (2016). Handling occlusion and large displacement through improved RGB-D scene flow estimation. IEEE Trans. Circuits Syst. Video Technol. 26, 1265–1278. doi: 10.1109/TCSVT.2015.2462011

CrossRef Full Text | Google Scholar

Williams, C., Vakanski, A., Lee, S., and Paul, D. (2019). Assessment of physical rehabilitation movements through dimensionality reduction and statistical modeling. Med. Eng. Phys. 74, 13–22. doi: 10.1016/j.medengphy.2019.10.003

PubMed Abstract | CrossRef Full Text | Google Scholar

Yan, S., Xi ong, Y., and Lin, D. (2018). Spatial temporal graph convolutional networks for skeleton-based action recognition. In Proceedings of the AAAI Conference on Artificial Intelligence.

Google Scholar

Zelai, S., and Begonya, G. (2016). Kinect-based virtual game for the elderly that detects incorrect body postures in real time. Sensors 16:704. doi: 10.3390/s16050704

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhou, D., Xiang, G., Liu, G., Mei, C., Dong, J., and Ying, L. (2011). Randomization in particle swarm optimization for global search ability. Expert Syst. Appl. 38, 15356–15364. doi: 10.1016/j.eswa.2011.06.029

CrossRef Full Text | Google Scholar

Keywords: rehabilitation assessment, upper limb, posture measurement, graph convolutional networks, motion range

Citation: Bai J, Wang Z, Lu X and Wen X (2023) Improved spatial–temporal graph convolutional networks for upper limb rehabilitation assessment based on precise posture measurement. Front. Neurosci. 17:1219556. doi: 10.3389/fnins.2023.1219556

Received: 09 May 2023; Accepted: 20 June 2023;
Published: 11 July 2023.

Edited by:

Lei Zhang, Nanjing Normal University, China

Reviewed by:

Xiaobin Xu, Hohai University, China
Hongdan Li, Shandong University, China

Copyright © 2023 Bai, Wang, Lu and Wen. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jing Bai, YmFpamluZ0Buaml0LmVkdS5jbg==

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.