Skip to main content

ORIGINAL RESEARCH article

Front. Aging Neurosci., 14 July 2022
Sec. Parkinson’s Disease and Aging-related Movement Disorders
This article is part of the Research Topic Advancing the Treatment Landscape in Parkinson’s Disease Using Sensor Technology and Data-Driven Modeling View all 6 articles

Recognition of Freezing of Gait in Parkinson’s Disease Based on Machine Vision

\r\nWendan Li,&#x;Wendan Li1,2†Xiujun Chen&#x;Xiujun Chen3†Jintao Zhang&#x;Jintao Zhang4†Jianjun LuJianjun Lu5Chencheng ZhangChencheng Zhang6Hongmin BaiHongmin Bai1Junchao LiangJunchao Liang1Jiajia WangJiajia Wang1Hanqiang DuHanqiang Du1Gaici XueGaici Xue1Yun LingYun Ling3Kang RenKang Ren3Weishen ZouWeishen Zou3Cheng ChenCheng Chen3Mengyan Li*Mengyan Li7*Zhonglue Chen,*Zhonglue Chen3,8*Haiqiang Zou,*Haiqiang Zou1,9*
  • 1Department of Neurosurgery, General Hospital of Southern Theater Command of PLA, Guangzhou, China
  • 2Graduate School, Guangzhou University of Chinese Medicine, Guangzhou, China
  • 3GYENNO SCIENCE Co., LTD., Shenzhen, China
  • 4Department of Neurology, The 960th Hospital of PLA, Taian, China
  • 5Department of Neurosurgery, Guangdong Second Provincial General Hospital, Guangzhou, China
  • 6Department of Neurosurgery, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
  • 7Department of Neurology, Guangzhou First People’s Hospital, Guangzhou, China
  • 8HUST-GYENNO CNS Intelligent Digital Medicine Technology Center, Wuhan, China
  • 9Branch of National Clinical Research Center for Geriatric Diseases, Chinese PLA General Hospital, Guangzhou, China

Background: Freezing of gait (FOG) is a common clinical manifestation of Parkinson’s disease (PD), mostly occurring in the intermediate and advanced stages. FOG is likely to cause patients to fall, resulting in fractures, disabilities and even death. Currently, the pathogenesis of FOG is unclear, and FOG detection and screening methods have various defects, including subjectivity, inconvenience, and high cost. Due to limited public healthcare and transportation resources during the COVID-19 pandemic, there are greater inconveniences for PD patients who need diagnosis and treatment.

Objective: A method was established to automatically recognize FOG in PD patients through videos taken by mobile phone, which is time-saving, labor-saving, and low-cost for daily use, which may overcome the above defects. In the future, PD patients can undergo FOG assessment at any time in the home rather than in the hospital.

Methods: In this study, motion features were extracted from timed up and go (TUG) test and the narrow TUG (Narrow) test videos of 50 FOG-PD subjects through a machine learning method; then a motion recognition model to distinguish between walking and turning stages and a model to recognize FOG in these stages were constructed using the XGBoost algorithm. Finally, we combined these three models to form a multi-stage FOG recognition model.

Results: We adopted the leave-one-subject-out (LOSO) method to evaluate model performance, and the multi-stage FOG recognition model achieved a sensitivity of 87.5% sensitivity and a specificity of 79.82%.

Conclusion: A method to realize remote PD patient FOG recognition based on mobile phone video is presented in this paper. This method is convenient with high recognition accuracy and can be used to rapidly evaluate FOG in the home environment and remotely manage FOG-PD, or screen patients in large-scale communities.

Introduction

Freezing of gait (FOG) is one of the most common motor symptoms in Parkinson’s disease (PD), mostly occurring in the intermediate and advanced stages of PD. FOG can be asymmetric, generally affecting one lower limb. Patients suddenly feel as if their feet are glued to the ground, it is difficult to lift their feet and take a step, and it is usually accompanied by a tremor in both legs (with a frequency of 6-8 Hz). This symptom usually lasts a few seconds but can sometimes exceed 30 s or last for a few minutes or more (Nutt et al., 2011). In the early stages of the disease, approximately 20% of PD patients report FOG episodes (Zhang et al., 2016), and its occurrence can increase up to 90% in the advanced stages (Hall et al., 2015). The greatest risk associated with FOG is falling. Prospective studies have revealed that 45-68% of PD patients fall each year (Wood et al., 2002; Allcock et al., 2009; Latt et al., 2009; Matinolli et al., 2011; Paul et al., 2014), and 60% of these falls are mainly caused by FOG (Pelicioni et al., 2019). Fractures from these falls can result in disability and prolonged bed rest, which can lead to a series of complications and even death. Therefore, the assessment of FOG symptoms is critical for preventive and protective measures.

Current methods for assessing FOG primarily fall into the following categories, but each method has some disadvantages:

(1) Evaluation methods based on specific gait tests or relevant rating scales (Giladi et al., 2000; Giladi et al., 2009; Nieuwboer et al., 2009; Seuthe et al., 2021): Although the rating scale evaluation method is simple and convenient, professional medical equipment and operation is not required. Therefore, this method is susceptible to environmental (Weiss et al., 2020) and subjective factors (Barthel et al., 2016). In addition, this assessment method is more difficult to implement for patients with cognitive impairments. Moreover, it is difficult to guarantee the consistency of evaluation results among different evaluators. Therefore, this method may lack high accuracy, repeatability and reliability.

(2) Evaluation methods based on neuroimaging examinations: Cerebral functional imaging technology, such as structural magnetic resonance imaging (MRI), functional MRI (fMRI), positron emission tomography (PET)/single-photon emission computed tomography (SPECT), etc., can dynamically detect cerebral functional activities and is now widely used in clinical practice (Djaldetti et al., 2018; Kim et al., 2018; Matar et al., 2019; Droby et al., 2021). These technologies provide potential imaging biomarkers for FOG research, but many researchers have only explained the correlation, not the causality, between examination results and FOG. A neuroimaging examination is considered reference data for FOG rather than an evaluation tool. In addition, the high expense and inconvenience of a neuroimaging exam make it unsuitable for clinical screening of large FOG populations.

(3) Evaluation methods based on smart devices: As objective evaluation methods, these technologies can remove the impact of human subjective factors. Currently, there are mainly two evaluation methods. The first method is based on wearable devices, and the second method is based on machine vision.

(1) Wearable device-based FOG evaluation methods: Both the sensitivity and specificity of FOG detection in PD patients based on sensor-related technology can reach a fairly high level (Sigcha et al., 2020), ideally suited for precision diagnosis in healthcare facilities. Although the current wearables are quite small in size and convenient to wear, the precision of equipment inevitably leads to an increase in cost and relatively complex installation or operation, and wearing multiple on-body sensors may induce some level of discomfort. In addition, the number, location, setting parameters and other operating points of sensors are not unified. Improper selection may lead to insufficient data collection or increase the difficulty of data analysis due to somewhat extreme environmental interference (Brognara et al., 2019), both of which pose significant challenges for wearable sensors in FOG recognition and monitoring.

(2) Machine vision-based FOG evaluation methods: Machine vision is a technology for evaluation that replaces human eyes with a machine. Compared with sensor wearables, machine vision-based methods do not require that a device be worn and neither induces discomfort in nor affects the motion of the evaluated patient; hence, it is an ideal objective evaluation method. There are two main ways to extract motion information with this method. (1) Some studies have leveraged the 3D motion capture system represented by the Kinect depth camera to extract motion information related to detecting gait disorders (Amini et al., 2019; Soltaninejad et al., 2019). However, these cameras require specialized equipment such as Microsoft Kinect, which is expensive. In addition, the complex structure of the assessment system can only be used in indoor environments, such as laboratory or clinic, resulting in a significant limitation of the evaluation environment, which is not conducive to the adoption of this technology in the screening and management of FOG populations. (2) Another method is using RGB technology of 2D keypoint recognition represented by OpenPose (Hu et al., 2019), which can estimate the joint coordinates of persons in the videos obtained using a monocular camera without external scales or markers. A study proposed a novel architecture of a graph convolutional neural network to detect FOG through 2D keypoint estimation and achieved good detection performance (Hu et al., 2020). Another study also used a deep learning method of convolutional 3D attention network to recognize FOG, which outperformed several state-of-the-art human action recognition methods (Renfei et al., 2018). The joint angles or other spatial parameters can be calculated by a home video camera or mobile phone camera at home without the special equipment needed to process 3D motion capture data. Although RGB cameras fail to cope with occluded bodies in multi-person tracking situations, RGB technology is low in cost, convenient, and easy to promote, which is very suitable for the screening of target populations in the community. Therefore, we finally chose RGB technology based on OpenPose as the preliminary scheme of our study (Viswakumar et al., 2019).

PD is a chronic disease, and patients need to occasionally go to the hospital to receive a follow-up check. However, the movements of PD patients are limited, which makes it inconvenient to visit a doctor. In addition, due to medical resource shortages, travel restrictions, and cross-infection in public places caused by the COVID-19 pandemic, it has been more difficult for PD patients to obtain diagnosis and treatment. Therefore, a remote assessment of FOG is necessary for doctors who require follow-up and management of long-term FOG patients.

To address the shortcomings associated with previous FOG recognition technologies, this study proposed constructing a FOG evaluation system based on mobile phone video. The timed up and go (TUG) test and narrow TUG (Narrow) videos of subjects were recorded by using an RGB camera, and used OpenPose to obtain the keypoint position signals of the human body and extract many representative time-domain and frequency-domain features. Then, these features were fed into the XGBoost classifier after feature selection to build the models. When modeling, we consider that a previous FOG recognition algorithm recognized FOG throughout the entire gait process, without considering the differences between the walking and turning stages. Due to the shooting angle, the visual aspects of walking and turning motions are completely different. Walking and turning gaits also vary in their kinematics. A logistic regression model determined that compared to a straight walk, turning has a reduced range of gait swing angle with significantly increased asymmetric gait and stride time (O’Day et al., 2020). Clinically, relevant brain and nervous system pathways during walking and turning are also different: turning tasks require more attention and involve greater interlimb coordination, increased coupling between posture and gait and modifications of locomotor patterns that require frontal lobe cognitive and executive functions that control posture transition (Nonnekes et al., 2019). Even under normal walking conditions, PD patients also often show longer turn times and a greater number of strides required to complete turning, which may be associated with impaired motion patterns when switching from walking to turning (Earhart, 2013). Based on the above reasons, we established motion recognition models and FOG recognition models that conform to different walking characteristics. Finally, we acquired a multi-stage FOG recognition model with acceptable performance through the leave-one-subject-out (LOSO) method, which realized automatic FOG recognition and monitoring.

Materials and Methods

Materials

Participants

The inclusion criteria were as follows: (1) participants were diagnosed with PD according to the Movement Disorder Society (MDS) diagnostic criteria (Postuma et al., 2015); (2) participants were diagnosed with FOG according to clinical manifestations and the FOG questionnaire; (3) participants were in the “on” state of medication; (4) participants could independently walk for more than 20 meters; (5) participants did not have cognitive dysfunction according to the Mini-Mental State Examination (MMSE >24 points) (Folstein et al., 1975); and (6) participants did not have any conditions affecting walking ability, such as hydrocephalus, cardiovascular and cerebrovascular disease, cognitive impairment, rheumatism, orthopedic disease, etc.

The exclusion criteria were as follows: participants with secondary PD causes, such as inflammatory, drug-induced, vascular, or toxin-induced PD, or participants with other neurodegenerative diseases, such as progressive supranuclear palsy (PSP) or multiple system atrophy (MSA) and other Parkinson-plus syndromes.

The research protocol in compliance with the Declaration of Helsinki was approved by the Scientific Ethics Committee of General Hospital of Southern Theater Command of PLA and the Ethics Committee of Guangzhou First People’s Hospital. All participants came from the General Hospital of Southern Theater Command of PLA and the Guangzhou First People’s Hospital, and written informed consent was obtained from all participants (or their legal guardians).

Dataset

Video Capture

Each subject completed a TUG (Shumway-Cook et al., 2000) and a narrow TUG, and the test routes are shown in Figure 1. A narrow test was added because it is easier to trigger FOG in PD subjects by passing through a narrow space (Snijders et al., 2008).

FIGURE 1
www.frontiersin.org

Figure 1. Test diagram: (A) TUG diagram; (B) narrow TUG diagram.

The gait test process of this study was as follows: The photographer was more than 4.4 m away from the patient and recorded the videos with a mobile phone [Huawei P40 (8G + 256G)]. The frame rate of the video was 30 frames per second, with a resolution of 544*960.

After the photographer gave walking instructions, the patient started to walk, and at the same time, the photographer began to record with the mobile phone. The patient walked 3 meters in a straight line at a comfortable and free speed, turned back at the end point, and finally returned to the starting point. A narrow TUG is designed to induce FOG more easily by setting up a narrow 0.6-meter tunnel during walking. The whole process is repeated twice.

To guarantee consistency and quality of test data, the experiment was performed in an open field, which ensured that the patient was in the video during the whole recording. To prevent the subjects from falling, the participants could use a walking stick, walking aid and other auxiliary tools. The participants with a high risk of falling were accompanied by a caregiver during the walk.

Clinical Information Record

After subjects completed the above gait tests, we collected their clinical information, including MDS-United Parkinson’s Disease Rating Scale (UPDRS) Part 3 scores (Stebbins et al., 2013), Hoehn & Yahr ratings (stage 1-5) (Hoehn and Yahr, 1967), New Freezing of Gait Questionnaire (NFOG-Q) scores (Giladi et al., 2000) and Parkinson’s Disease Questionnaire-39 (PDQ-39) scores (Moore et al., 2007).

Video Annotation

To build a binary classification model, we independently annotated the turning stage and FOG stage. The video annotations of the turning and FOG stages were frame-based.

Regarding the annotation of the turning stage, the data reviewer annotated the turn-start frame and the turn-end frame. The annotation of the FOG stage was completed by two independent movement disorder specialists, and this served as the gold standard evaluation for FOG detection. More specifically, two independent neurologists separately identified the start and end of FOG episodes, and in case of discrepancies, a third neurologist was invited to make a joint decision. The three neurologists performed a common assessment to resolve any ambiguities. The start of the FOG episode was defined when the patient hesitated for more than one second at the start of walking or if it looked as if he or she was unsuccessfully trying to initiate or continue locomotion. A transient and clinically significant break in locomotion without any apparent reason was also defined as a freezing episode. The end of an episode was defined as the time when an effective step had been performed with a relatively normal length and swing, and the step also had to be followed by a continuous normal walk (Schaafsma et al., 2003). We recorded the entire gait test, such that the data were collected from the beginning of the gait test after the subjects heard the instructions to the end of the gait test. Due to the complexity of the definition of FOG, it is difficult to label all real FOG in practice. Based on the recognition of FOG for labeling purposes, we labeled the parts of the video that could be marked as FOG with certainty, and the parts of the video where it was unclear whether it was FOG would be labeled as unknown. For example, when an uninstructed stop occurred, we cannot tell whether the individual stopped to rest or FOG occurred, and these instances were marked unknown.

The dataset for this study was composed of the video of the subjects’ gait tests, the recorded scale information, and the video annotation.

Methods

The five steps of this study were keypoint position signal extraction, signal preprocessing, feature extraction, modeling and algorithm evaluation. The algorithm flowchart is shown in Figure 2.

FIGURE 2
www.frontiersin.org

Figure 2. Algorithm flowchart.

Keypoint Position Signal Extraction

After videotaping the subjects’ tests via mobile phone, we used OpenPose (Cao et al., 2021) to extract position signals of 25 keypoints of the human skeleton to conduct 2D human motion perception. The OpenPose human gesture recognition project is an open-source library developed by US Carnegie Mellon University based on a convolutional neural network and supervised learning with Caffe as the framework that achieves human body motion, facial expression, finger movement and other posture estimates easily and accurately from images.

Signal Preprocessing

Human movement mainly has frequency components between 0 and 20 Hz. In addition, most previous research on FOG recognition based on wearable sensors used a 15-Hz lowpass filter for denoising (Rodriguez-Martin et al., 2017; Sigcha et al., 2020). According to the Nyquist–Shannon sampling theorem (Nyquist, 1928; Shannon, 1949), a 30-Hz sampling frequency will sample motion information below 15 Hz. A higher sampling frequency will result in more high-frequency information, but there will be additional mobile requirements that are not conducive to future applications.

2D images have the problem of objects being near appear large and those that are far appear small, which results in different scales at different distances. Thus, we normalized the keypoint position signals from the subjects’ bodies to eliminate the effects of the different scales. First, we calculated the minimum enclosing rectangle of the human body based on 25 keypoints and expanded the length and width by 30%; then, we converted the original coordinate system to a coordinate system with the upper left corner vertex of the enclosing rectangle as the coordinate origin to obtain the position coordinates of the 25 keypoints after coordinate system conversion. Next, a ratio of 80 to the height of the enclosing rectangle was regarded as the scale factor, and this scale factor was multiplied by the position coordinate of each keypoint to obtain the normalized position coordinates. The keypoint position signal normalization diagram is shown in Figure 3.

FIGURE 3
www.frontiersin.org

Figure 3. Keypoint position signal normalization diagram.

Based on the normalized position coordinates of the keypoints in each frame, we calculated the speed signal, acceleration signal, and knee joint angle signal from the keypoints, and we calculated the absolute value signals of 8 pairs of keypoint position differences (right hip and left hip, right knee and left knee, right ankle and left ankle, right big toe and left big toe, right shoulder and left shoulder, right elbow and left elbow, right wrist and left wrist, and right ear and left ear). Among these measures, the speed signal reflects the subject’s movement speed, and the acceleration signal reflects the subject’s change in velocity. When the subject has FOG, leg movements become more obvious. By adding the angle signals from the subject’s knees, bending information of the leg joint can be extracted. Furthermore, the absolute value signals of 8 pairs of keypoint position differences showed a trend of first decreasing and then increasing when turning around, and thus, reflected the turning process.

Feature Extraction

Before feature extraction, we first performed sliding window processing on the signal. Sliding window signal processing is a routine processing operation, and we will not focus on this as the research priority. According to previous studies, we concluded that a window size of 2 s would yield good results (Zach et al., 2015); we adopted a sliding window with a step length of 0.1 s and a window size of 2 s to convert the signal into windows and calculated the time-domain and frequency-domain features of each window.

With respect to the motion recognition model, we extracted the minimum value, crest factor, frequency amplitude peak value, center frequency and other time-domain and frequency-domain features to reflect the differences between the walking and turning stages. With respect to the FOG recognition model, we also extracted time-domain and frequency-domain features and added features such as the freezing index (FI), the area under the power spectrum of multiple frequency bands including low-frequency bands and high-frequency bands, as well as features that may contribute to the recognition of FOG. The FI was defined as the power in the “freeze” band (3-8 Hz) divided by the power in the “locomotor” band (0.5-3 Hz). Some researchers (Moore et al., 2013) proposed that the FI can detect FOG episodes based on changes in the inertial signal power spectrum. In addition, Bächlin et al. (2009) used the power band (0.5-8 Hz) to avoid false detections when standing. In clinical observation, FOG is frequently accompanied by fast alternating trembling movements (a large increase in power in the ‘freeze’ band) of the lower extremity with both feet on the floor (Schaafsma et al., 2003; Nutt et al., 2011). The “locomotor” band is the energy band representing low frequency. It can be used to discriminate FOG (i.e., power is significantly decreased or absent in the locomotor band) from a normal gait (i.e., most of the power is in the locomotor band). Thus, FI was considered to be beneficial for FOG recognition, and this feature was added to the feature set.

Modeling

Next, a motion recognition model, the Walk-FOG recognition model and Turn-FOG recognition model, were built based on the collated data.

When building the motion recognition model, an excessive number of model features may lead to overfitting or other problems. As such, we carried out feature selection to improve the generalization ability of the classifiers and reduce the time required to train the classifier. First, we calculated the information gain of each feature based on XGBoost (Chen and Guestrin, 2016); the larger the information gain of the feature is, the greater the contribution of this feature to the model. Then, we used the forward feature selection strategy based on the information gain to select the best feature set. In addition, we searched for optimal model parameter combinations using grid search and the LOSO method to obtain the optimal XGBoost-based motion recognition model. The model hyperparameters for the search include the learning rate, the number of estimators (n_estimators), the max depth (max_depth), the sampling ratio (subsample), the feature sampling ratio for each tree (colsample_bytree).

When building FOG recognition models for the walking and turning stages, due to unbalanced samples caused by sudden FOG, we first adopted the SMOTE (Chawla et al., 2002) algorithm to balance the samples in the training set. Then, we performed the same feature selection strategy as we did with the motion recognition model. Considering that the difficulties in distinguishing samples at the transition stages are non-FOG or FOG, if they are entered during model training, the results will be affected by ambiguous labels, and performance would decline. To obtain a more precise classifier model for the FOG and non-FOG classes, we removed the samples at the transition stages of FOG and non-FOG for training but not for testing (i.e., the samples with a duration of 0.5 s for FOG and non-FOG were removed, as shown in Figure 4) (Mazilu et al., 2013). Finally, we obtained the optimal model parameter combination of the XGBoost-based FOG recognition model using grid search and the LOSO method, thereby acquiring the optimal Walk-FOG recognition model and Turn-FOG recognition model. The model hyperparameters for the search were the same as those for the motion recognition model.

FIGURE 4
www.frontiersin.org

Figure 4. Removal of the transition samples between FOG and non-FOG.

Multi-Stage Freezing of Gait Recognition Model

We applied the motion recognition model to segment the walking and turning stages and utilized the Walk-FOG recognition model and Turn-FOG recognition model for FOG recognition in the walking and turning stages, respectively. With the strategy of first segmenting the motion stages and then recognizing FOG, we formed a multi-stage FOG recognition model. The evaluation of the FOG recognition model was episode-based. A FOG sequence composed of continuous FOG windows was considered a FOG episode; a non-FOG sequence composed of continuous non-FOG windows was considered a non-FOG episode.

Some short-lived false positive (FP) FOG events occasionally appeared in the recognition results. Therefore, with the aim of reducing false detections, we modified the recognition results of the window. Specifically, we took the 10% quantile of all the marked FOG durations as the threshold and used this value as the minimum duration necessary for a FOG episode. If the duration of a FOG episode was less than this value, we removed this FOG episode.

Algorithm Evaluation

Evaluation Metrics of the Motion Recognition Model

Motion recognition model evaluation metrics are window-based sensitivity, specificity, accuracy, geometric mean (GM) and area under the curve (AUC). GM is the square root of sensitivity and specificity, and this indicator has the beneficial property of averaging out sensitivity and specificity scores while penalizing unbalanced pairs. AUC is the area under the receiver operating characteristic (ROC) curve, a common indicator used to evaluate the performance of a classification model; the higher the AUC value is, the better the model performance.

Evaluation Metrics of the Walk-FOG Recognition Model, Turn-FOG Recognition Model and Multi-Stage FOG Recognition Model

The evaluation metrics of these three models are episode-based sensitivity, specificity, accuracy and GM. Sensitivity is defined as the proportion of the number of correctly predicted FOG episodes (at least one window is predicted as FOG in FOG episodes) to the total number of FOG episodes; specificity is defined as the proportion of the number of correctly predicted non-FOG episodes (no window is predicted as FOG in non-FOG episodes) to the total number of non-FOG episodes; accuracy is the proportion of the number of correctly predicted episodes to the total number of episodes; GM is the square root of sensitivity and specificity.

Evaluation Method

Previous studies (Mazilu et al., 2012; Tripoliti et al., 2013; Sato et al., 2019), despite high sensitivity and specificity, all have some obvious problems in their evaluation methods. For example, these evaluation methods include leave-one-window-out, which trains the model using all windows except one window. Therefore, training samples with high similarity to the test samples are used for model training. With suspected data leakage, the accuracy is very high. To prevent the model evaluation results from being unrealistically high, we used the LOSO method to evaluate the performance of this algorithm on this dataset. LOSO is one of several cross-validation methods; our dataset contains 50 subjects, therefore, there is 50-fold cross-validation. For each fold of cross-validation, the data from one subject are reserved for testing, and the data of the remaining subjects are used for training. This validation method can guarantee that the data of each subject only appear in the training set or test set, which helps to statistically estimate the performance for unseen subjects (Das et al., 2011; Rodriguez-Martin et al., 2017; Hu et al., 2019).

Statistical Analysis

All continuous demographic data and clinical data are expressed as the mean ± SD or as a percentage. Boxplots (Toit et al., 1986) are used to explore the correlations between the walking and turning stages, FOG and non-FOG stages and features. We adopted the Wilcoxon rank-sum test to analyze whether the features used in the model analysis significantly differed between groups, such as feature differences between the walking and turning stages and differences between the FOG and non-FOG stages. Considering that the sample size was large (>30), that there were individual differences among samples, and that there were autocorrelations between time series samples, we designed a hypothesis testing method suitable for this case. First, we calculated the mean value of the feature for each subject in the two groups to obtain two sets of paired samples from the same subject; then, the paired sample Wilcoxon test was used to test whether these features of the subjects differed between groups. All figure generations and statistical analyses were performed by R 4.1.0 or Python 3.8.

Results

Dataset

This study included convenience sample 50 PD-FOG subjects who met the criteria. Their basic information is shown in Table 1. Each subject was videotaped in the TUG and narrow TUG, resulting in a total of 200 videos being collected. After eliminating the problems of incomplete video shooting, such as video jitter or walking routes that did not conform to the requirement, 100 qualified videos were selected and further excluding unqualified videos that contained severe joint occlusion and incorrect skeleton recognition, 89 qualified videos were used for modeling (a flowchart describing the dataset collection is shown in Figure 5). The total length of the videos was 50.78 min; from this time period, the length of video including turning was 12.88 min, and that including walking was 37.9 min. Thirteen subjects had FOG during the gait test (26%), and a total of 33 FOG episodes were captured, including 25 FOG episodes in the walking stage and 8 in the turning stage, as shown in Table 2; the median FOG duration was 8.2 s (0.9-66 s).

TABLE 1
www.frontiersin.org

Table 1. Demographics of the subjects.

FIGURE 5
www.frontiersin.org

Figure 5. Dataset collection flowchart.

TABLE 2
www.frontiersin.org

Table 2. Frequency of FOG in 13 PD patients.

Feature Analysis

According to the forward feature selection strategy based on the information gain of XGBoost, 50, 20, and 12 features were selected for the motion recognition model, Walk-FOG recognition model, and Turn-FOG recognition model, respectively. We present these features and performed analyses.

Motion Recognition Model

After feature selection, the top 50 features of information gain were selected to build the model, of which the top 10 are shown in Table 3. Six of the top 10 features were obtained by calculating the absolute value signal of keypoint position differences, which shows that such signals best reflected the difference between walking and turning motions. To analyze the contribution of particular features to the model, we enumerated two important features used in motion recognition models and showed the stage comparison chart and boxplot, observed their numerical changes across the two stages, and analyzed the differences between the walking and turning groups. The stage comparison chart and boxplot of the minimum value (top 1) were calculated based on the absolute value signal of the position difference between the left and right shoulders in the motion recognition model (Figures 6A,B). As shown in Figure 6A, the numerical value of this feature decreased sharply when turning because the absolute value signal of the keypoint position difference first decreases to 0 and then increases during the turning process. As a result, the minimum value of this signal also showed the same change; as shown in Figure 6B, the overall numerical value of this feature when walking was significantly greater than that when turning, with a statistically significant difference (p < 0.001). Additionally, the numerical range while walking was smaller than that while turning, indicating that the feature fluctuated less when walking. The stage comparison chart and boxplot of the crest factor (top 6) were calculated by the absolute value signal of the position difference between the right elbow and the left elbow in the motion recognition model (Figures 6C,D). The crest factor is the ratio of the maximum value to the root mean square. Figure 6C shows that the numerical value of this feature increased when turning because the maximum value of this signal was greater than the degree of fluctuation in the signal. Figure 6D shows that the overall numerical value of this feature when turning was apparently greater than that when walking, with a significant difference between groups (p < 0.001). In conclusion, we believe that these features show synchronous changes and significant numerical differences when walking and turning, which markedly contributed to motion classification of the model and could facilitate subsequent research.

TABLE 3
www.frontiersin.org

Table 3. Feature importance for the motion recognition model based on the XGBoost gain (top 10).

FIGURE 6
www.frontiersin.org

Figure 6. Motion feature recognition boxplot and stage comparison chart analysis. (A) The stage comparison chart of the minimum value calculated by the absolute value signal of the position difference between the left and right shoulders. (B) Boxplot of the minimum value calculated by the absolute value signal of the position difference between the left and right shoulders. The feature was significantly different between the turning and walking groups (p < 0.001). (C) Stage comparison chart of the crest factor calculated by the absolute value signal of the position difference between the right wrist and the left elbow. (D) Boxplot of the crest factor calculated by the absolute value signal of the position difference between the right wrist and the left elbow. The feature was significantly different between the turning and walking groups (p < 0.001).

Freezing of Gait Recognition Model

Similarly, we selected the top 20 features according to the information gain of the XGBoost algorithm to build the Walk-FOG recognition model (Table 4), and the top 12 features were used to build the Turn-FOG recognition model (Table 5). Comparing the features of the Walk-FOG recognition model and Turn-FOG recognition model, we found a great discrepancy between the important features, which illustrates the necessity of staged recognition for FOG from the perspective of selected features. We found that the features of the left and right limbs were different, which may be because the results were data-driven, and the samples were not balanced on FOG of the left and right limbs.

TABLE 4
www.frontiersin.org

Table 4. Feature importance for the Walk-FOG recognition model based on the XGBoost gain (top 20).

TABLE 5
www.frontiersin.org

Table 5. Feature importance for the turn-FOG recognition model based on the XGBoost gain (top 12).

As shown in Tables 4, 5, many features used for building the Walk-FOG recognition model and Turn-FOG recognition model were the area under the power spectrum of the 0.5-3 Hz frequency band calculated from various signals, which are conducive to FOG recognition. The 0.5-3 Hz frequency band is the energy band representing low frequency. As Moore et al. (2013) showed in their study, the area under the power spectrum in this band increased during normal walking and decreased during FOG. They also showed a decrease in FI when patients experienced FOG, which is consistent with the reduced area under the power spectrum of the 0.5-3 Hz frequency band as the denominator in the calculation of FI. The “locomotor” band (0.5-3 Hz) of the FI is useful for recognizing FOG. In addition, the Turn-FOG recognition model used the features that included the area under the power spectrum of the 3.5-15 Hz frequency band (Table 5), and this result was similar to that of Moore et al. (2013). They found that the area under the power spectrum of the 3-8 Hz high frequency band increased when subjects experienced FOG.

To further illustrate the contribution of features to the FOG recognition model, we analyzed two features used in the Walk-FOG recognition model and Turn-FOG recognition model and presented them in a visualized manner. The stage comparison chart and boxplot of the area under the power spectrum of the 0.5-3 Hz frequency band (top 1) was calculated based on the Y-axis acceleration signal from the right heel in the Walk-FOG recognition model (Figures 7A,B). The 0.5-3 Hz frequency band represents the low-frequency energy band. When FOG occurs, there is no significant movement of the subject’s legs, so the low-frequency energy will be considerably reduced. Therefore, as shown in Figure 7A, the numerical value of this feature was significantly reduced and remained low for a long time; as shown in Figure 7B, the numerical value of this feature was very low when FOG occurred, and there was a significant numerical difference between the FOG and non-FOG groups (p = 0.010). The stage comparison chart and boxplot of the area under the power spectrum of the 1-1.5 Hz frequency band (top 8) was calculated based on the Y-axis speed signal from the left heel in the Walk-FOG recognition model (Figures 7C,D). The 1-1.5 Hz frequency band is also a low-frequency energy band. When the subject has FOG, the energy in this frequency band also decreased. As shown in Figure 7C, this feature was reduced to close to zero when FOG occurred. Moreover, there was also a significant numerical difference between the FOG and non-FOG groups (p = 0.004) (Figure 7D).

FIGURE 7
www.frontiersin.org

Figure 7. Feature boxplot and stage comparison chart analysis in the walking stage. (A) Stage comparison chart of the area under the power spectrum of the 0.5-3 Hz frequency band calculated by the Y-axis acceleration signal from the right heel. (B) Boxplot of the area under the power spectrum of the 0.5-3 Hz frequency band calculated by the Y-axis acceleration signal from the right heel. The feature was significantly different between the FOG and non-FOG groups (p = 0.010). (C) Stage comparison chart of the area under the power spectrum of the 1-1.5 Hz frequency band calculated by the Y-axis speed signal from the left heel. (D) Boxplot of the crest factor calculated by the area under the power spectrum of the 1-1.5 Hz frequency band calculated by the Y-axis speed signal from the left heel. The feature was significantly different between the FOG and non-FOG groups (p = 0.004).

The stage comparison chart and boxplot of the area under the power spectrum of the 0.5-3 Hz frequency band (top 1) was calculated based on the X-axis acceleration signal from the right heel in the Turn-FOG recognition model (Figures 8A,B). The stage comparison chart and boxplot of the area under the power spectrum of the 0.5-3 Hz frequency band (top 3) was calculated based on the X-axis speed signal from the left heel in the Turn-FOG recognition model (Figures 8C,D). Both features involve the area under the power spectrum of the 0.5-3 Hz frequency band. When the numerical value of this feature suddenly decreased, it indicated that the subject may have FOG. As shown in Figures 8A,C, when FOG occurred, the numerical value of this feature suddenly decreased. As shown in Figures 8B,D, both numerical values of the two features were lower when FOG occurred, with smaller overall fluctuations as well, and the P values also passed the significance test (p = 0.031 and p = 0.031, respectively).

FIGURE 8
www.frontiersin.org

Figure 8. Feature boxplot and stage comparison chart analysis in the turning stage. (A) Stage comparison chart of the area under the power spectrum of the 0.5-3 Hz frequency band calculated by the X-axis acceleration signal from the right heel. (B) Boxplot of the area under the power spectrum of the 0.5-3 Hz frequency band calculated by the X-axis acceleration signal from the right heel. The feature was significantly different between the FOG and non-FOG groups (p = 0.031). (C) Stage comparison chart of the area under the power spectrum of the 0.5-3 Hz frequency band calculated by the X-axis speed signal form the left heel. (D) Boxplot of the area under the power spectrum of the 0.5-3 Hz frequency band calculated by the X-axis speed signal from the left heel. The feature was significantly different between the FOG and non-FOG groups (p = 0.031).

Model Performance

According to the above modeling methods, the motion recognition model has a recognition sensitivity of 81.21%, specificity of 90.16% and accuracy up to 87.75% (see Table 6). We also built FOG recognition models for the walking and turning stages. The sensitivity of the Walk-FOG recognition model was 71.43%, and the specificity was 85.71%; the sensitivity of the Turn-FOG recognition model was 61.54%, and the specificity was 92.59%. Finally, we combined the motion recognition model, Walk-FOG recognition model and Turn-FOG recognition model into a multi-stage FOG recognition model, with a specific strategy of using the motion recognition model to segment different motion stages and then to use the FOG recognition model in each stage to recognize FOG. Finally, we modified the window recognition results of the FOG recognition model and obtained an episode-based sensitivity and specificity of 87.50 and 79.82%, respectively, as shown in Table 6.

TABLE 6
www.frontiersin.org

Table 6. Performance of each model.

Model Comparison

We built the non-staged FOG recognition model using the same method as the staged FOG recognition model (i.e., the multi-stage FOG recognition model) and compared the performance of the two models (see Table 6). The results suggested that both the sensitivity and specificity of the staged FOG recognition model was higher by 3%, proving better recognition performance of the staged FOG recognition model than the non-staged FOG recognition model.

Discussion

This study proposes a new method of automatically recognizing FOG based on mobile phone video. We collected 89 qualified videos from 50 PD subjects, built a multi-stage FOG recognition model, evaluated the recognition performance of this model using the LOSO method, and obtained a relatively sound result (sensitivity: 87.50%; specificity: 79.82%). Although the specificity is slightly lower than the sensitivity, community-based screening emphasizes higher sensitivity. Thus, this model meets the preliminary criteria for clinical needs (Leeflang et al., 2013; Barry et al., 2014).

The TUG, a simple and effective way to assess an individual’s basic functional mobility, includes stages where FOG is most likely to occur, such as starting, turning, and reaching goals (Shumway-Cook et al., 2000). When designing the test, we also added a 3-m narrow test to induce FOG. When considering how long to make the TUG for use in a gait analysis based on machine vision, we read the relevant literature and found that gait parameters [including step length (SL), step width (SW), step duration (SD), single-stance duration (SSD) and double-stance duration (DSD)] extracted from 3-m TUG videos can be used as an effective and stable gait quantification method (Aberg et al., 2021). A previous study segmented patients’ starting, turning, and walking tasks from the video of the 3-m TUG to classify patients based on disease status (Tang et al., 2022). In addition, a similar study formulated vision-based FOG detection from 3-m TUG videos from 45 PD patients(Hu et al., 2019). This study showed that the 3-m TUG is widely used for machine vision gait analysis, which makes the result of the study more comparative. Therefore, using a 3-m TUG is a sound choice.

Bächlin et al. (2010) and Moore et al. (2013) found that when subjects experienced FOG, the area under the power spectrum of the 0.5-3 Hz frequency band decreased, and the area under the power spectrum of the 3-8 Hz frequency band increased. Therefore, these two features were used in FOG recognition. The energy of the low frequency band reflects normal activity. When the energy of the low frequency band is reduced, the normal activity decreases, while the FOG is usually accompanied by high-frequency tremor. As a result, the high frequency band reflects the FOG. Therefore, low-frequency and high-frequency features are useful for FOG recognition. As shown in Table 5, we used the low-frequency feature, the area under the power spectrum of the 0.5-3 Hz frequency band and the high-frequency feature, the area under the power spectrum of the 3.5-15 Hz frequency band, to build the FOG recognition model. We used the XGBoost algorithm combined with the forward feature selection strategy to select the features. Many useful frequency domain and time domain features can be obtained through this method, as shown in Tables 4, 5. All the low-frequency and high-frequency features used for FOG recognition are shown in Table 8. In addition to discovering low-frequency features and high-frequency that can be used for FOG recognition as previous, we also applied time domain features (Rodriguez-Martin et al., 2017; Sigcha et al., 2020) in FOG recognition and found that some time domain features have a great contribution in the FOG recognition model, such as the clearance factor calculated by position signal of L big toe for the Walk-FOG recognition model (see Table 4) and the min calculated by the position signal of R small toe for the Turn-FOG recognition model (see Table 5). These features, which are important for detecting FOG, were data-driven.

TABLE 7
www.frontiersin.org

Table 7. NFOG-Q prediction results.

TABLE 8
www.frontiersin.org

Table 8. The low-frequency and high-frequency features for the Walk-FOG recognition model and Turn-FOG recognition model.

This study realized FOG recognition algorithm with relatively high sensitivity and specificity, which has preliminary clinical feasibility. Considering that the clinical need to assess the severity of FOG, we also attempted to predict scores of the third item (FOG frequency when turning), the sixth item (longest duration of FOG appearing at the start of the step) and the seventh item (evaluation of FOG impact on walking in daily life) on the NFOG-Q that has strong clinical objectivity. In practice, after using OpenPose to extract the position signals of the subject’s keypoints based on their skeleton, we calculated time-domain and frequency-domain features. Then, we built an ordered logistic regression model and evaluated the prediction performance of the model with the LOSO method. The prediction accuracy of all three items exceeded 70% (see Table 7). For these measures, ACC ± 0 represents the accuracy for a correctly predicted score, and ACC ± 1 represents the accuracy for a prediction error of one point. To the best of our knowledge, we propose the method for NFOG-Q score prediction in our study by machine vision to evaluate the subject’s FOG severity, which has great convenience and solves the problem of bias in evaluations caused by the subjectivity associated with self-evaluations.

The limitations of this paper are as follows:

(1) Since the included subjects did not experience FOG often during shooting, there were few FOG samples. The reasons are for this are mainly due to the characteristic of the sudden occurrence of FOG, filming environment (FOG is less likely to occur in a wide hospital corridor than in a home environment) (Snijders et al., 2008), test requirements in this study (for the sake of safety and feasibility of the study, the included PD-FOG subjects were in the “ON” state of medication during the test and were less prone to FOG), and the exclusion of more severe advanced PD subjects (i.e., those no longer in the “ON” state of medication or no longer able to stand and walk). Because the patients performed the test in the “ON” state, the “ON” state FOG has not been validated to determine whether it can be generalizable to the “OFF” state. A study found that FOG in the “OFF” state can persist in the full “ON” state after sufficient dopaminergic treatment, and the “ON” state FOG may be the same as the “OFF” state FOG but requires further verification (Lucas et al., 2019).

(2) Because the videos were shot in a real scene, some factors may have affected FOG recognition. For example, the patients had irregular motions during the test (e.g., a patient stood still for a short while due to a desire to rest). To optimize this problem, some gait parameters can be considered features to identify FOG, avoiding the influence of irregular motions, such as gait speed, cadence, and time to stop or freeze. In addition, factors such as shooting angle may have affected the effectiveness of the FOG recognition model. In our study, we asked subjects to walk toward the photographer, which eliminated the problem of the shooting angle. An affine transformation will be performed to adjust to various cameras in applications.

(3) Due to the perspective nature, the increase in the shooting distance leads to a decrease in the signal-to-noise ratio (SNR) of the keypoint position signals extracted by OpenPose. However, this relationship is acceptable. Since our purpose was to verify the feasibility of machine vision methods for FOG monitoring and assessment, as long as the results showed that the model had achieved relatively good performance and met clinical expectations, optimization of the SNR was not the focus of this study. Theoretically, if the SNR can be improved, better performance can be achieved. Therefore, the SNR can be optimized in the next step.

(4) Regarding the recognition of severe patients who needed support during the test, since OpenPose can recognize multiple skeletons, if the shooting condition was ignored in actual application, the skeleton of both the patient and the caregiver would be exposed together in the video, which would interfere with FOG recognition. In response to this problem, we will study other algorithms to assist the subjects’ skeleton recognition, thus improving information extraction effectiveness.

To further increase the representativeness and higher accuracy of our FOG recognition system, future studies should evaluate a larger sample of patients (e.g., covering residential communities and inpatient wards as well as in “OFF” state FOG). In addition, we did not examine the possible impact of fewer FOG phenotypes such as trembling, akinetic, or shuffling with the algorithms (Mazzetta et al., 2019). PD-FOG patients with many different phenotypes should be included in future studies to improve the practicality of this system. We can also combine wearable sensors to accurately detect suspected PD-FOG patients to improve the accuracy of FOG recognition.

As a new technology, FOG recognition based on machine vision has many challenges. For example, levodopa (L-dopa) and other drugs can improve space- and time-related gait parameters (such as step length and speed) (Smulders et al., 2016), thus causing a large difference between home evaluation and outpatient evaluation. In this regard, we will need to address differences in the assessment of patients under the influence of drugs to make objective assessments. If this problem can be solved, FOG recognition based on machine vision will play a key role in the identification of high-risk individuals and observation of curative effects.

Conclusion

Freezing of gait is one of the major motor symptoms of PD patients that can cause falls. Since PD patients are mostly elderly, falling results in fractures and even death. Thus, it is critical to identify methods to detect FOG and prevent patients from falling. In this study, a multi-stage FOG recognition model and the NFOG-Q score prediction model with good performance were developed. This brand-new diagnostic evaluation service not only solves the impact of individual subjectivity but also reduces the economic burden on PD patients. The current outbreak of the COVID-19 pandemic affects all aspects of healthcare. PD-FOG patients are a particularly vulnerable group who have been directly and indirectly affected by this pandemic, resulting in a lack of timely diagnosis and treatment. Fortunately, the implementation of the FOG evaluation system based on machine vision can overcome the above problems. In this way, full-process management and timely intervention in the PD-FOG population can be truly achieved.

Data Availability Statement

The original contributions presented in this study are included in the article/supplementary material, further inquiries can be directed to the corresponding author/s.

Ethics Statement

The studies involving human participants were reviewed and approved by the Scientific Ethics Committee of General Hospital of Southern Theater Command of PLA and the Ethics Committee of Guangzhou First People’s Hospital. The patients/participants provided their written informed consent to participate in this study.

Author Contributions

ML and HZ provided the sample. WL, JLu, HB, JLi, JW, HD, and GX collected data from the sample. WL, XC, JZ, and HZ drafted the manuscript. XC, WZ, CC, and ZC performed the statistical analyses. JZ, YL, KR, CZ, ML, ZC, and HZ designed and supervised the study and double checked the statistical analyses. All authors contributed to the conceptual design, collection, writing, editing, and generation of figures and tables for this manuscript and approved the submitted version.

Funding

This study was funded by the National 863 Program project (grant no. 2012AA02A514), National Key R&D Program of China (grant no. 2018YFC1312000) and Guangzhou Municipal Health Commission (grant no. 20201A011004).

Conflict of Interest

XC, YL, KR, WZ, CC, and ZC were employed by GYENNO SCIENCE Co., LTD.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We thank all the patients who participated in this study. We acknowledge the support of General Hospital of Southern Theatre Command of PLA, GYENNO SCIENCE Co., LTD., Guangzhou First People’s Hospital, Ruijin Hospital, Shanghai Jiao Tong University School of Medicine, Guangdong Second Provincial General Hospital and The 960th Hospital of the PLA Joint Logistics Support Force.

Abbreviations

PD, Parkinson’s disease; LOSO, leave one subject out; FOG, freezing of gait; TUG, timed up and go test; Narrow, narrow gap timed up and go test; Walk-FOG recognition model, FOG recognition model for the walking stage; Turn-FOG recognition model, FOG recognition model for the turning stage.

References

Aberg, A. C., Olsson, F., Ahman, H. B., Tarassova, O., Arndt, A., Giedraitis, V., et al. (2021). Extraction of gait parameters from marker-free video recordings of timed up-and-go tests: validity, inter- and intra-rater reliability. Gait Posture 90, 489–495. doi: 10.1016/j.gaitpost.2021.08.004

PubMed Abstract | CrossRef Full Text | Google Scholar

Allcock, L. M., Rowan, E. N., Steen, I. N., Wesnes, K., Kenny, R. A., and Burn, D. J. (2009). Impaired attention predicts falling in Parkinson’s disease. Parkinsonism Relat. Disord. 15, 110–115. doi: 10.1016/j.parkreldis.2008.03.010

PubMed Abstract | CrossRef Full Text | Google Scholar

Amini, A., Banitsas, K., and Young, W. R. (2019). Kinect4FOG: monitoring and improving mobility in people with Parkinson’s using a novel system incorporating the Microsoft Kinect v2. Disabil. Rehabil. Assist. Technol. 14, 566–573. doi: 10.1080/17483107.2018.1467975

PubMed Abstract | CrossRef Full Text | Google Scholar

Bächlin, M., Hausdorff, J. M., Roggen, D., Giladi, N., Plotnik, M., and Tröster, G. (2009). “Online detection of freezing of gait in Parkinson’s disease patients: a performance characterization,” in Proceedings of the 4th International ICST Conference on Body Area Networks, BODYNETS 2009, (Los Angeles, CA), 1–8. doi: 10.4108/ICST.BODYNETS2009.5852

CrossRef Full Text | Google Scholar

Bächlin, M., Plotnik, M., Roggen, D., Maidan, I., Hausdorff, J. M., Giladi, N., et al. (2010). Wearable assistant for Parkinson’s disease patients with the freezing of gait symptom. IEEE Trans. Inf. Technol. Biomed. 14, 436–446. doi: 10.1109/TITB.2009.2036165

PubMed Abstract | CrossRef Full Text | Google Scholar

Barry, E., Galvin, R., Keogh, C., Horgan, F., and Fahey, T. (2014). Is the timed up and go test a useful predictor of risk of falls in community dwelling older adults: a systematic review and meta-analysis. BMC Geriatr. 14:14. doi: 10.1186/1471-2318-14-14

PubMed Abstract | CrossRef Full Text | Google Scholar

Barthel, C., Mallia, E., Debu, B., Bloem, B. R., and Ferraye, M. U. (2016). The practicalities of assessing freezing of gait. J. Parkinsons Dis. 6, 667–674. doi: 10.3233/JPD-160927

PubMed Abstract | CrossRef Full Text | Google Scholar

Brognara, L., Palumbo, P., Grimm, B., and Palmerini, L. (2019). Assessing gait in Parkinson’s disease using wearable motion sensors: a systematic review. Diseases 7:18. doi: 10.3390/diseases7010018

PubMed Abstract | CrossRef Full Text | Google Scholar

Cao, Z., Hidalgo, G., Simon, T., Wei, S. E., and Sheikh, Y. (2021). OpenPose: realtime multi-person 2D pose estimation using part affinity fields. IEEE Trans. Pattern Anal. Mach. Intell. 43, 172–186. doi: 10.1109/TPAMI.2019.2929257

PubMed Abstract | CrossRef Full Text | Google Scholar

Chawla, N. V., Bowyer, K. W., Hall, L. O., and Kegelmeyer, W. P. (2002). SMOTE: synthetic minority over-sampling technique. J. Artif. Intell. Res. 16, 321–357.

Google Scholar

Chen, T., and Guestrin, C. (2016). “XGBoost: a scalable tree boosting system,” in Proceedings of the 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (San Francisco, CA: Association for Computing Machinery), 785–794.

Google Scholar

Das, S., Trutoiu, L., Murai, A., Alcindor, D., Oh, M., De la Torre, F., et al. (2011). Quantitative measurement of motor symptoms in Parkinson’s disease: a study with full-body motion capture data. Annu. Int. Conf. IEEE Eng. Med. Biol. Soc. 2011, 6789–6792. doi: 10.1109/IEMBS.2011.6091674

PubMed Abstract | CrossRef Full Text | Google Scholar

Djaldetti, R., Rigbi, A., Greenbaum, L., Reiner, J., and Lorberboym, M. (2018). Can early dopamine transporter imaging serve as a predictor of Parkinson’s disease progression and late motor complications? J. Neurol. Sci. 390, 255–260. doi: 10.1016/j.jns.2018.05.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Droby, A., Pelosin, E., Putzolu, M., Bommarito, G., Marchese, R., Mazzella, L., et al. (2021). A multimodal imaging approach demonstrates reduced midbrain functional network connectivity is associated with freezing of gait in Parkinson’s disease. Front. Neurol. 12:583593. doi: 10.3389/fneur.2021.583593

PubMed Abstract | CrossRef Full Text | Google Scholar

Earhart, G. M. (2013). Dynamic control of posture across locomotor tasks. Mov. Disord. 28, 1501–1508. doi: 10.1002/mds.25592

PubMed Abstract | CrossRef Full Text | Google Scholar

Folstein, M. F., Folstein, S. E., and McHugh, P. R. (1975). Mini-mental state”. A practical method for grading the cognitive state of patients for the clinician. J. Psychiatr. Res. 12, 189–198. doi: 10.1016/0022-3956(75)90026-6

CrossRef Full Text | Google Scholar

Giladi, N., Shabtai, H., Simon, E. S., Biran, S., Tal, J., and Korczyn, A. D. (2000). Construction of freezing of gait questionnaire for patients with Parkinsonism. Parkinsonism Relat. Disord. 6, 165–170. doi: 10.1016/s1353-8020(99)00062-0

CrossRef Full Text | Google Scholar

Giladi, N., Tal, J., Azulay, T., Rascol, O., Brooks, D. J., Melamed, E., et al. (2009). Validation of the freezing of gait questionnaire in patients with Parkinson’s disease. Mov. Disord. 24, 655–661. doi: 10.1002/mds.21745

PubMed Abstract | CrossRef Full Text | Google Scholar

Hall, J. M., Shine, J. M., O’Callaghan, C., Walton, C. C., Gilat, M., Naismith, S. L., et al. (2015). Freezing of gait and its associations in the early and advanced clinical motor stages of Parkinson’s disease: a cross-sectional study. J. Parkinsons Dis. 5, 881–891. doi: 10.3233/JPD-150581

PubMed Abstract | CrossRef Full Text | Google Scholar

Hoehn, M. M., and Yahr, M. D. (1967). Parkinsonism: onset, progression and mortality. Neurology 17, 427–442. doi: 10.1212/wnl.17.5.427

PubMed Abstract | CrossRef Full Text | Google Scholar

Hu, K., Wang, Z., Mei, S., Ehgoetz, M. K., Yao, T., Lewis, S. J. G., et al. (2020). Vision-Based freezing of gait detection with anatomic directed graph representation. IEEE J. Biomed. Health Inform. 24, 1215–1225. doi: 10.1109/JBHI.2019.2923209

PubMed Abstract | CrossRef Full Text | Google Scholar

Hu, K., Wang, Z., Wang, W., Martens, K., Wang, L., Tan, T., et al. (2019). Graph sequence recurrent neural network for vision-based freezing of gait detection. IEEE Trans. Image Process. 29, 1890–1901. doi: 10.1109/TIP.2019.2946469

PubMed Abstract | CrossRef Full Text | Google Scholar

Kim, R., Lee, J., Kim, Y., Kim, A., Jang, M., Kim, H. J., et al. (2018). Presynaptic striatal dopaminergic depletion predicts the later development of freezing of gait in de novo Parkinson’s disease: an analysis of the PPMI cohort. Parkinsonism Relat. Disord. 51, 49–54. doi: 10.1016/j.parkreldis.2018.02.047

PubMed Abstract | CrossRef Full Text | Google Scholar

Latt, M. D., Lord, S. R., Morris, J. G., and Fung, V. S. (2009). Clinical and physiological assessments for elucidating falls risk in Parkinson’s disease. Mov. Disord. 24, 1280–1289. doi: 10.1002/mds.22561

PubMed Abstract | CrossRef Full Text | Google Scholar

Leeflang, M. M., Rutjes, A. W., Reitsma, J. B., Hooft, L., and Bossuyt, P. M. (2013). Variation of a test’s sensitivity and specificity with disease prevalence. CMAJ 185, E537–E544. doi: 10.1503/cmaj.121286

PubMed Abstract | CrossRef Full Text | Google Scholar

Lucas, M. J., Goldstein, F. C., Sommerfeld, B., Bernhard, D., Perez, P. S., and Factor, S. A. (2019). Freezing of Gait can persist after an acute levodopa challenge in Parkinson’s disease. NPJ Parkinsons Dis. 5:25. doi: 10.1038/s41531-019-0099-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Matar, E., Shine, J. M., Gilat, M., Ehgoetz, M. K., Ward, P. B., Frank, M. J., et al. (2019). Identifying the neural correlates of doorway freezing in Parkinson’s disease. Hum. Brain Mapp. 40, 2055–2064. doi: 10.1002/hbm.24506

PubMed Abstract | CrossRef Full Text | Google Scholar

Matinolli, M., Korpelainen, J. T., Sotaniemi, K. A., Myllyla, V. V., and Korpelainen, R. (2011). Recurrent falls and mortality in Parkinson’s disease: a prospective two-year follow-up study. Acta Neurol. Scand. 123, 193–200. doi: 10.1111/j.1600-0404.2010.01386.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Mazilu, S., Calatroni, A., Gazit, E., Roggen, D., Hausdorff, J. M., and Tröster, G. (2013). “Feature learning for detection and prediction of freezing of gait in Parkinson’s disease,” in Machine Learning and Data Mining in Pattern Recognition. MLDM 2013. Lecture Notes in Computer Science, Vol. 7988, ed. P. Perner (Berlin: Springer), 144–158.

Google Scholar

Mazilu, S., Hardegger, M., Zhu, Z., Roggen, D., Tröster, G., Plotnik, M., et al. (2012). “Online detection of freezing of gait with smartphones and machine learning techniques,” in Proceedings of the 6th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth) and Workshops, San Diego, CA, 123–130.

Google Scholar

Mazzetta, I., Zampogna, A., Suppa, A., Gumiero, A., Pessione, M., and Irrera, F. (2019). Wearable sensors system for an improved analysis of freezing of gait in Parkinson’s disease using electromyography and inertial signals. Sensors 19:948. doi: 10.3390/s19040948

PubMed Abstract | CrossRef Full Text | Google Scholar

Moore, O., Peretz, C., and Giladi, N. (2007). Freezing of gait affects quality of life of peoples with Parkinson’s disease beyond its relationships with mobility and gait. Mov. Disord. 22, 2192–2195. doi: 10.1002/mds.21659

PubMed Abstract | CrossRef Full Text | Google Scholar

Moore, S. T., Yungher, D. A., Morris, T. R., Dilda, V., MacDougall, H. G., Shine, J. M., et al. (2013). Autonomous identification of freezing of gait in Parkinson’s disease from lower-body segmental accelerometry. J. Neuroeng. Rehabil. 10:19. doi: 10.1186/1743-0003-10-19

PubMed Abstract | CrossRef Full Text | Google Scholar

Nieuwboer, A., Rochester, L., Herman, T., Vandenberghe, W., Emil, G. E., Thomaes, T., et al. (2009). Reliability of the new freezing of gait questionnaire: agreement between patients with Parkinson’s disease and their carers. Gait Posture 30, 459–463. doi: 10.1016/j.gaitpost.2009.07.108

PubMed Abstract | CrossRef Full Text | Google Scholar

Nonnekes, J., Ruzicka, E., Nieuwboer, A., Hallett, M., Fasano, A., and Bloem, B. R. (2019). Compensation strategies for gait impairments in Parkinson disease: a review. JAMA Neurol. 76, 718–725. doi: 10.1001/jamaneurol.2019.0033

PubMed Abstract | CrossRef Full Text | Google Scholar

Nutt, J. G., Bloem, B. R., Giladi, N., Hallett, M., Horak, F. B., and Nieuwboer, A. (2011). Freezing of gait: moving forward on a mysterious clinical phenomenon. Lancet Neurol. 10, 734–744. doi: 10.1016/S1474-4422(11)70143-0

CrossRef Full Text | Google Scholar

Nyquist, H. (1928). Certain topics in telegraph transmission theory. Trans. Am. Inst. Electr. Eng. 47, 617–644. doi: 10.1109/T-AIEE.1928.5055024

CrossRef Full Text | Google Scholar

O’Day, J., Syrkin-Nikolau, J., Anidi, C., Kidzinski, L., Delp, S., and Bronte-Stewart, H. (2020). The turning and barrier course reveals gait parameters for detecting freezing of gait and measuring the efficacy of deep brain stimulation. PLoS One 15:e231984. doi: 10.1371/journal.pone.0231984

PubMed Abstract | CrossRef Full Text | Google Scholar

Paul, S. S., Sherrington, C., Canning, C. G., Fung, V. S., Close, J. C., and Lord, S. R. (2014). The relative contribution of physical and cognitive fall risk factors in people with Parkinson’s disease: a large prospective cohort study. Neurorehabil. Neural Repair 28, 282–290. doi: 10.1177/1545968313508470

PubMed Abstract | CrossRef Full Text | Google Scholar

Pelicioni, P., Menant, J. C., Latt, M. D., and Lord, S. R. (2019). Falls in Parkinson’s disease subtypes: risk factors, locations and circumstances. Int. J. Environ. Res. Public Health 16:2216. doi: 10.3390/ijerph16122216

PubMed Abstract | CrossRef Full Text | Google Scholar

Postuma, R. B., Berg, D., Stern, M., Poewe, W., Olanow, C. W., Oertel, W., et al. (2015). MDS clinical diagnostic criteria for Parkinson’s disease. Mov. Disord. 30, 1591–1601. doi: 10.1002/mds.26424

PubMed Abstract | CrossRef Full Text | Google Scholar

Renfei, S., Zhiyong, W., Kaylena, E. M., and Simon, L. (2018). “Convolutional 3D attention network for video based freezing of gait recognition,” in Proceedings of the 2018 Digital Image Computing: Techniques and Applications (DICTA) (Canberra: IEEE), 1–7. doi: 10.1109/DICTA.2018.8615791

CrossRef Full Text | Google Scholar

Rodriguez-Martin, D., Sama, A., Perez-Lopez, C., Catala, A., Moreno, A. J., Cabestany, J., et al. (2017). Home detection of freezing of gait using support vector machines through a single waist-worn triaxial accelerometer. PLoS One 12:e171764. doi: 10.1371/journal.pone.0171764

PubMed Abstract | CrossRef Full Text | Google Scholar

Sato, K., Nagashima, Y., Mano, T., Iwata, A., and Toda, T. (2019). Quantifying normal and Parkinsonian gait features from home movies: practical application of a deep learning-based 2D pose estimator. PLoS One 14:e223549. doi: 10.1371/journal.pone.0223549

PubMed Abstract | CrossRef Full Text | Google Scholar

Schaafsma, J. D., Balash, Y., Gurevich, T., Bartels, A. L., Hausdorff, J. M., and Giladi, N. (2003). Characterization of freezing of gait subtypes and the response of each to levodopa in Parkinson’s disease. Eur. J. Neurol. 10, 391–398. doi: 10.1046/j.1468-1331.2003.00611.x

PubMed Abstract | CrossRef Full Text | Google Scholar

Seuthe, J., Kuball, K., Hoffmann, A. K., Weisser, B., Deuschl, G., and Schlenstedt, C. (2021). Validation of the German version of the new freezing of gait questionnaire for people with Parkinson’s disease. Parkinsons Dis. 2021:8841679. doi: 10.1155/2021/8841679

PubMed Abstract | CrossRef Full Text | Google Scholar

Shannon, C. E. (1949). Communication in the presence of noise. Proc. IRE 37, 10–21. doi: 10.1109/JRPROC.1949.232969

CrossRef Full Text | Google Scholar

Shumway-Cook, A., Brauer, S., and Woollacott, M. (2000). Predicting the probability for falls in community-dwelling older adults using the timed up & go test. Phys. Ther. 80, 896–903.

Google Scholar

Sigcha, L., Costa, N., Pavon, I., Costa, S., Arezes, P., López, J. M., et al. (2020). Deep learning approaches for detecting freezing of gait in Parkinson’s disease patients through on-body acceleration sensors. Sensors 20:1895. doi: 10.3390/s20071895

PubMed Abstract | CrossRef Full Text | Google Scholar

Smulders, K., Dale, M. L., Carlson-Kuhta, P., Nutt, J. G., and Horak, F. B. (2016). Pharmacological treatment in Parkinson’s disease: effects on gait. Parkinsonism Relat. Disord. 31, 3–13. doi: 10.1016/j.parkreldis.2016.07.006

PubMed Abstract | CrossRef Full Text | Google Scholar

Snijders, A. H., Nijkrake, M. J., Bakker, M., Munneke, M., Wind, C., and Bloem, B. R. (2008). Clinimetrics of freezing of gait. Mov. Disord. 23(Suppl. 2), S468–S474. doi: 10.1002/mds.22144

PubMed Abstract | CrossRef Full Text | Google Scholar

Soltaninejad, S., Cheng, I., and Basu, A. (2019). Kin-FOG: automatic simulated freezing of gait (FOG) assessment system for Parkinson’s disease. Sensors 19:2416. doi: 10.3390/s19102416

PubMed Abstract | CrossRef Full Text | Google Scholar

Stebbins, G. T., Goetz, C. G., Burn, D. J., Jankovic, J., Khoo, T. K., and Tilley, B. C. (2013). How to identify tremor dominant and postural instability/gait difficulty groups with the movement disorder society unified Parkinson’s disease rating scale: comparison with the unified Parkinson’s disease rating scale. Mov. Disord. 28, 668–670. doi: 10.1002/mds.25383

PubMed Abstract | CrossRef Full Text | Google Scholar

Tang, Y. M., Wang, Y. H., Feng, X. Y., Zou, Q. S., Wang, Q., Ding, J., et al. (2022). Diagnostic value of a vision-based intelligent gait analyzer in screening for gait abnormalities. Gait Posture 91, 205–211. doi: 10.1016/j.gaitpost.2021.10.028

PubMed Abstract | CrossRef Full Text | Google Scholar

Toit, S., Steyn, A., and Stumpf, R. H. (1986). Graphical exploratory data analysis. J. Am. Stat. Assoc. 31:116

Google Scholar

Tripoliti, E. E., Tzallas, A. T., Tsipouras, M. G., Rigas, G., Bougia, P., Leontiou, M., et al. (2013). Automatic detection of freezing of gait events in patients with Parkinson’s disease. Comput. Methods Programs Biomed. 110, 12–26. doi: 10.1016/j.cmpb.2012.10.016

PubMed Abstract | CrossRef Full Text | Google Scholar

Viswakumar, A., Rajagopalan, V., Ray, T., and Parimi, C. (2019). “Human gait analysis using OpenPose,” in Proceedings of the Fifth International Conference on Image Information Processing (ICIIP), (Piscataway, NJ: IEEE), 310–314.

Google Scholar

Weiss, D., Schoellmann, A., Fox, M. D., Bohnen, N. I., Factor, S. A., Nieuwboer, A., et al. (2020). Freezing of gait: understanding the complexity of an enigmatic phenomenon. Brain 143, 14–30. doi: 10.1093/brain/awz314

PubMed Abstract | CrossRef Full Text | Google Scholar

Wood, B. H., Bilclough, J. A., Bowron, A., and Walker, R. W. (2002). Incidence and prediction of falls in Parkinson’s disease: a prospective multidisciplinary study. J. Neurol. Neurosurg. Psychiatry 72, 721–725. doi: 10.1136/jnnp.72.6.721

PubMed Abstract | CrossRef Full Text | Google Scholar

Zach, H., Janssen, A. M., Snijders, A. H., Delval, A., Ferraye, M. U., Auff, E., et al. (2015). Identifying freezing of gait in Parkinson’s disease during freezing provoking tasks using waist-mounted accelerometry. Parkinsonism Relat. Disord. 21, 1362–1366. doi: 10.1016/j.parkreldis.2015.09.051

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhang, H., Yin, X., Ouyang, Z., Chen, J., Zhou, S., Zhang, C., et al. (2016). A prospective study of freezing of gait with early Parkinson disease in Chinese patients. Medicine 95:e4056 doi: 10.1097/MD.0000000000004056

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: Parkinson’s disease, freezing of gait, XGBoost, machine vision, machine learning

Citation: Li W, Chen X, Zhang J, Lu J, Zhang C, Bai H, Liang J, Wang J, Du H, Xue G, Ling Y, Ren K, Zou W, Chen C, Li M, Chen Z and Zou H (2022) Recognition of Freezing of Gait in Parkinson’s Disease Based on Machine Vision. Front. Aging Neurosci. 14:921081. doi: 10.3389/fnagi.2022.921081

Received: 15 April 2022; Accepted: 21 June 2022;
Published: 14 July 2022.

Edited by:

Ramesh Kandimalla, Indian Institute of Chemical Technology (CSIR), India

Reviewed by:

J. Lucas McKay, Emory University, United States
Frederico Pieruccini-Faria, Western University, Canada

Copyright © 2022 Li, Chen, Zhang, Lu, Zhang, Bai, Liang, Wang, Du, Xue, Ling, Ren, Zou, Chen, Li, Chen and Zou. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Mengyan Li, ZG9jbG15QDE2My5jb20=; Zhonglue Chen, Y2hlbnpob25nbHVlQGd5ZW5uby5jb20=; Haiqiang Zou, ZHJfem91QDE2My5jb20=

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.