- 1Human Performance and Robotics Laboratory, Department of Mechanical and Aerospace Engineering, California State University Long Beach, Long Beach, CA, United States
- 2Department of Electrical Engineering, California State University Long Beach, Long Beach, CA, United States
- 3Department of Biomedical Engineering, California State University Long Beach, Long Beach, CA, United States
We performed an electronic database search of published works from 2012 to mid-2021 that focus on human gait studies and apply machine learning techniques. We identified six key applications of machine learning using gait data: 1) Gait analysis where analyzing techniques and certain biomechanical analysis factors are improved by utilizing artificial intelligence algorithms, 2) Health and Wellness, with applications in gait monitoring for abnormal gait detection, recognition of human activities, fall detection and sports performance, 3) Human Pose Tracking using one-person or multi-person tracking and localization systems such as OpenPose, Simultaneous Localization and Mapping (SLAM), etc., 4) Gait-based biometrics with applications in person identification, authentication, and re-identification as well as gender and age recognition 5) “Smart gait” applications ranging from smart socks, shoes, and other wearables to smart homes and smart retail stores that incorporate continuous monitoring and control systems and 6) Animation that reconstructs human motion utilizing gait data, simulation and machine learning techniques. Our goal is to provide a single broad-based survey of the applications of machine learning technology in gait analysis and identify future areas of potential study and growth. We discuss the machine learning techniques that have been used with a focus on the tasks they perform, the problems they attempt to solve, and the trade-offs they navigate.
1 Introduction
Smart Gait (SG) is a term for any integrated human gait data analysis system utilizing Artificial Intelligence (AI). It is a growing research field capitalizing on the advancements in modern sensing technologies, automation, cloud computing, data analytics, parallel processing, and Internet of Things (IoT).
Some of the most prominent tasks SG performs are gait phase detection, gait event prediction, human activity recognition, fall detection, recognition of a person’s age and gender, abnormal gait detection such as fatigued state, stroke and neurological disease (ND), Parkinson’s Disease (PD), estimation of joint angles and moments, the walking person’s intent recognition and trajectory prediction, human pose estimation, localization and mapping, person identification, re-identification and authentication, step counting, assessment of physical skill and mobility, balance assessment, fall risk assessment, gait modeling, and simulation. Often SG performs a combination of two or more tasks simultaneously.
SG systems can analyze single or multiple gaits simultaneously. The multi-gait SG performs tasks such as crowd and occupancy sensing, crowd behavior prediction, multi-gait recognition as in identifying a person walking with one or more others, generating multi-gait for animation and virtual environments, and detecting abnormal gait in crowds or indoors for security applications. SG is thus a smart tool in the toolbox of experts in many fields such as health and wellness, security, user privacy, forensics, enhanced user experience, animation, energy, wearables and related fields like insurance, longevity, geriatrics, workplace safety, and productivity. SG can also easily be integrated with other smart systems that utilize heart rate, audio, haptic, speech, etc., for an even wider reach across many applications and industries. As such, SG is a component of many smart devices, smart homes, stores, cities, and energy grids.
This work reviews most of the research in the field of smart gait in recent years. It provides a broad-based survey of the current state of the field and identifies future areas of potential study and growth. First, we performed an electronic database search on six well-known electronic libraries: IEEE Xplore, Science Direct, PubMed Central, Google Scholar, ACM, and Web of Science. We searched for “artificial intelligence” AND “human” AND “gait” in the full text (when available), title, keywords, and abstract. Due to the sheer volume of the work and based on the assumption that most of the more recent works build and improve upon previous work, we limited our search to 2012 to mid-year 2021. We started our work in early 2020 by reviewing papers published in 2012–2019, and updated our review with papers from 2020 and the first half of 2021 as we continued our work through the end of July 2021. All references were downloaded into EndNote. Figure 1 offers an overview of the number of references located by this original search and reflects the fast growth of SG research. Google Scholar, even though at first glance it shows the number of references in the 10,000 range, it allows downloads only up to 1,000 references. However, during our work, we came across other relevant papers that we included in this review. For instance, to identify all the work in studying fatigue by SG, we performed another search to include the word “fatigue” in our search.
Our search resulted in over 6,000 references. First, all papers were reviewed by title, and those not relevant to this study were excluded. These included papers with 1) no “gait” - such as research on upper limb movement 2) no “human” - such as research on animal and robotic gait, and 3) no AI - such as human gait studies that do not incorporate machine learning methods. After the exclusions by title, our database was reduced to about 3500 papers, which were then reviewed by title and abstract. The final selection criteria included at least one paper for each application, even smaller niche examples such as depression detection by gait (Fang et al., 2019). In areas where many papers existed for the same application, such as neurological disease gait detection, the most impactful papers were selected by looking at the overall influence of the paper by the number of citations on Google Scholar and the journal’s impact factor.
Additionally, 393 review papers were originally excluded from this study. None of the works in our database search presents a comprehensive review of the applications of AI in gait studies. These works fall in one of the three categories as below:
1) studies in related fields but not exactly within the inclusion criteria of this work, For instance, Najafi and Mishra (2021) compiled a narrative review on digital health technologies for diabetic foot syndrome. While ML algorithms are probably embedded in some of the technologies they discuss, the study does not mention or discuss artificial intelligence.
2) studies that partially overlap with our studies, such as the literature review on PD diagnosis by Mei et al. (2021), that in addition to PD diagnosis by gait, also discuss other modalities such as voice, handwriting, magnetic resonance imaging (MRI), etc.
3) studies that are in the scope of this review but only cover a specific topic such as human motion trajectory prediction (Rudenko et al., 2020), wearable sensing technologies for sports biomechanics (Taborri et al., 2020), self-powered sensors and systems (Wu et al., 2020), person re-Identification (Wang et al., 2016), (Nambiar et al., 2019), (Karanam et al., 2019), machine learning in soft robotics (Kim et al., 2021), ambient assisted living technologies (mostly AI-enabled and gait-related) (Cicirelli et al., 2021), human action recognition (Gurbuz and Amin, 2019), biomechanics (Halilaj et al., 2018), gait recognition (Kusakunniran, 2020), (Singh et al., 2018), (Wan C. et al., 2018), gait event detection and gait phase recognition (Prasanth et al., 2021), clinical gait diagnostics of knee osteoarthritis (Parisi et al., 2020), knee pathology assessment (Abid et al., 2019), data preprocessing in gait classification (Burdack et al., 2019), age estimation (Aderinola et al., 2021), and banchamrk datasets (Nunes et al., 2019). A survey paper by Alzubaidi et al. (2021) provides an overview of deep learning, with helpful definitions and a discussion of strengths, limitations, and future trends of various deep learning techniques. Similarly, Abiodun et al. (2019) review Artificial Neural Network (ANN) applications in pattern recognition. Our paper’s goal is to review applications of SG technology, in a wider, all-encompassing overview. The above-mentioned papers were not used in our study; the reader is encouraged to read them if they wish to go deeper into any of the discussed topics.
We identified six main applications of SG: 1) Gait analysis where analyzing techniques are improved through AI algorithms, 2) Health and Wellness, with applications in gait monitoring for abnormal gait detection, recognition of human activities, fall detection, and sports performance, 3) Human Pose Tracking covering one-person or multi-person tracking and localization systems, 4) Gait-based biometrics with applications in person identification, authentication, and re-identification as well as gender and race recognition 5) Smart Gait devices and environments ranging from smart socks, shoes, and other wearables, to smart homes and smart retail stores that incorporate continuous monitoring and control systems and 6) Animation that reconstructs human motion through gait modeling, simulation, and machine learning techniques. The categories are sometimes not easy to separate and overlap. For instance, gait phase detection algorithms could belong to Health and Wellness, SG Devices, or Gait Recognition categories.
The paper is organized in the following sections: Section 2) Smart Gait Vocabulary, sections 3–8 discuss the applications of AI in 3) Gait Analysis 4) Health and wellness 5) Tracking Human Pose 6) Gait Based Biometrics 7) Smart Gait devices and 8) Animation. In sections 9 and 10, we conclude with 9) Discussion and Future Trends and (10) Conclusions.
2 The Smart Gait Vocabulary
Here we introduce certain key definitions of essential terms commonly used in SG studies (See Table 1).
3 Gait Analysis
Machine learning techniques are successfully utilized to improve many aspects of gait analysis. In this collection of papers AI 1) helps with data aggregation and pre-processing, 2) works along with another AI to improve its performance, interpretability, or accuracy, and 3) classifies gait phases and predicts gait events.
3.1 Data Aggregation and Pre-processing Using ML Techniques
Deep learning models can automatically extract the optimal features directly from raw spatiotemporal gait data without requiring data preprocessing or engineering (end-to-end approach) (Costilla-Reyes et al., 2021), (Morbidoni et al., 2019), etc. In other cases, when conventional machine learning techniques are deployed, much work goes into feature extraction and selection to ensure that the input features explain most of the variance in the data and achieve good performance and high accuracy of the algorithm. Of the six commonly used data pre-processing techniques, Ground Reaction Force (GRF) filtering, time derivative, time normalization, data reduction, weight normalization, and data scaling, (Burdack et al., 2020) found that only GRF filtering and supervised data reduction techniques such as Principal Component Analysis (PCA) increased the performance of ML classifiers, with Random Forest (RF) being more robust in feature reduction than Support Vector Machines (SVM), Multi-Layer Perceptron (MLP) and Convolutional Neural Networks (CNN).
PCA is a dimension reduction technique that transforms the original feature space into a set of linearly uncorrelated variables called principal components (PCs). The first few PCs alone are usually enough to account for most of the variance in the data. For instance, the first 3–6 PCs alone accounted for 84–99% of the overall variance (Dolatabadi et al., 2017). Other variants of PCA have been suggested in the SG literature, such as kernel-based PCA (Semwal et al., 2015), multi-linear PCA that achieves sparse and discriminative tensor to vector projection (Zhang et al., 2015), or a combination of PCA, linear discriminant analysis (LDA) and other feature reduction techniques (Wu C. et al., 2019). (Phinyomark et al., 2015) attempt to evaluate the importance of intermediate to higher-order PCs in running biomechanics, finding that low order PCs (that account for up to 90% of the cumulative variance in the data) can be successfully used for age and gender recognition. Still, the more subtle running behavior patterns such as between-group variations in improvements after a 6-weeks rehabilitation program of runners with patellofemoral pain (PFP) can be captured by intermediate and high order PC’s (that explain 90–99% and 99–100% of the variance of the data correspondingly). PCA, some variation of PCA, or a combination of PCA and other linear dimension reduction techniques were used in most papers we reviewed. t-Distributed Stochastic Neighbor Embedding (t-SNE) was used by Costilla-Reyes et al. (2021) to achieve a 2D representation and visualization of different age clusters of the trained data. Besides feature reduction, AI is employed for data augmentation (Bhattacharya et al., 2020) or data engineering (Johnson et al., 2021).
Other considerations in carefully selecting the input features are avoiding overfitting (Zhang et al., 2020a), improving interpretability, especially in medical applications (Dindorf et al., 2020b), (Horst et al., 2019), reducing the energy expenditure of wearable sensors (Lan et al., 2020), (Russell et al., 2021), improving patient comfort (Di Nardo et al., 2020), fairness to people with disabilities (Trewin et al., 2019) and user experience (Kim et al., 2020). To achieve the highest possible accuracy, usually, more complex sensing technology is required. The goal is to apply feature reduction and selection to avoid redundant features such that only the significant features are extracted from the minimum sensing hardware (Khademi et al., 2019). Information Gain (IG) and not PCA is used to improve interpretability by preserving and ranking the original features (Onodera et al., 2017).
3.2 AI2AI
We define AI2AI as an AI-based procedure whose purpose is to make another AI better. Better AI is defined based on the task. As indicated above, it includes better performance, higher accuracy, better interpretability, robustness, lower cost, lower energy expenditure, reduced overall system complexity, unobtrusive sensing technology, and automatic and real-time processing. For instance, Zhu et al. (2020) use an Improved Artificial Fish Swarm Algorithm (IAFSA) to optimize the parameters of the RF algorithm for knee contact force (KCF) prediction. Bhattacharya et al. (2020) use a novel Conditional Variational Autoencoder (CVAE) trained on an annotated dataset of 4,227 human gaits recorded on video to generate thousands of new realistic annotated gaits. Their data augmentation generative network improves the accuracy of their novel Spatial-Temporal Graph Convolutional Network (ST-GCN) algorithm to classify the four human emotions: happy, sad, angry, and neutral by 6%. Lu et al. (2021b) apply transfer learning and domain adaptation to label the data for a cross-domain human activity recognition task.
Similarly, as a first phase to creating an activity recognition algorithm for construction workers (Kim and Cho, 2020), the authors first implemented five ML algorithms 32 times each, once for every possible combination of the number of Inertial Measurement Unit (IMU) sensors and their location in the body, and compared their performance by a cross-validation (CV) technique. This systematic approach of evaluating how many sensors are sufficient for activity recognition and where they should be placed in the construction workers’ body revealed that using two motion sensors located at a certain distance will achieve motion recognition performance like using all 17 motion sensors located throughout the entire body. Using the first phase results, the authors then proceed to build a long-short term memory (LSTM) algorithm for activity recognition that achieves 94.7% accuracy with just two sensors placed in the hip and neck.
Another example of AI2AI is using AI to label the training data for another AI. Labeling data manually often requires a lot of time and expertise, is expensive, requires an elaborate lab-like setup involving obtrusive sensing technology, and must comply with safety protocols. A computer vision algorithm trained on ImageNet (He et al., 2016) labels the data with sufficient accuracy for many applications such as sports biomechanics, training, and rehabilitation (Cronin et al., 2019). Similarly, a stream-based Active Learning (AL) algorithm minimizes data labeling effort (Loy et al., 2012).
3.3 Gait Phase Classification and Gait Event Prediction
Gait analysis evaluates a person’s walking pattern, which is seen as a sequence of gait cycles, where each gait cycle follows the movement of a single limb from heel-strike to heel-strike again (Gage et al., 1995). The two main gait phases are the stance phase and the swing phase. Depending on the reason for gait analysis, detecting just these two phases can be enough. That simplification permits less complex and cheaper gait analysis, which is desirable, especially in wearable systems (Di Nardo et al., 2020). A more common four-phase cycle includes initial contact, mid-stance, pre-swing, and swing (Jiang et al., 2018). The importance of AI in these studies is in facilitating real-time gait analysis, appreciated in many control devices like orthotics and prosthetics, rehabilitation monitoring, and fall detection systems for aging-in-place applications. (See Table 2).
4 Health and Wellness
Clinical gait analysis, though by itself not reliable for a definitive diagnosis of neurological disease or other impairment, often suggests a pathology if it detects a pattern different from a typical walking behavior considered the normal gait. Normal gait is a controlled, coordinated, and repetitive series of limb movements that advance the body in the desired direction with minimum energy expenditure (Gage et al., 1995). There are four reasons for performing a clinical gait analysis: diagnosis, assessment, monitoring, and prediction (Baker et al., 2016). In these applications, the AI can continuously monitor and learn data, look for patterns, classify human activities and detect the anomaly. If connected to a display, it is an excellent monitoring and assessment tool. AI also predicts a future gait event, in which case it can either alert a human operator such as a clinician, caretaker, or facility supervisor or, if integrated with a control device, activate an automatic response to prevent falls or injury. For instance, WeedGait passively monitors the gait of a person and then alerts them if they are at risk of Driving Under the Influence of Marijuana (DUIM) (Li et al., 2019), while an in-home rehabilitation system provides qualitative and quantitative feedback to post-stroke survivors (Lee et al., 2019).
We identified four major applications of ML gait analysis in health and wellness: 1) detecting abnormal gait due to a person’s condition or disease, 2) sports management, 3) fall detection, and 4) activity recognition.
4.1 Abnormal Gait
AI is well suited at learning patterns and detecting an anomaly in the data based on a pre-defined abnormal event (supervised learning) or a clustering algorithm (unsupervised learning), or a combination of the two. A very wide range of human diseases and conditions can affect the way a person walks such as Parkinson’s (Flagg et al., 2021), (Wahid et al., 2015), Huntington’s (Acosta-Escalante et al., 2018), ALS (Aich et al., 2018), idiopathic normal-pressure hydrocephalus (iNPH) (Ishikawa et al., 2019), ASD (Hasan et al., 2018) neuromuscular disease (Gotlin et al., 2018), pediatric hereditary spastic paraplegia (HSP) (Pulido-Valdeolivas et al., 2018), aging (Strath et al., 2015), (Costilla-Reyes et al., 2021), dementia (Kenney et al., 2018), (Arifoglu and Bouchachia, 2017), fatigue (Zhang J. et al., 2014), depression (Fang et al., 2019), anxiety (Zhao et al., 2019), emotional state (Bhattacharya et al., 2020), dual task, or walking while performing a cognitive task (Costilla-Reyes et al., 2021), knee osteoarthritis, (Kotti et al., 2017), stroke (PSH gait), (Cui et al., 2018), (Clark et al., 2015), diabetes (Sutkowska et al., 2019), COVID-19 (Maghded et al., 2020), inflammation (Lasselin et al., 2020), (Renner et al., 2021), level of physical activity (Renner et al., 2021), kidney disease (Yadollahpour et al., 2018), vertigo (Cao et al., 2021), sleep quality (Liu X. et al., 2019), Trendelenburg (Michalopoulos et al., 2016), arthritis (Karg et al., 2015), (Struss et al., 2018), idiopathic toe walking (Kim et al., 2019), drunkenness (Arnold et al., 2015) and influence of marijuana (Li et al., 2019). Thus, monitoring human gait can provide key insight into a person’s health. Gait data can serve as a biomarker by itself or in association with other biomarkers, demographic data, other measured or calculated body and health parameters (Andersson et al., 2014), and patient-generated health data (PGHD) (Jim et al., 2020).
The main role of ML techniques for this cluster of studies is in classifying 1) healthy control from pathology or 2) healthy control from multiple classes of functional gait disorders. Some frequently used words in these papers, such as unobtrusive, plug-and-play system, automatic, affordable, integrated into the home or clinic, adaptable, etc., shed light on the perceived values of these applied systems and where most of the research is focused.
The benefits of utilizing machine learning in abnormal gait detection for health applications are as follows:
1) Deep learning models can automatically extract the optimal features directly from raw spatiotemporal gait data without the need for data preprocessing or engineering. (Costilla-Reyes et al., 2021).
2) A combination of ML techniques can deploy simultaneously to perform more than one task automatically. For instance, ML can simultaneously detect pathological gait and automatically identify which body part has been affected the most by the disease (Dolatabadi et al., 2017).
3) ML can process a lot of data fast, including active online learning from streams of live and repository data (Flagg et al., 2021).
4) ML community can benefit from the collaborative effort. It fosters competition against established benchmarks and open research promoted by initiatives such as the Parkinson’s Disease Digital Biomarker (PDDB) DREAM Challenge (Zhang et al., 2020a)
5) ML-driven gait-based studies can benefit from large-scale population data through smartphones, smartwatches, and other wearable devices (Frey et al., 2019), as well as individual or small-scale data easily collected at low cost, via pervasive techniques such as MS Kinect (Dolatabadi et al., 2017), smartphone (Pepa et al., 2020), triboelectric nanogenerator (TENG) smart shoes (Zou et al., 2020) and socks (Zhang et al., 2020g) which makes them suitable for implementation at home, clinic, etc. These systems stand to benefit from the recent fast and continuing advancements in sensing technology, textiles, parallel processing, cloud computing, and IoT grids.
From a practical viewpoint, using the analogy of a human operator driving a car, some common issues with these studies are:
1) Can’t drive faster than sight: the problem of data labeling.
The accuracy of the algorithm is prone to human error during the training stage. For example, to label the training data in fatigue/non-fatigue states, subjective self-reported thresholds were used for the participants in a study (Zhang J. et al., 2014). Still, the participant’s perception of fatigue is not necessarily aligned with the physical changes and walking patterns due to fatigue (Baghdadi et al., 2021). Even in stricter clinical settings where experts manually label the data, there is the problem of changed gait due to the controlled environment. Researchers usually adopt measures to mitigate these issues to some degree by studying normal walking behavior versus treadmill walking (Morbidoni et al., 2019), discarding the first few gait strides from the walking test data, instructing participants to watch a smiley on the wall to get distracted from targeting the force plate at the expense of a natural walking behavior (Horst et al., 2019) or hiding the floor sensors under the walking surface (Slijepcevic et al., 2020).
There is a need to minimize the distance between the experimental setup and the real-life application of a system. For instance, the use of a home monitoring system for abnormal gait was studied (Guo et al., 2019), but the subjects used in the study were healthy subjects imitating abnormal gaits such as in-toeing, out-toeing, drop foot, supination, and pronation. To study the gait of dementia patients, an artificially generated dataset was proposed (Arifoglu and Bouchachia, 2017) where a dataset of normal ADL data was injected with instances of skipped or repeated activities during the day and sleep disruptions at night to mimic the abnormal activities that people with dementia (PwD) would manifest. A biped that imitates human motion was used to train the data for the recognition of anterior cruciate ligament (ACL) injury of the knee (Zeng et al., 2020). Often, studies are performed on normal gait subjects and cannot be extended to a pathological gait. Trewin et al. (2019) discuss considerations of fairness for people with disabilities and outline a few guideposts on problem scoping, data sourcing and pre-processing, AI model training, and deployment. They advocate inclusive, participatory, and value-sensitive design.
Similarly, there is also the problem of a small sample size. Given the clinical nature of these studies and the impaired gait participants they require, the barriers to experimentally collecting sufficient data are understandable. For instance, 19 healthy participants had to consent to intravenous injection of lipopolysaccharide to induce inflammation in a fatigued gait study and were paid 3500SEK each (Lasselin et al., 2020). In some scenarios, the number of input features is larger than the sample size (Zhang et al., 2015). In others, the sample size is too small to be representative enough to support any assertions fully and reduces the paper to the level of an exploratory effort (Baghdadi et al., 2021). In others, researchers generate synthetic data (Arifoglu and Bouchachia, 2017) that reflect features similar to the disease for an abnormal gait detection task; apply deep AL technique to reduce the number of required labels and consequently the time cost of manual labeling in a gait phase detection task. Finally, there is a developing trend to take measurements out of the lab (Russell et al., 2021) and into the subjects’ natural environments and implement deep learning techniques to do the labeling (Costilla-Reyes et al., 2021), (Cronin et al., 2019).
Many of these techniques in aggregating, pre-processing, and learning the data, often represent work-around strategies to cost, user privacy, and clinical constraints. They both simplify the systems and introduce some errors in them simultaneously.
2) Can’t see in the dark: The Black Box problem.
This refers to the low interpretability of complex AI systems which can pose a problem, especially in medical applications. Explainable AI (XAI) is a recent trend in AI research that attempts to address this concern and the related issues of transparency, trustworthiness, and clinical acceptance (Dindorf et al., 2020b), (Khodabandehloo et al., 2021). Interpretable deep gait is the first attempt to make deep learning gait analysis more interpretable using layer-wise relevance propagation (LPR) while still achieving high accuracy (Horst et al., 2019).
3) Can’t leave the parking lot: the research to commercialization gap and the need for government approval.
Medical devices that utilize data and ML techniques will need Food and Drug Administration (FDA) approval and general buy-in from medical professionals and their patients. Doctors, clinicians, therapists, carers, et al. will need to be willing and capable of embracing the newness of the technology. To date, the authors are not aware of any FDA-approved medical devices that utilize AI and gait data. Neurodegenerative diseases are not symptomatic until years after their onset, but clinical usefulness needs to be demonstrated for the approval of a medical device. The cost of developing, installing, and maintaining such systems also becomes a barrier to their commercialization and practical usefulness (König et al., 2015).
The most recent advancements seem to address some of these concerns, but those are only in the beginning phases, and the long-term implications on user safety and privacy, as well as their actual performance, remain to be proved.
4.1.1 Fatigued Gait
Fatigue is defined as “a lower level of strength, physical capacity, and performance” (Lu et al., 2017). Detecting the onset of fatigue and creating systems that manage the associated risks is an important part of production quality and human factors engineering in the workplace.
Measuring and analyzing gait for fatigue monitoring makes sense because 1) walking is a task that is possible to track via unintrusive technology suitable to workplace settings, such as video, wearable sensors, radar, and force plates on the floor or any combination of these. 2) Walking is a significant part of occupational tasks for workers in manufacturing, mining, construction, nursing, warehouse and distribution centers (Baghdadi et al., 2021). 3) in the context of advanced manufacturing (known as Industry 4.0) for example (but other modern occupational settings as well), in which a worker’s daily tasks involve interacting with automation, computing, and sensing technologies, the most relevant features extracted from readily available gait data can be pre-processed and analyzed through ML techniques, often real-time, which makes it possible to detect the onset of fatigue with accuracy and speed, at a relatively low cost. 4) Lastly, fatigue in the walking behavior is correlated to fatigue in other physical tasks, and while safety systems in the workplace should be custom-tailored to the relevant tasks, detecting fatigue in a worker’s gait could serve as an excellent general fatigue monitoring technique applicable to most industrial settings.
There are two main goals in fatigued gait studies: 1) intra-person recognition or continuous recognition of the person walking in different fatigue states 2) inter-person recognition or the recognition of fatigue in an individual. The first one answers the question: Can we still identify the person by their gait in a fatigued state? The second one answers the question: Can we recognize the onset of fatigue to avoid overtraining and injury in sports or improve worker safety in the workplace? Researchers conclude that a person’s gait pattern maintains its individuality even while manifesting situation-dependency, as is the case with fatigue (Janssen et al., 2011).
Only seven features extracted from one wearable sensor are needed for fatigue detection, with an average accuracy of greater than 0.85 (Sedighi Maman et al., 2020). Still, in the workplace, the individual fatigue detection accuracy of 0.85 may mean high misclassification rates across many subjects (Baghdadi et al., 2021), thus the authors suggest a multivariate hierarchical time series clustering algorithm using Dynamic Time Wrapping (DTW) as a dissimilarity measure. Detecting fatigue from smartphone sensors was suggested (Maghded et al., 2020) as part of a multi-modal sensing and machine learning framework to detect Covid-19 and predict its severity and outcome through an app on the user’s phone.
Overall, fatigue studies utilizing gait data and AI point to the importance of SG in managing fatigue in the workplace, sports performance management, rehabilitation exercises, reducing fall risk in the elderly, and finally, as part of an integrated system for overall health management, both at the individual level and in public health applications, as is the case with Covid-19 related studies. The fact that most studies we reviewed collected gait data from just one sensor or just the smartphone (Karvekar, 2019) shows that effort is already invested in making these systems unintrusive, cost-effective, and adaptable (see Table 3).
4.1.2 Neurological Disease
Gait-based detection and classification algorithms for disease diagnosis and monitoring are one of the major applications we saw, and Parkinsonian gait, with its many classifiable features, such as freezing of gait (FoG), shuffling steps, slow gait, gait asymmetry, etc., is the most prevalent disease in the studies (See Table 4). Researchers propose an automated, accurate, and sensor-free gait detection deep learning algorithm that depends on video recordings from pervasive devices such as smartphones, web cameras, and surveillance cameras as cheaper and more accessible alternatives to Vicon camera systems (Ajay et al., 2018). Procházka et al. (2015) move away from expensive, complex camera systems and recommend using MS Kinect image and depth sensors for synchronized data acquisition and spatial modeling of a moving person. The recommended Bayesian Classification (BC) algorithm distinguishes PD gait from healthy gait based on decision boundaries of three features: gait speed, stride length, and age, with an achieved accuracy of 94.1%. Wahid et al. (2015) contribute to the gait-based PD detection by suggesting that the spatial-temporal gait data be normalized first using multiple regression to account for the patient’s age, height, body mass, gender, and walking speed. (Wan S. et al., 2018) employ a deep MLP to analyze both movement and speech data captured through a smartphone and estimate the severity of PD. Ye et al. (2018) go a step further and propose a Neural Network (NN) combined with Fuzzy Logic (FL) approach that recognizes the gait of patients with neurodegenerative disease (ND) from normal gait. Since the motor function impairment in various NDs such as ALS, HD, and PD is caused by different factors, the particle swarm optimization (PSO) algorithm was used along with an adaptive neuro-fuzzy inference system (ANFIS) to classify the non-linear gait dynamics. Bilgin (2017) also attempts to distinguish ALS from other ND diseases and healthy patients using Discrete Wavelet Transform (DWT), LDA, and Naïve Bayesian Classifier (NBC).
One of the most encouraging recent developments in ND gait research using AI is the DREAM PDDB Challenge launched by Sage Bionetworks that promotes an open and competitive research infrastructure with large-scale data for developing digital signatures of PD. Similar challenges in computer vision, such as the ImageNet Large Scale Visual Recognition Challenge (ILSVRC) (Russakovsky et al., 2015), have shown to be conducive to tremendous growth with classification accuracy improving year after year against the previously established benchmark. The DREAM PDDB challenge utilizes the self-reported and sensor data collected through the mPower app (Bot et al., 2016) from 15,000 PD and healthy control (HC) subjects. The best performing team as of July 2021 reported Area Under the Receiver-Operating Characteristic Curve (AUROC) of 0.86 with a deep CNN algorithm that employs three spatial and temporal data augmentation techniques to deal with overfitting.
Another initiative, still in the early phases but very promising, is the Early Detection of Neurodegenerative Diseases (EDoN) project. Launched in February 2020, it is a global research initiative that has secured generous funding, prominent cross-disciplinary expertise, and UK government support (Wakefield, 2020). It is already in the first phase of collecting smartwatch data (gait data along with heart rate, sleep, navigation data, etc.) from volunteers in the Greater Boston area through a partnership with Boston University Alzheimer’s Disease Research Centre (BUADRC). The data will be used to generate a digital “fingerprint” for dementia which, in the application phase, will detect dementia 10–15 years earlier than the current clinical methods.
4.2 Sports
The general trend we see in sports and fitness applications of ML-based gait analysis systems is to create low-cost, adaptable, fast, easy, and scalable systems that reach the right balance of accuracy and comfort for the given application (See Table 5). In sports management and sports injury prevention, gait data is lower limb movements signals acquired from various sensing technologies: video, IMUs, force plates, etc. SG monitors walking and running activities but also water sports, basketball, and football.
A deep learning algorithm trained with a relatively small number of labeled images was able to predict the locations of joint markers with the same accuracy as a human labeler, thus providing a low-cost system for kinematic analysis with sufficient accuracy for applications in sports biomechanics, training, coaching, and rehabilitation (Cronin et al., 2019). A kinematic and kinetic analysis aided by ML techniques was carried out to identify the effects of the shoes on the biomechanics of running. The conclusion: changes in the midsole resilience are more subject-dependent, but the changes in the upper shoe structure seem to be more subject-independent (Onodera et al., 2017). A smart exercise mat unobtrusively recognizes which exercise the subject is performing and counts the number of repetitions. It is soft, cheap, and smart, and it has an accuracy like pedometers when it comes to monitoring strength-related exercises performed on the ground. This is a great alternative to the exercise mats athletes commonly use in the gym (Sundholm et al., 2014).
SG systems in sports applications are diverse, performing data engineering (Johnson et al., 2021) and data labeling (Cronin et al., 2019), evaluating the role of the shoe structure on running biomechanics (Onodera et al., 2017), monitoring fatigue to prevent injury (Russell et al., 2021), (Zhang J. et al., 2014), counting steps (Kang et al., 2018), assessing physical autonomy and functional ability (Khan et al., 2015), articulating real-time control of an electrical muscle stimulation (EMS) device for sports training (Hassan et al., 2017), predicting and preventing injury (Taborri et al., 2021), achieving multi-player tracking, identification, and re-identification (Zhang R. et al., 2020), classifying and counting different sports activities (Sundholm et al., 2014), and recognizing and analyzing sports behavior (Guo and Wang, 2021).
Researchers voice concern over the validity of the laboratory setting-based AI models versus real-world scenario-based models in sports. CNN and LSTM perform better than SVM (previously suggested in the literature) in the football shot and pass detection task in three scenarios closer to the real-world setting. The integrity of the collected data, selected features, and evaluation method must be reconsidered once AI systems are deployed in the real world (Stoeve et al., 2021). Estimating kinematic data (that would usually be collected in a lab, using force plates) from kinetic data that is easily measured in the field using IMU sensors is the focus of the study by (Johnson et al., 2021). Further studies will be needed to translate the research done in the lab to systems that can reliably and accurately deploy in their practical, real-world setting, especially for time-sensitive applications such as preventing sports injury in near real-time after detecting a potentially harmful event. Additionally, an area of potential growth in the future will be the application of sports monitoring, injury prevention, and training optimization for people with disabilities (Rum et al., 2021).
4.3 Fall Detection and Human Activity Recognition
Gait analysis is beneficial in monitoring the activities of daily living (ADL) in the elderly to improve the quality of their lives and health care in their homes and outside hospitals. HAR using wearables poses four main concerns, and often there are tradeoffs to navigate: 1) Energy considerations 2) Activity recognition accuracy 3) robustness over different users and different activities 4) user experience. These SG systems recognize several ADLs such as bending, squatting, walking, lying down, rolling out of bed, and the transitions between them to detect falls, minimize the false alarms on lying down versus falling events while trying to keep these systems low cost, automatic, adaptable, and unobtrusive. To the effect of the low-cost fall detection systems, (Ma et al., 2014b) propose to extract curvature scale space (CSS) features of human silhouette from video clips recorded with an inexpensive Kinect depth camera. They find that their Extreme Learning Machine (ELM) algorithm, combined with a variable-length PSO algorithm, performs no worse than state-of-the-art systems that depend on expensive, complex multi-camera systems. The performance of the algorithm is enhanced by using calibration techniques to address issues of misplaced or misaligned sensors (Yu et al., 2018). The detection time is essential for fall prevention systems, such that a control device has enough time to respond and prevent the fall (Mori et al., 2020). Researchers have been concerned with the false alarms in automatic fall detection systems, especially in activities such as lying in bed and falling. Chelli and Pätzold (2019) report 100% fall detection accuracy without any false alarms when implementing quadratic SVM and ensemble bagged tree (EBT) algorithms on acceleration and angular velocity data from two public datasets. In an earlier paper, Hakim et al. (2017) also reported near-flawless accuracies on their SVM classifier using smartphone data.
Gait data for these studies were collected from force sensors and three-axis accelerometers concealed under intelligent tiles (Daher et al., 2017), built-in smartphone IMU sensors (Hakim et al., 2017), wearable sEMG sensors (Xi et al., 2017), wearable motion sensors (Özdemir and Barshan, 2014), or from an integrated data collection system such as inertial sensor and Bluetooth nodes data captured on a smartphone (Santoyo-Ramón et al., 2018). Sensor data from both a smartwatch and a smartphone is shown to work better than either one individually (Weiss et al., 2019). The smart home prides itself in capturing data real-time and device-free, utilizing wi-fi enabled IoT platforms (Yang et al., 2018). Finally, radar data from continuous-wave radar systems can be used for activity recognition and fall detection providing an unobtrusive solution for data collection and posing no privacy concerns (Wu Q. et al., 2015), (Seyfioğlu et al., 2018). CapSense sensing technology uses KEH capacitor voltage traces to recognize among five different activities with an accuracy of over 90% (Lan et al., 2017). In subsequent work, the accuracy was improved by using two capacitors, one in the sole of a shoe and one in front (Lan et al., 2020). The overall system energy usage was also reduced by 75% since the conventional machine learning algorithms such as NB or KNN used accumulated voltage data as input, reducing the computational load of the system. Future work will make these sensors fully self-powered and battery-free and address activity recognition accuracy and user experience issues.
In summary, SG systems have performed with increasing accuracy, minimizing false alarms, thus providing accurate, low-cost, automatic activity recognition and fall detection systems (see Table 6).
5 Tracking Human Pose
Applications of gait analysis to indoor environments include indoor positioning and localization algorithms. Wang et al. (2015) designed four schemes for indoor positioning. They found that WiFi Pseudo-Odometry integration with a combination of topology-constrained KNN and a multi-threshold Pedestrian Dead Reckoning (PDR) algorithm achieves higher accuracy with a smaller number of particles when a floor map is used. Other researchers (Tariq et al., 2017) found that RF performs best out of all the Weka ML collection algorithms using capacitive sensors for indoor person localization. Robertson et al. (2013) achieve SLAM using distortions of the local magnetic field. A foot-mounted sensor, for instance, can serve to localize a moving person or robot and generate a map of the indoor space while exploiting odometry. Low-cost 2D LiDAR has been recommended to preserve user’s privacy in human tracking (Hasan et al., 2020) (See Table 7).
6 Gait Based Biometrics
Gait is a soft biometric feature enabling the identification of people by their gait. The individuality of the gait pattern persists over time (Horst et al., 2017) and many pathologies. The main applications of gait recognition are person identification, person re-identification, person authentication, gender recognition, age estimation (Dindorf et al., 2020a), occupancy sensing (Yang et al., 2018), crowd density estimation (Zhou et al., 2018), crowd monitoring and anomaly detection for video surveillance applications (Sun et al., 2017), and multi-player tracking and identification (Zhang R. et al., 2020). We will look into the first four categories in more detail in sections 6.1 through 6.4.
6.1 Person Identification
Human identification is the process of determining an individual’s identity. Methods include but are not limited to those based on vision, identification cards, and biological data. Human gait-based identification is the process of determining an individual’s identity by their distinctive walking style. Person identification and re-identification through gait recognition have a growing importance in security systems and video surveillance in public spaces such as airports, banks, shopping malls, etc., since gait provides a non-invasive biometric feature. The task includes identifying a subject from a camera and matching him/her to persons in the other cameras with non-overlapping fields of view, an operation known as “tag-and-track” (Wu et al., 2015c). These gait recognition systems face challenges due to variable parameters that influence the size and quality of video and image inputs, such as camera viewpoint, lighting, occlusion, image resolution, and the subjects’ dressing and carrying conditions. For this reason, most papers we reviewed tried to address one or more of these difficulties by improving previously studied and implemented algorithms. To validate their work, authors conduct experiments to analyze the performance of their algorithms against benchmarks using public datasets. For instance, when CNN was first proposed for gait recognition (Wu et al., 2017), the authors evaluated their proposed algorithm on the CASIA-B (Yu et al., 2006), OU-ISIR (Iwama et al., 2012), and USF (Sarkar et al., 2005).
Most person identification SG systems deal with the case when gait data is extracted from video. These systems are model-based (Hu et al., 2012), (An et al., 2020), or appearance-based (Guan et al., 2012; Zhang W. et al., 2019), also called model-free. Gait Energy Image (GEI) is an average of all the human body silhouette images in one gait cycle (Han and Bhanu, 2005). It is widely used in appearance-based person identification algorithms because it allows for a simple representation of gait data, removing noise while preserving relevant gait information. It strikes a good balance between reducing computational cost and maintaining a good gait recognition rate. On the other hand, GEI is sensitive to appearance variations such as camera view angles and whether a subject is wearing a coat or carrying a bag, and it loses temporal information. There have been efforts to replace GEI with a representation that preserves more of the temporal gait information such as Chrono Gait Image (CGI) (Wang et al., 2011) or to fuse GEI features with temporal features (Liu et al., 2018). Chao et al. (2019) proposed GaitSet, which uses unordered sets of equally sized gait silhouettes as the input to a CNN architecture with set pooling, providing an effective way to preserve spatial and temporal information without the sequential constraints. argued that not all body parts contain discriminative information for gait recognition tasks, and they proposed GaitPart, a part-based and micro-motion model that preserves only the relevant part-dependent spatial features and the local short-range temporal information. An et al. (2020) acknowledges the challenges of appearance-based models and proposes a pose-based model approach for gait recognition. The gait poses are extracted from video using deep learning techniques AlphaPose (Fang et al., 2017) and OpenPose (Cao et al., 2017). Tracking the 3-D pose of walking pedestrians in video surveillance systems in cases where multiple people move together and cast a shadow or cause occlusion has also been attempted (Rogez et al., 2014).
Besides video, gait data can come from depth sensors, force plates, radar, Wi-Fi-enabled IoT devices, and IMU sensors. Radar-Id, a radar-based human identification algorithm (Cao et al., 2018), employs a CNN with architecture similar to that of AlexNet (Krizhevsky et al., 2012) to learn the necessary features from raw micro-Doppler spectrograms directly without a need to explicitly design the features. This algorithm has excellent anti-noise performance, and it can identify one person amid up to 20 other people. Similarly, Vandersmissen et al. (2018) use radar device data for indoor person identification and intruder detection. Costilla-Reyes et al. (2019) propose a footstep recognition system that can differentiate between the legitimate users (clients) and the impostor users of a biometric system from sensors on the floor. Biometric recognition by gait makes it possible to identify intruders since gait is difficult to fake. Deep CNN architectures that utilize footstep representations extracted from GRFs serve as automatic continuous person identification and verification systems with applications in security and anomaly detection at airports, workplace environments, and smart homes. Zhong and Deng (2014) propose a gait representation using accelerometer and gyroscope data invariant to sensor orientation. Haque et al. (2016) present a method for human identification when given only depth images. Zou et al. (2018) propose Auto-ID, a human identification system that collects CSI measurements data from WiFi-enabled IoT devices referred to as shapelet “signatures” of human identification. Yang et al. (2018) also use the CSI curve of the human body for occupancy sensing and activity recognition. This system is good for anomaly detection, such as identifying intruders in a smart home automatically. A person walking in a smart space equipped with RFID devices affects the radio frequency (RF) signals. The effect can be captured by Received Signal Strength Indicator (RSSI) and the phase of the RF signal and then used for person identification. A system that employs TSNet, a tag selection deep reinforcement learning algorithm, PCA for feature reduction, and an attention-based LSTM algorithm performs RFID-based gait recognition that can easily be integrated into a smart home or smart office environment (Luo et al., 2020).
Person identification studies have benefitted from the developments in computer vision, established benchmarks, and public datasets. The code proposed in the studies is often made public, encouraging continued and collaborative research (Chao et al., 2019). Recent studies employ deep learning approaches that extract gait representations directly from raw gait data and learn discriminative features for human identification (see Table 8)
6.2 Person Re-Identification
Human reidentification is the task of identifying images of the same person from non-overlapping camera views at different times and locations. A re-identification problem has three major components: identifying which human parts should be compared, constructing invariant features to represent those parts, and computing an appropriate similarity metric between them. (Saghafi et al., 2014). SG as a soft biometric feature allows continuous tracking and behavior analysis of a person over a large camera network for forensics, surveillance, and security applications. Traditional ReID methods usually focused on building robust feature representations of the gait and estimated the similarity between a probe and gallery image by calculating their Euclidian distances. This method faces challenges in cross-view and cross-walking conditions, such as when the gait pair is in different camera viewpoints and carrying and dressing conditions (Wu et al., 2017). Wu et al. (2015c) proposed that current classifiers be enhanced with a combination of Pose Prior (PP) algorithm and subject-discriminative feature selection algorithm to construct a view-invariant ReID system. One solution is to view the ReID task as a link probability prediction problem where each person represents an instance node in a graph structure. The ReID algorithm computes the likelihood of the link between the two (Liu H. et al., 2021). Another solution, suggested by Chtourou et al. (2021), includes an offline and an online phase. During the offline phase, an optimized GEI feature representation is constructed combining a dynamic selection of most relevant parts and a transformation of the probe or the gallery image, so the two of them have the same view before a matching score is calculated. This offline phase serves to train the Part View Transformation Model (PVTM), which will be used online to transform the gallery image to the same view as the probe image before classification.
ReID algorithms involve feature learning and metric learning. They learn gait features attributable to a person and then learn a similarity measure which should be greater if a gait pair belongs to different people than when it belongs to the same person. Song et al. (2018) introduced binary segmentation masks and region-level contrastive learning. Joint feature learning and similarity measure learning have also been attempted to perform both tasks well. The algorithm simultaneously extracts local convolutional features and enhances the discrimination capability by focusing only on distinct regions when looking for similarities between videos. It jointly learns features and similarity values for a pair or triplet of values (Wu L. et al., 2019). A deep Siamese attention architecture that consists of a conventional GRU and an attention mechanism can learn spatiotemporal representations and similarity metrics and learn to discriminate which local spatial representations are relevant (Wu L. et al., 2019). With large datasets of gait images, one faces the issue of multiple pedestrians having similar appearances. Chen et al. (2015) observed that the similarity metric was larger for images of the same pedestrian than for two different pedestrians Multiple deep metric learning utilizing multiple stacked auto-encoder networks and classification networks has been used to characterize different pedestrian images belonging to the same person based on multiple similarity probabilities (Xiong et al., 2019). These deep learning networks have integrated the feature learning and dissimilarity learning tasks of the traditional ReID systems into a unified deep neural network that learns representations robust to variations in image quality, background clutter, camera viewpoint and subjects’ carrying and dressing conditions directly from raw gait images at pixel level. (See Table 9)
6.3 Person Authentication
Proper authentication includes user authentication: whether the user has authorized access, and person identification: who the current user is (Liang et al., 2020). Sensor-based user authentication uses biological features categorized in physical, physiological, and behavioral (Hernández-Álvarez et al., 2021). Gait provides behavioral biometric-based authentication, which by comparison with knowledge-based (passwords, personal identification number (PIN)) and physiological biometric-based (face recognition, fingerprint) is unobtrusive, continuous, less prone to attacks, and easily tracked through wearable devices, videos, and smartphones in the context of IoT environments. Gait-based authentication is studied either by itself (Qin et al., 2019) or as part of an authentication system that uses multiple modalities (Hintze et al., 2019) (Acien et al., 2019) (See Table 10). Authors (Acien et al., 2019) showed that the fusion with behavioral data improves the authentication system results.
6.4 Gender Recognition
SG performs gender recognition reliably and unobtrusively (See Table 11). The gender classification task is usually conducted alongside human identification, and human re-identification tasks (Lu et al., 2014) since a person’s gender can serve as a soft feature in identifying a person. For instance, gender classification is done first to prune a subset of subjects before human identification is performed (Castro et al., 2017), (Meena and Sarawadekar, 2020). Gender classification is also done first to improve the accuracy of a subsequent age estimation algorithm (Zhang S. et al., 2019). Gender and age classification, in certain commercial and electronic consumer applications specifically, can be sufficient to enhance user experience (Duan et al., 2018).
The authors have explored different input features for this task. Joint Swing Energy (JSE) is a static feature extracted from the skeleton, namely the distance of the body joints from anatomical planes. It can easily be extracted from gait data and performs well with various classifiers to recognize someone’s gender (Kwon and Lee, 2021). Histogram of Gradient (HG) method reduces the three-dimensional (3D) accelerometer and gyroscope data from smartphones into 1D temporal descriptors, used as input for a bootstrapped DT algorithm (Jain and Kanhangad, 2018). Small walking speed variations do not affect the classification accuracy, but larger variations do, suggesting that spatial features are probably better suited for gender recognition tasks using conventional classifiers. Authors (Wazzeh et al., 2019) suggested that extracting features that are invariant to walking speeds variations could improve the performance of their gender recognition algorithm. Castro et al. (2017) proposed a CNN-based end-to-end approach that uses optical flow maps extracted from very low-resolution video to represent each person by their gait signature and recognize their gender and identity with high accuracy. A multi-task CNN setup, where the deep network learns multiple attributes simultaneously, improves the accuracy further (Zhang S. et al., 2019).
Some authors focus on recognizing gender from gait data when subjects walk in different directions (Lu et al., 2014), at different walking speeds (Jain and Kanhangad, 2018), carrying a bag, wearing a coat (El-Alfy and Binsaadoon, 2019), etc. SG for gender recognition is non-intrusive, does not require the subject to cooperate, and has better performance invariant to carrying and clothing conditions, even with the low-resolution quality of videos.
7 Smart Gait Devices and Environments
This section reviews systems that incorporate ML, IoT, and advanced sensing and textile technologies for automatic, real-time gait data processing to perform an intelligent task with health, sports, entertainment, and security applications.
7.1 Smart Gait Devices
The domain of smart gait devices and environments is exciting, brave, creative, extensive, and ever-growing (See Table 12). SG devices include wearable shoes (Zou et al., 2020), socks (Zhang et al., 2020f), kneepads and anklets (Totaro et al., 2017), insoles (Low et al., 2020), as well as devices attached to the body, such as smartphones (Poniszewska-Maranda et al., 2019), smartwatches (San-Segundo et al., 2018), (Sigcha et al., 2021), etc., implantable medical devices such as ActiGait (Sturma et al., 2019), wearable robotics (Shi et al., 2019) such as prosthetics (Gao et al., 2020) orthotics (Zhang et al., 2020e), (Choo et al., 2021), assistive devices such as smart walkers (Jimenez et al., 2018), and environmental devices such as smart tiles (Daher et al., 2017). SG devices use gait data to facilitate health monitoring, including passive mental health assessment (Rabbi et al., 2011) and transfer data to control devices for health, sports, security, and entertainment applications. For instance, UbiHeld (Ubiquitous Healthcare for Elderly) incorporates gait and other data from a smartphone as well as additional data from an inexpensive Kinect camera to keep a status of the overall health, location, and activities of the elderly at home (Ghose et al., 2013).
The requirements for soft smart wearable garments can be conflicting: stretchable, entirely conformable to the body, designed ergonomically and esthetically pleasing, small size and weight, flexible, washable, robust, unobtrusive, reliable, and durable (Yang and Yin, 2021). Person identification is an integral part of smart gait devices due to the need for customized user experience, which introduces the need to preserve users’ privacy. Additionally, continuous, seamless user authentication is vital to prevent malicious attacks on users’ medical records without the burden of frequently entering a personal identification number (PIN) (Xu et al., 2021). Finally, other requirements include water-proof capability, mechanical durability, and connectivity with other smart devices and environments depending on the specific use (Zou et al., 2020). To accommodate these requirements, the main focus of the research is on sensing technology. Sensors for wearables are IMUs, capacitive sensors (Lan et al., 2020), KEH (Xu et al., 2017; Xu et al., 2018), (Ma et al., 2018), solar cells (Sandhu et al., 2021), resistive sensors, stretchable conductive micro fluids (Low et al., 2020), and TENGs (Zhang et al., 2020g).
There is generally a trade-off between the complexity of sensing technology in smart gait devices and the accuracy of the intelligent task they perform (Khademi et al., 2019). Depending on the task and granted sufficient accuracy, one can stop the chase for performance and focus on patient comfort (Di Nardo et al., 2020). A multi-object optimization (MOO) technique is implemented (Khademi et al., 2019) to navigate this trade-off: it selects an optimal feature subset that maximizes accuracy while minimizing sensing hardware. Fusing different bio-signals in hBCI systems for gait applications improves classification accuracy and the number of control commands. Still, it introduces the problem of channel configuration, information transfer rate, and temporal synchronization between the modalities (Khan et al., 2021).
SG devices will be an integral part of smart cities, smart buildings, smart homes, smart transportation, smart factories, energy grids, and e-Healthcare and are poised for tremendous growth in the future as IoT in the 5G framework will facilitate better and faster inter-connectivity between the devices (Chettri and Bera, 2019), end-to-end deep learning algorithms will provide real-time intelligence, and sensing technology will be embedded more comfortably into our everyday objects and clothing.
7.2 Smart Homes
A smart home environment is defined as one that acquires and applies knowledge about its residents and their physical surroundings to improve their experience in that setting (See Table 13). The smart home sensors, the wearable sensors, and the classifying algorithms in the CASAS smart home (Cook et al., 2015) serve to perform health monitoring, early detection of disease, health care, and treatment. The smart home integrates wearable sensing technology, AI technology, and sensor fusion technology to automatically control home appliances via a gesture recognition algorithm, to turn lights on and off via an indoor positioning algorithm, and to set the alarm off via a fire detection algorithm (Hsu et al., 2017). With a focus on XAI, a numerical score of anomaly level is output on the monitor along with natural language explanations (Khodabandehloo et al., 2021). The privacy and security of health data are major concerns of smart homes. With the recent advancements in blockchain technology and its wider acceptance, a possible solution is to build smart blockchain networks to securely store and share health data (Taralunga and Florea, 2021) (Cernian et al., 2020).
Of interest in this discussion is the project HABITAT (Home Assistance Based on the Internet of Things for the Autonomy of Everybody) that had only four key AI applications at the time this paper was written: an indoor localization system, smart armchair, smart belt, and a wall panel and mobile devices as the user interface (Borelli et al., 2019). The smart home embeds smart objects into objects of everyday life. The belt, which assesses body movement, is of interest to this study since it captures postural transitions and gait biometric data. The AI modules follow the Event Calculus (EC) modelling approach, which makes it easy to define the properties of the system. The system is expandable, more smart devices can be incorporated into it, and is an example of what is possible when activities and objects of our daily lives become “smart.” In summary, AI algorithms take care of the “smart” part of the home. Instead of being a passive recipient of care, the patient becomes an active agent of an intelligent health ecosystem (Najafi and Mishra, 2021).
SG extends beyond the context of smart home and will play a major role in the future smart cities (Lozano Domínguez and Mateo Sanguino Tde, 2021), smart factories (Pech et al., 2021), smart retail stores (Zhang et al., 2019a), smart rehabilitation labs (Sessoms et al., 2015), and smart devices (Lozano Domínguez and Mateo Sanguino Tde, 2021)
8 Animation and Virtual Environments
For this review, we consider studies requiring and using gait data and implementing an ML algorithm to create animated characters for movies, video games, and virtual reality environments. These studies include research in gait modeling, motion reconstruction, and character control. (See Table 14)
In Virtual Reality (VR) applications, the virtual character is controlled real-time by the user’s behavior. The task of the AI algorithm is to correctly recognize the user’s gait phases, with no delay, for a good real-time visual representation of the avatar’s motion. In video games, deep reinforcement learning is the algorithm of choice where the policy controls the action at each step. The policy is defined carefully to positively reward the desired action at each step, such as maintaining balance, tracking pose and orientation, alignment, mimicking the reference motion, surviving perturbations, and negatively rewarding actions such as falls. The avatars must be real-looking, preserving both the motion extracted from the skeleton features and the shape (Loper et al., 2014), (Huang et al., 2015).
Responsiveness to user demand, quality of visual representation, robustness to different walking styles and terrains, system runtime performance, adaptation to disturbances, maintaining balance, and retargeting to different morphologies are some of the key goals of these systems. Usually, motion data is extracted from expensive high-quality multi-camera motion capture systems. Authors have studied the possibility of retrieving high-quality representations from low-cost video clips recorded with pervasive monocular videos such as YouTube clips (Peng et al., 2018b) and reconstructing human pose from inertial measurements, in the case when direct-line-of-sight camera recording is not available due to occlusion for instance (Huang et al., 2018).
9 Discussion and Future Trends
In this study, we reviewed different applications of the smart gait with a focus on the various tasks artificial intelligence algorithms perform across many industries and disciplines, such as health and wellness, security, forensics, and energy management. We further identify four emerging trends in the SG research: 1) population-wide scale health data will be available and empower end-to-end automatic, ubiquitous, and continuous deep learning approaches for big data-driven intelligent systems, 2) Fast-growing innovations in other technologies such as cloud computing, smart textiles, blockchain, and 5G will offer new opportunities and pose new challenges for SG systems and demand fast congruent growth, 3) SG systems will need to address concerns about user privacy, safety, comfort, and experience captured by the paradigm, “human-in-the-loop” (Sedighi Maman et al., 2020); these will be enforced by human rights advocates and regulatory bodies, and 4) the need for AI in health applications to benefit from the fast, low-cost, and accuracy of “black box” intelligent systems while still being transparent and understandable, as captured by the paradigm of XAI. SG is a valuable tool in kinetic and kinematic analysis, disease monitoring, diagnosis, and rehabilitation, sports performance, fall risk assessment, detection and prevention, gait-based biometrics for person identification, re-identification and continuous automatic authentication, age and gender recognition, physical skill and mobility assessment, fitness tracking, gait modeling and simulation, crowd monitoring and anomaly detection, human pose estimation, indoor tracking, and localization. SG is often integrated with other smart systems that utilize biomedical signals such as electrocardiograms, body parameters such as temperature, blood pressure, respiration rate, energy expenditure, and heart rate, and environmental signals such as room temperature and humidity. SG is often part of a smart IoT framework embedded with sensors, monitoring devices, and AI-enabled actuators that are all connected and in continuous communication. SG is everywhere, in our smart devices, smart homes, classrooms, cars, stores, cities, and energy grids. Smart Gait research will continue to grow fast in the future and will benefit from advancements in other technologies such as sensors, blockchain, IoT, textiles, 5G, cloud computing, and big data. These new technologies will also pose new demands and offer new opportunities for smart gait research. SG research will continue to address user privacy issues, security of health data, patient comfort, worker safety in the workplace, enhanced user experience, fatigue monitoring and injury prevention in sports.
We also identify these three needs for the SG research community: 1) inter-disciplinarity 2) the need for SG to become an organized field of study with specific definitions and good practices in place, and 3) the need for open competition and collaboration. Looking into the future, digital technologies such as smartwatches, headbands, and smartphones will be exploited to collect population-wide scale data to facilitate health monitoring and early detection of various diseases. One aspect of such systems is the multimodality of the data (Frey et al., 2019), (Abeysekara et al., 2020). Their integrated approach will demand cross-disciplinary research and collaborations. The gait research teams will need to include experts across many disciplines. We see the collaborative culture becoming more prevalent in the future both, within the SG community and between the SG community and the larger research community across many fields. With the recent development of programming languages and the open research community around them on Github, Kaggle, and other online and app-based forums, it is possible and important that the Smart Gait research community is open and the datasets, AI code, and suggested strategies for improvement are available for future collaborative work. For instance, some authors have made the data and the ML toolbox public and available for other researchers (Baghdadi et al., 2021) and (Horst et al., 2019). Wherever such open and collaborative efforts have flourished in the past, the results have been outstanding. Computer vision studies have seen tremendous results by reducing the cost of entry to new research and encouraging collaboration and competition towards a goal. The authors of this paper are excited to see similar growth in the SG research community.
10 Conclusion
The utilization of AI in gait analysis is a growing field. In this paper, we refer to it as the Smart Gait. It is compellingly multi-disciplinary, drawing from cutting-edge research in multiple mathematical and engineering fields, and it continues to welcome new approaches and new data-capturing methods. With the constant advances in wearable sensors technology, cloud computing, and new advances in ML, the progress is formidable. Coupled with the growing realization of real-world applications such as non-invasive person identification, person re-identification, intruder detection, medical diagnosis and treatment, advanced fall detection, etc., will keep the demand for new gait detection and analysis methods high and keep AI at the forefront of the research. The field should only grow and expand in scope over the next 10–15 years. From smart home devices to smart grids and smart cities, AI is here to stay, and it will become very pervasive. AI-driven gait-based systems will take the shape of a chair you sit on, shoes you wear, a mat you exercise on. The smart door of your home will open as it recognizes your walk via a continuous SG authentication system, and the phone will lock itself in the hands of the thief as the abnormal gait is recognized. A crowd density algorithm will warn you of high Covid-19 risk when you enter a store.
With the growing demand and promising results, accessibility remains a limitation. ML is a tool, and as such, it should be in the hands of those who need it; early detection of PD in the hands of the clinician, monitoring of treatment in the hands of the caregiver, occupancy sensing in the hands of a family who cares about energy management in their home, indoor tracking and localization in the hands of the police in your town police station, kinematic analysis in the hands of a competing athlete. For ML systems to be more accessible, the technology will need to become easy to understand and implement. It must become less expensive and more scalable. XAI is a trend that will continue. The experts in related fields and the general population will continue to become more AI savvy. In a not-too-distant future, knowing how to use an open-source Python library or write your line or two of code in R will be as common a task as writing an email today. With that will come privacy concerns, ethical concerns, and a need to adjust our laws and regulations as a society.
Author Contributions
EH conducted the systematic review of the papers. I-HK helped with the review of the papers and the writing of the manuscript. ED planned the systematic review, helped with the review of the papers, and contributed to the writing.
Funding
The work has been supported by the California State University Small Faculty Grant.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Abeysekara, L. L., Tran, H., Pathirana, P. N., Horne, M., Power, L., and Szmulewicz, D. (2020). “Multi-domain Data Capture and Cloud Buffered Multimodal Evaluation Platform for Clinical Assessment of Cerebellar Ataxia,” in 2020 42nd Annual International Conference of the IEEE Engineering in Medicine & Biology Society (EMBC), 5640–5643.
Abid, M., Mezghani, N., and Mitiche, A. (2019). Knee Joint Biomechanical Gait Data Classification for Knee Pathology Assessment: A Literature Review. Appl. bionics Biomech. 2019, 7472039. doi:10.1155/2019/7472039
Abiodun, O. I., Kiru, M. U., Jantan, A., Omolara, A. E., Dada, K. V., Umar, A. M., et al. (2019). Comprehensive Review of Artificial Neural Network Applications to Pattern Recognition. IEEE Access 7, 158820–158846. doi:10.1109/access.2019.2945545
Acien, A., Morales, A., Vera-Rodriguez, R., Fierrez, J., and Tolosana, R. (2019). “MultiLock: Mobile Active Authentication Based on Multiple Biometric and Behavioral Patterns,” in 1st International Workshop on Multimodal Understanding and Learning for Embodied Applications (Nice, France: Association for Computing Machinery).
Acosta-Escalante, F. D., Beltran-Naturi, E., Boll, M. C., Hernandez-Nolasco, J. A., and Pancardo Garcia, P. (2018). Meta-Classifiers in Huntington's Disease Patients Classification, Using iPhone's Movement Sensors Placed at the Ankles. IEEE Access 6, 30942–30957. doi:10.1109/access.2018.2840327
Aderinola, T. B., Connie, T., Ong, T. S., Yau, W.-C., and Teoh, A. B. J. (2021). Learning Age from Gait: A Survey. IEEE Access 9, 100352–100368. doi:10.1109/access.2021.3095477
Aich, S., Choi, K., and Hee-Cheol, K. (2018). “A Machine Learning Approach to Discriminate the Parkinsons Disease from Other Neurological Diseases Using Different Feature Selection Techniques Based on Gait Parameters,” in International Conference on Future Information & Communication Engineering, 355–358.
Ajay, J., Song, C., Wang, A., Langan, J., Li, Z., and Xu, W. (2018). “A Pervasive and Sensor-free Deep Learning System for Parkinsonian Gait Analysis,” in 2018 IEEE EMBS International Conference on Biomedical & Health Informatics (BHI), 108–111.
Alotaibi, M., and Mahmood, A. (2015). “Improved Gait Recognition Based on Specialized Deep Convolutional Neural Networks,” in 2015 IEEE Applied Imagery Pattern Recognition Workshop (AIPR), 1–7. doi:10.1109/aipr.2015.7444550
Alvarez-Alvarez, A., Trivino, G., and Cordon, O. (2012). Human Gait Modeling Using a Genetic Fuzzy Finite State Machine. IEEE Trans. Fuzzy Syst. 20, 205–223. doi:10.1109/tfuzz.2011.2171973
Alzubaidi, L., Zhang, J., Humaidi, A. J., Al-Dujaili, A., Duan, Y., Al-Shamma, O., et al. (2021). Review of Deep Learning: Concepts, CNN Architectures, Challenges, Applications, Future Directions. J. Big Data 8, 53. doi:10.1186/s40537-021-00444-8
Amisha, F., Malik, P., Pathania, M., and Rathaur, V. (2019). Overview of Artificial Intelligence in Medicine. J. Fam. Med Prim Care 8, 2328–2331. doi:10.4103/jfmpc.jfmpc_440_19
An, W., Yu, S., Makihara, Y., Wu, X., Xu, C., Yu, Y., et al. (2020). Performance Evaluation of Model-Based Gait on Multi-View Very Large Population Database with Pose Sequences. IEEE Trans. Biom. Behav. Identity Sci. 2, 421–430. doi:10.1109/tbiom.2020.3008862
Andersson, V., Dutra, R., and Araújo, R. (2014). “Anthropometric and Human Gait Identification Using Skeleton Data from Kinect Sensor,” in Proceedings of the 29th Annual ACM Symposium on Applied Computing (Gyeongju, Republic of Korea: Association for Computing Machinery). doi:10.1145/2554850.2555147
Anguita, D., Ghio, A., Oneto, L., Parra, X., and Reyes-Ortiz, J. L. (2013). “A Public Domain Dataset for Human Activity Recognition Using Smartphones,” in ESANN 2013 proceedings, European Symposium on Artificial Neural Networks, Computational Intelligence, 3.
Anguita, D., Ghio, A., Oneto, L., Parra, X., and Reyes-Ortiz, J. L. (2012). “Human Activity Recognition on Smartphones Using a Multiclass Hardware-Friendly Support Vector Machine,” in International workshop on ambient assisted living (Berlin, Germany: Springer), 216–223.
Arifoglu, D., and Bouchachia, A. (2017). Activity Recognition and Abnormal Behaviour Detection with Recurrent Neural Networks. Proced. Computer Sci. 110, 86–93. doi:10.1016/j.procs.2017.06.121
Arnold, Z., Larose, D., and Agu, E. (2015). “Smartphone Inference of Alcohol Consumption Levels from Gait,” in 2015 International Conference on Healthcare Informatics, 417–426. doi:10.1109/ichi.2015.59
Baghdadi, A., Cavuoto, L. A., Jones-Farmer, A., Rigdon, S. E., Esfahani, E. T., and Megahed, F. M. (2021). Monitoring Worker Fatigue Using Wearable Devices: A Case Study to Detect Changes in Gait Parameters. J. Qual. Technol. 53, 47–71. doi:10.1080/00224065.2019.1640097
Baghdadi, A., Megahed, F. M., Esfahani, E. T., and Cavuoto, L. A. (2018). A Machine Learning Approach to Detect Changes in Gait Parameters Following a Fatiguing Occupational Task. Ergonomics 61, 1116–1129. doi:10.1080/00140139.2018.1442936
Baker, R., Esquenazi, A., Benedetti, M. G., and Desloovere, K. (2016). Gait Analysis: Clinical Facts. Eur. J. Phys. Rehabil. Med. 52, 560–574.
Barbosa, I. B., Cristani, M., Del Bue, A., Bazzani, L., and Murino, V. (2012). “Re-identification with Rgb-D Sensors,” in European Conference on Computer Vision (Berlin, Germany: Springer), 433–442.
Barshan, B., and Yuksek, M. C. (2014). Recognizing Daily and Sports Activities in Two Open Source Machine Learning Environments Using Body-Worn Sensor Units. Computer J. 57, 1649–1667. doi:10.1093/comjnl/bxt075
Belkacem, A. N., Jamil, N., Palmer, J. A., Ouhbi, S., and Chen, C. (2020). Brain Computer Interfaces for Improving the Quality of Life of Older Adults and Elderly Patients. Front Neurosci. 14, 692. doi:10.3389/fnins.2020.00692
Bergamin, K., Clavet, S., Holden, D., and Forbes, J. R. (2019). DReCon: Data-Driven Responsive Control of Physics-Based Characters. ACM Trans. Graph. 38, 206. doi:10.1145/3355089.3356536
Bhattacharya, U., Mittal, T., Chandra, R., Randhavane, T., Bera, A., and Manocha, D. (2020). “Step: Spatial Temporal Graph Convolutional Networks for Emotion Perception from Gaits,” in Proceedings of the AAAI Conference on Artificial Intelligence, 1342–1350.
Bialkowski, A., Denman, S., Sridharan, S., Fookes, C., and Lucey, P. (2012). “A Database for Person Re-identification in Multi-Camera Surveillance Networks,” in 2012 International Conference on Digital Image Computing Techniques and Applications (DICTA) (Fremantle, WA, Australia: IEEE), 1–8.
Bilgin, S. (2017). The Impact of Feature Extraction for the Classification of Amyotrophic Lateral Sclerosis Among Neurodegenerative Diseases and Healthy Subjects. Biomed. Signal Process. Control 31, 288–294. doi:10.1016/j.bspc.2016.08.016
Bin Tariq, O., Lazarescu, M. T., Iqbal, J., and Lavagno, L. (2017). Performance of Machine Learning Classifiers for Indoor Person Localization with Capacitive Sensors. IEEE Access 5, 12913–12926. doi:10.1109/access.2017.2721538
Borelli, E., Paolini, G., Antoniazzi, F., Barbiroli, M., Benassi, F., Chesani, F., et al. (2019). HABITAT: An IoT Solution for Independent Elderly. Sensors (Basel) 19, 1258. doi:10.3390/s19051258
Bossard, L., Guillaumin, M., and Van Gool, L. (2013). “Event Recognition in Photo Collections with a Stopwatch HMM,” in Proceedings of the IEEE International Conference on Computer Vision, 1193–1200.
Bot, B. M., Suver, C., Neto, E. C., Kellen, M., Klein, A., Bare, C., et al. (2016). The mPower Study, Parkinson Disease mobile Data Collected Using ResearchKit. Sci. Data 3, 160011. doi:10.1038/sdata.2016.11
Burdack, J., Horst, F., Giesselbach, S., Hassan, I., Daffner, S., and Schöllhorn, W. I. (2019). Systematic Comparison of the Influence of Different Data Preprocessing Methods on the Classification of Gait Using Machine Learning. arXiv preprint arXiv:1911.04335.
Burdack, J., Horst, F., Giesselbach, S., Hassan, I., Daffner, S., and Schöllhorn, W. I. (2020). Systematic Comparison of the Influence of Different Data Preprocessing Methods on the Performance of Gait Classifications Using Machine Learning. Front. Bioeng. Biotechnol. 8, 260. doi:10.3389/fbioe.2020.00260
Cao, P., Xia, W., Ye, M., Zhang, J., and Zhou, J. (2018). Radar‐ID: Human Identification Based on Radar micro‐Doppler Signatures Using Deep Convolutional Neural Networks. IET Radar, Sonar & Navigation 12, 729–734. doi:10.1049/iet-rsn.2017.0511
Cao, Z., Zhu, C., Zhou, Y., Wang, Y., Chen, M., Ju, Y., et al. (2021). Risk Factors Related Balance Disorder for Patients with Dizziness/vertigo. BMC Neurol. 21, 186. doi:10.1186/s12883-021-02188-7
Cao, Z., Simon, T., Wei, S.-E., and Sheikh, Y. (2017). “Realtime Multi-Person 2d Pose Estimation Using Part Affinity fields,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 7291–7299.
Castro, F. M., Marín-Jiménez, M. J., Guil, N., and De La Blanca, N. P. (2017). “Automatic Learning of Gait Signatures for People Identification,” in International Work-Conference on Artificial Neural Networks (Berlin, Germany: Springer), 257–270.
Cernian, A., Tiganoaia, B., Sacala, I., Pavel, A., and Iftemi, A. (2020). PatientDataChain: A Blockchain-Based Approach to Integrate Personal Health Records. Sensors (Basel) 20, 6538. doi:10.3390/s20226538
Chang, W. J., Chen, L. B., Su, J. P., Chen, M. C., and Yang, T. C. (2021). “A Fall Risk Prediction System Based on 3D Space Human Skeleton Torso Images,” in 2021 IEEE International Conference on Consumer Electronics (ICCE), 1–2.
Chao, H., He, Y., Zhang, J., and Feng, J. (2019). “Gaitset: Regarding Gait as a Set for Cross-View Gait Recognition,” in Proceedings of the AAAI Conference on Artificial Intelligence, 8126–8133.
Chelli, A., and Patzold, M. (2019). A Machine Learning Approach for Fall Detection and Daily Living Activity Recognition. IEEE Access 7, 38670–38687. doi:10.1109/access.2019.2906693
Chen, J., Zhang, Z., and Wang, Y. (2015). Relevance Metric Learning for Person Re-identification by Exploiting Listwise Similarities. IEEE Trans. Image Process. 24, 4741–4755. doi:10.1109/tip.2015.2466117
Chen Wang, C., Junping Zhang, J., Liang Wang, L., Jian Pu, J., and Xiaoru Yuan, X. (2012). Human Identification Using Temporal Information Preserving Gait Template. IEEE Trans. Pattern Anal. Mach. Intell. 34, 2164–2176. doi:10.1109/tpami.2011.260
Chen, X., Weng, J., Lu, W., and Xu, J. (2018). Multi-Gait Recognition Based on Attribute Discovery. IEEE Trans. Pattern Anal. Mach. Intell. 40, 1697–1710. doi:10.1109/tpami.2017.2726061
Cheng, D. S., Cristani, M., Stoppa, M., Bazzani, L., and Murino, V. (2011). “Custom Pictorial Structures for Re-identification,” in Bmvc: Citeseer, 6.
Chettri, L., and Bera, R. (2019). A Comprehensive Survey on Internet of Things (IoT) toward 5G Wireless Systems. IEEE Internet Things J. 7, 16–32.
Choo, Y. J., Kim, J. K., Kim, J. H., Chang, M. C., and Park, D. (2021). Machine Learning Analysis to Predict the Need for Ankle Foot Orthosis in Patients with Stroke. Sci. Rep. 11, 8499. doi:10.1038/s41598-021-87826-3
Chtourou, I., Fendri, E., and Hammami, M. (2021). Person Re-identification Based on Gait via Part View Transformation Model under Variable Covariate Conditions. J. Vis. Commun. Image Representation 77, 103093. doi:10.1016/j.jvcir.2021.103093
Cicirelli, G., Marani, R., Petitti, A., Milella, A., and D'Orazio, T. (2021). Ambient Assisted Living: A Review of Technologies, Methodologies and Future Perspectives for Healthy Aging of Population. Sensors (Basel) 21, 3549. doi:10.3390/s21103549
Clark, R. A., Vernon, S., Mentiplay, B. F., Miller, K. J., Mcginley, J. L., Pua, Y. H., et al. (2015). Instrumenting Gait Assessment Using the Kinect in People Living with Stroke: Reliability and Association with Balance Tests. J. Neuroeng Rehabil. 12, 15. doi:10.1186/s12984-015-0006-8
Cook, D. J., Schmitter-Edgecombe, M., and Dawadi, P. (2015). Analyzing Activity Behavior and Movement in a Naturalistic Environment Using Smart Home Techniques. IEEE J. Biomed. Health Inform. 19, 1882–1892. doi:10.1109/jbhi.2015.2461659
Costilla-Reyes, O., Scully, P., Leroi, I., and Ozanyan, K. B. (2021). Age-Related Differences in Healthy Adults Walking Patterns under a Cognitive Task with Deep Neural Networks. IEEE Sensors J. 21, 2353–2363. doi:10.1109/jsen.2020.3021349
Costilla-Reyes, O., Vera-Rodriguez, R., Scully, P., and Ozanyan, K. B. (2019). Analysis of Spatio-Temporal Representations for Robust Footstep Recognition with Deep Residual Neural Networks. IEEE Trans. Pattern Anal. Mach. Intell. 41, 285–296. doi:10.1109/tpami.2018.2799847
Cronin, N. J., Rantalainen, T., Ahtiainen, J. P., Hynynen, E., and Waller, B. (2019). Markerless 2D Kinematic Analysis of Underwater Running: A Deep Learning Approach. J. Biomech. 87, 75–82. doi:10.1016/j.jbiomech.2019.02.021
Cui, C., Bian, G.-B., Hou, Z.-G., Zhao, J., Su, G., Zhou, H., et al. (2018). Simultaneous Recognition and Assessment of post-stroke Hemiparetic Gait by Fusing Kinematic, Kinetic, and Electrophysiological Data. IEEE Trans. Neural Syst. Rehabil. Eng. 26, 856–864. doi:10.1109/tnsre.2018.2811415
Daher, M., Diab, A., El Badaoui El Najjar, M., Ali Khalil, M., and Charpillet, F. (2017). Elder Tracking and Fall Detection System Using Smart Tiles. IEEE Sensors J. 17, 469–479. doi:10.1109/jsen.2016.2625099
Dapeng Tao, D., Lianwen Jin, L., Yongfei Wang, Y., and Xuelong Li, X. (2015). Person Reidentification by Minimum Classification Error-Based KISS Metric Learning. IEEE Trans. Cybern. 45, 242–252. doi:10.1109/tcyb.2014.2323992
Di Nardo, F., Morbidoni, C., Cucchiarelli, A., and Fioretti, S. (2020). Recognition of Gait Phases with a Single Knee Electrogoniometer: A Deep Learning Approach. Electronics 9, 355. doi:10.3390/electronics9020355
Dindorf, C., Konradi, J., Wolf, C., Taetz, B., Bleser, G., Huthwelker, J., et al. (2020a). General Method for Automated Feature Extraction and Selection and its Application for Gender Classification and Biomechanical Knowledge Discovery of Sex Differences in Spinal Posture during Stance and Gait. Comput. Methods Biomech. Biomed. Engin 24, 299–307. doi:10.1080/10255842.2020.1828375
Dindorf, C., Teufl, W., Taetz, B., Bleser, G., and Fröhlich, M. (2020b). Interpretability of Input Representations for Gait Classification in Patients after Total Hip Arthroplasty. Sensors (Basel) 20, 4385. doi:10.3390/s20164385
Ding, M., and Fan, G. (2015). Multilayer Joint Gait-Pose Manifolds for Human Gait Motion Modeling. IEEE Trans. Cybern. 45, 2413–2424. doi:10.1109/tcyb.2014.2373393
Dolatabadi, E., Taati, B., and Mihailidis, A. (2017). An Automated Classification of Pathological Gait Using Unobtrusive Sensing Technology. IEEE Trans. Neural Syst. Rehabil. Eng. 25, 2336–2346. doi:10.1109/tnsre.2017.2736939
Duan, M., Li, K., Yang, C., and Li, K. (2018). A Hybrid Deep Learning CNN-ELM for Age and Gender Classification. Neurocomputing 275, 448–461. doi:10.1016/j.neucom.2017.08.062
El-Alfy, E.-S. M., and Binsaadoon, A. G. (2019). Automated Gait-Based Gender Identification Using Fuzzy Local Binary Patterns with Tuned Parameters. J. Ambient Intell. Hum. Comput 10, 2495–2504. doi:10.1007/s12652-018-0728-0
Esteva, A., Robicquet, A., Ramsundar, B., Kuleshov, V., Depristo, M., Chou, K., et al. (2019). A Guide to Deep Learning in Healthcare. Nat. Med. 25, 24–29. doi:10.1038/s41591-018-0316-z
Fan, C., Peng, Y., Cao, C., Liu, X., Hou, S., Chi, J., et al. (2020). “GaitPart: Temporal Part-Based Model for Gait Recognition,” in 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 14213–14221. doi:10.1109/cvpr42600.2020.01423
Fang, H.-S., Xie, S., Tai, Y.-W., and Lu, C. (2017). “Rmpe: Regional Multi-Person Pose Estimation,” in Proceedings of the IEEE international conference on computer vision, 2334–2343.
Fang, J., Wang, T., Li, C., Hu, X., Ngai, E., Seet, B.-C., et al. (2019). Depression Prevalence in Postgraduate Students and its Association with Gait Abnormality. IEEE Access 7, 174425–174437. doi:10.1109/access.2019.2957179
Farah, J. D., Baddour, N., and Lemaire, E. D. (2017). “Gait Phase Detection from Thigh Kinematics Using Machine Learning Techniques,” in 2017 IEEE International Symposium on Medical Measurements and Applications (MeMeA), 263–268.
Feigl, T., Gruner, L., Mutschler, C., and Roth, D. (2020). “Real-Time Gait Reconstruction for Virtual Reality Using a Single Sensor,” in 2020 IEEE International Symposium on Mixed and Augmented Reality Adjunct (ISMAR-Adjunct), 84–89.
Felzenszwalb, P., Mcallester, D., and Ramanan, D. (2008). “A Discriminatively Trained, Multiscale, Deformable Part Model,” in 2008 IEEE Conference on Computer Vision and Pattern Recognition, 1–8.
Flagg, C., Frieder, O., Macavaney, S., and Motamedi, G. (2021). “Real-time Streaming of Gait Assessment for Parkinson's Disease,” in Proceedings of the 14th ACM International Conference on Web Search and Data Mining (Virtual Event, Israel: Association for Computing Machinery). doi:10.1145/3437963.3441701
Frey, A.-L., Karran, M., Jimenez, R. C., Baxter, J., Adeogun, M., Bose, N., et al. (2019). Harnessing the Potential of Digital Technologies for the Early Detection of Neurodegenerative Diseases (edon).
Gage, J. R., Deluca, P. A., and Renshaw, T. S. (1995). Gait Analysis. J. Bone Joint Surg. 77, 1607–1623. doi:10.2106/00004623-199510000-00017
Gao, C., Gehlhar, R., Ames, A. D., Liu, S. C., and Delbruck, T. (2020). “Recurrent Neural Network Control of a Hybrid Dynamical Transfemoral Prosthesis with EdgeDRNN Accelerator,” in 2020 IEEE International Conference on Robotics and Automation (ICRA), 5460–5466.
Gao, Q., Wang, J., Ma, X., Feng, X., and Wang, H. (2017). CSI-based Device-free Wireless Localization and Activity Recognition Using Radio Image Features. IEEE Trans. Veh. Technol. 66, 10346–10356. doi:10.1109/tvt.2017.2737553
Gholami, M., Napier, C., and Menon, C. (2020). Estimating Lower Extremity Running Gait Kinematics with a Single Accelerometer: A Deep Learning Approach. Sensors 20, 2939. doi:10.3390/s20102939
Ghose, A., Sinha, P., Bhaumik, C., Sinha, A., Agrawal, A., and Choudhury, A. D. (2013). “UbiHeld: Ubiquitous Healthcare Monitoring System for Elderly and Chronic Patients,” in Proceedings of the 2013 ACM conference on Pervasive and ubiquitous computing adjunct publication (Zurich, Switzerland: Association for Computing Machinery).
Goldberger, A. L., Amaral, L. A., Glass, L., Hausdorff, J. M., Ivanov, P. C., Mark, R. G., et al. (2000). PhysioBank, PhysioToolkit, and PhysioNet: Components of a New Research Resource for Complex Physiologic Signals. Circulation 101, e215–20. doi:10.1161/01.cir.101.23.e215
Gray, D., and Tao, H. (2008). “Viewpoint Invariant Pedestrian Recognition with an Ensemble of Localized Features,” in European conference on computer vision (Berlin, Germany: Springer), 262–275.
Guan, Y., Li, C., and Hu, Y. (2012). “Robust Clothing-Invariant Gait Recognition,” in 2012 Eighth International Conference on Intelligent Information Hiding and Multimedia Signal Processing, 321–324. doi:10.1109/iih-msp.2012.84
Guo, Y., Deligianni, F., Gu, X., and Yang, G.-Z. (2019). 3-D Canonical Pose Estimation and Abnormal Gait Recognition with a Single RGB-D Camera. IEEE Robot. Autom. Lett. 4, 3617–3624. doi:10.1109/lra.2019.2928775
Guo, Y., and Wang, X. (2021). Applying TS-DBN Model into Sports Behavior Recognition with Deep Learning Approach. J. Supercomput 77, 12192–12208. doi:10.1007/s11227-021-03772-x
Gurbuz, S. Z., and Amin, M. G. (2019). Radar-based Human-Motion Recognition with Deep Learning: Promising Applications for Indoor Monitoring. IEEE Signal Process. Mag. 36, 16–28. doi:10.1109/msp.2018.2890128
Hakim, A., Huq, M. S., Shanta, S., and Ibrahim, B. S. K. K. (2017). Smartphone Based Data Mining for Fall Detection: Analysis and Design. Proced. Computer Sci. 105, 46–51. doi:10.1016/j.procs.2017.01.188
Halilaj, E., Rajagopal, A., Fiterau, M., Hicks, J. L., Hastie, T. J., and Delp, S. L. (2018). Machine Learning in Human Movement Biomechanics: Best Practices, Common Pitfalls, and New Opportunities. J. Biomech. 81, 1–11. doi:10.1016/j.jbiomech.2018.09.009
Han, J., and Bhanu, B. (2005). Individual Recognition Using Gait Energy Image. IEEE Trans. Pattern Anal Mach Intell. 28, 316–322. doi:10.1109/TPAMI.2006.38
Haque, A., Alahi, A., and Fei-Fei, L. (2016). “Recurrent Attention Models for Depth-Based Person Identification,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), 1229–1238. doi:10.1109/cvpr.2016.138
Hasan, C. Z. C., Jailani, R., and Tahir, N. M. (2018). “ANN and SVM Classifiers in Identifying Autism Spectrum Disorder Gait Based on Three-Dimensional Ground Reaction Forces,” in TENCON 2018 - 2018 IEEE Region 10 Conference, 2436–2440.
Hasan, M., Hanawa, J., Goto, R., Fukuda, H., Kuno, Y., and Kobayashi, Y. (2020). “Tracking People Using Ankle-Level 2D LiDAR for Gait Analysis,” in International Conference on Applied Human Factors and Ergonomics (Berlin, Germany: Springer), 40–46.
Hassan, M., Daiber, F., Wiehr, F., Kosmalla, F., and Krüger, A. (2017). FootStriker: An EMS-Based Foot Strike Assistant for Running. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 1, 2. doi:10.1145/3053332
Hausdorff, J. M., Lertratanakul, A., Cudkowicz, M. E., Peterson, A. L., Kaliton, D., and Goldberger, A. L. (2000). Dynamic Markers of Altered Gait Rhythm in Amyotrophic Lateral Sclerosis. J. Appl. Physiol. 88, 2045. doi:10.1152/jappl.2000.88.6.2045
He, K., Zhang, X., Ren, S., and Sun, J. (2016). “Deep Residual Learning for Image Recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 770–778.
He, Y., Zhang, J., Shan, H., and Wang, L. (2019). Multi-Task GANs for View-specific Feature Learning in Gait Recognition. IEEE Trans.Inform.Forensic Secur. 14, 102–113. doi:10.1109/tifs.2018.2844819
Hernández-Álvarez, L., De Fuentes, J. M., González-Manzano, L., and Hernández Encinas, L. (2021). Privacy-Preserving Sensor-Based Continuous Authentication and User Profiling: A Review. Sensors (Basel) 21, 92. doi:10.3390/s21010092
Hintze, D., Füller, M., Scholz, S., Findling, R. D., Muaaz, M., Kapfer, P., et al. (2019). CORMORANT: Ubiquitous Risk-Aware Multi-Modal Biometric Authentication across Mobile Devices. Proc. ACM Interact. Mob. Wearable Ubiquitous Technol. 3, 85. doi:10.1145/3351243
Hirzer, M., Beleznai, C., Roth, P. M., and Bischof, H. (2011). “Person Re-identification by Descriptive and Discriminative Classification,” in Scandinavian conference on Image analysis (Berlin, Germany: Springer), 91–102.
Hofmann, M., Geiger, J., Bachmann, S., Schuller, B., and Rigoll, G. (2014). The TUM Gait from Audio, Image and Depth (GAID) Database: Multimodal Recognition of Subjects and Traits. J. Vis. Commun. Image Representation 25, 195–206. doi:10.1016/j.jvcir.2013.02.006
Holden, D., Saito, J., and Komura, T. (2016). A Deep Learning Framework for Character Motion Synthesis and Editing. ACM Trans. Graph. 35, 138. doi:10.1145/2897824.2925975
Horst, F., Lapuschkin, S., Samek, W., Müller, K. R., and Schöllhorn, W. I. (2019). Explaining the Unique Nature of Individual Gait Patterns with Deep Learning. Sci. Rep. 9, 2391. doi:10.1038/s41598-019-38748-8
Horst, F., Mildner, M., and Schöllhorn, W. I. (2017). One-year Persistence of Individual Gait Patterns Identified in a Follow-Up Study - A Call for Individualised Diagnose and Therapy. Gait & Posture 58, 476–480. doi:10.1016/j.gaitpost.2017.09.003
Hsu, Y. L., Chou, P. H., Chang, H. C., Lin, S. L., Yang, S. C., Su, H. Y., et al. (2017). Design and Implementation of a Smart Home System Using Multisensor Data Fusion Technology. Sensors (Basel) 17, 1631. doi:10.3390/s17071631
Hu, M., Wang, Y., Zhang, Z., Zhang, D., and Little, J. J. (2012). Incremental Learning for Video-Based Gait Recognition with LBP Flow. IEEE Trans. Cybern 43, 77–89. doi:10.1109/TSMCB.2012.2199310
Hu, M., Wang, Y., Zhang, Z., Little, J. J., and Huang, D. (2013). View-invariant Discriminative Projection for Multi-View Gait-Based Human Identification. IEEE Trans.Inform.Forensic Secur. 8, 2034–2045. doi:10.1109/tifs.2013.2287605
Huang, P., Tejera, M., Collomosse, J., and Hilton, A. (2015). Hybrid Skeletal-Surface Motion Graphs for Character Animation from 4D Performance Capture. ACM Trans. Graph. 34, 17. doi:10.1145/2699643
Huang, Y., Kaufmann, M., Aksan, E., Black, M. J., Hilliges, O., and Pons-Moll, G. (2018). Deep Inertial Poser: Learning to Reconstruct Human Pose from Sparse Inertial Measurements in Real Time. ACM Trans. Graph. 37, 185.
Iosa, M., Capodaglio, E., Pelà, S., Persechino, B., Morone, G., Antonucci, G., et al. (2021). Artificial Neural Network Analyzing Wearable Device Gait Data for Identifying Patients with Stroke Unable to Return to Work. Front Neurol. 12, 650542. doi:10.3389/fneur.2021.650542
Ishikawa, M., Yamada, S., and Yamamoto, K. (2019). Agreement Study on Gait Assessment Using a Video-Assisted Rating Method in Patients with Idiopathic normal-pressure Hydrocephalus. PLoS One 14, e0224202. doi:10.1371/journal.pone.0224202
Iwama, H., Okumura, M., Makihara, Y., and Yagi, Y. (2012). The Ou-Isir Gait Database Comprising the Large Population Dataset and Performance Evaluation of Gait Recognition. IEEE Trans.Inform.Forensic Secur. 7, 1511–1521. doi:10.1109/tifs.2012.2204253
Jain, A., and Kanhangad, V. (2018). Gender Classification in Smartphones Using Gait Information. Expert Syst. Appl. 93, 257–266. doi:10.1016/j.eswa.2017.10.017
Janssen, D., Schöllhorn, W. I., Newell, K. M., Jäger, J. M., Rost, F., and Vehof, K. (2011). Diagnosing Fatigue in Gait Patterns by Support Vector Machines and Self-Organizing Maps. Hum. Move. Sci. 30, 966–975. doi:10.1016/j.humov.2010.08.010
Jaouedi, N., Boujnah, N., and Bouhlel, M. S. (2020). A New Hybrid Deep Learning Model for Human Action Recognition. J. King Saud Univ. - Computer Inf. Sci. 32, 447–453. doi:10.1016/j.jksuci.2019.09.004
Jiang, X., Chu, K. H. T., Khoshnam, M., and Menon, C. (2018). A Wearable Gait Phase Detection System Based on Force Myography Techniques. Sensors (Basel) 18, 1279. doi:10.3390/s18041279
Jim, H. S. L., Hoogland, A. I., Brownstein, N. C., Barata, A., Dicker, A. P., Knoop, H., et al. (2020). Innovations in Research and Clinical Care Using Patient‐generated Health Data. CA A. Cancer J. Clin. 70, 182–199. doi:10.3322/caac.21608
Jimenez, M. F., Mello, R. C. D., Frizera, A., and Bastos, T. (2018). “Assistive Device for Guiding Visually Impaired People with Mobility Disorders,” in Companion of the 2018 ACM/IEEE International Conference on Human-Robot Interaction (Chicago, IL, USA: Association for Computing Machinery). doi:10.1145/3173386.3176910
Johnson, W. R., Mian, A., Robinson, M. A., Verheul, J., Lloyd, D. G., and Alderson, J. A. (2021). Multidimensional Ground Reaction Forces and Moments from Wearable Sensor Accelerations via Deep Learning. IEEE Trans. Biomed. Eng. 68, 289–297. doi:10.1109/tbme.2020.3006158
Jorquera Valero, J. M., Sánchez Sánchez, P. M., Fernández Maimó, L., Huertas Celdrán, A., Arjona Fernández, M., De Los Santos Vílchez, S., et al. (2018). Improving the Security and QoE in Mobile Devices through an Intelligent and Adaptive Continuous Authentication System. Sensors (Basel) 18, 3769. doi:10.3390/s18113769
Kang, X., Huang, B., and Qi, G. (2018). A Novel Walking Detection and Step Counting Algorithm Using Unconstrained Smartphones. Sensors (Basel) 18, 297. doi:10.3390/s18010297
Karanam, S., Gou, M., Wu, Z., Rates-Borras, A., Camps, O., and Radke, R. J. (2019). A Systematic Evaluation and Benchmark for Person Re-identification: Features, Metrics, and Datasets. IEEE Trans. Pattern Anal. Mach. Intell. 41, 523–536. doi:10.1109/tpami.2018.2807450
Karg, M., Seiberl, W., Kreuzpointner, F., Haas, J.-P., and Kulic, D. (2015). Clinical Gait Analysis: Comparing Explicit State Duration HMMs Using a Reference-Based Index. IEEE Trans. Neural Syst. Rehabil. Eng. 23, 319–331. doi:10.1109/tnsre.2014.2362862
Karvekar, S. B. (2019). Smartphone-based Human Fatigue Detection in an Industrial Environment Using Gait Analysis.
Kastaniotis, D., Theodorakopoulos, I., Economou, G., and Fotopoulos, S. (2013). “Gait-based Gender Recognition Using Pose Information for Real Time Applications,” in 2013 18th International Conference on Digital Signal Processing (DSP) (Fira, Greece: IEEE), 1–6.
Kastaniotis, D., Theodorakopoulos, I., and Fotopoulos, S. (2016). Pose-based Gait Recognition with Local Gradient Descriptors and Hierarchically Aggregated Residuals. J. Electron. Imaging 25, 063019. doi:10.1117/1.jei.25.6.063019
Kaur, R., Chen, Z., Motl, R., Hernandez, M. E., and Sowers, R. (2020). Predicting Multiple Sclerosis from Gait Dynamics Using an Instrumented Treadmill–A Machine Learning Approach. IEEE Trans. Biomed. Eng. 68, 2666. doi:10.1109/TBME.2020.3048142
Kenney, K., Iacono, D., Edlow, B. L., Katz, D. I., Diaz-Arrastia, R., Dams-O’Connor, K., et al. (2018). Dementia after Moderate-Severe Traumatic Brain Injury: Coexistence of Multiple Proteinopathies. J. Neuropathol. Exp. Neurol. 77, 50–63. doi:10.1093/jnen/nlx101
Keyes, O. (2018). The Misgendering Machines: Trans/HCI Implications of Automatic Gender Recognition. Proc. ACM Hum.-Comput. Interact. 2, 88. doi:10.1145/3274357
Khademi, G., Mohammadi, H., and Simon, D. (2019). Gradient-based Multi-Objective Feature Selection for Gait Mode Recognition of Transfemoral Amputees. Sensors 19, 253. doi:10.3390/s19020253
Khan, A., Mellor, S., Berlin, E., Thompson, R., Mcnaney, R., Olivier, P., et al. (2015). “Beyond Activity Recognition: Skill Assessment from Accelerometer Data,” in Proceedings of the 2015 ACM International Joint Conference on Pervasive and Ubiquitous Computing (Osaka, Japan: Association for Computing Machinery).
Khan, H., Naseer, N., Yazidi, A., Eide, P. K., Hassan, H. W., and Mirtaheri, P. (2020). Analysis of Human Gait Using Hybrid EEG-fNIRS-Based BCI System: A Review. Front Hum. Neurosci. 14, 613254. doi:10.3389/fnhum.2020.613254
Khan, R. A., Naseer, N., Qureshi, N. K., Noori, F. M., Nazeer, H., and Khan, M. U. (2018). fNIRS-Based Neurorobotic Interface for Gait Rehabilitation. J. Neuroengineering Rehabil. 15, 7. doi:10.1186/s12984-018-0346-2
Khodabandehloo, E., Riboni, D., and Alimohammadi, A. (2021). HealthXAI: Collaborative and Explainable AI for Supporting Early Diagnosis of Cognitive Decline. Future Generation Computer Syst. 116, 168–189. doi:10.1016/j.future.2020.10.030
Kim, D., Kim, S. H., Kim, T., Kang, B. B., Lee, M., Park, W., et al. (2021). Review of Machine Learning Methods in Soft Robotics. PLoS One 16, e0246102. doi:10.1371/journal.pone.0246102
Kim, K., and Cho, Y. K. (2020). Effective Inertial Sensor Quantity and Locations on a Body for Deep Learning-Based Worker's Motion Recognition. Automation in Construction 113, 103126. doi:10.1016/j.autcon.2020.103126
Kim, M. J., Cho, M. E., and Jun, H. J. (2020). Developing Design Solutions for Smart Homes through User-Centered Scenarios. Front Psychol. 11, 335. doi:10.3389/fpsyg.2020.00335
Kim, S., Soangra, R., Grant-Beuttler, M., and Aminian, A. (2019). Wearable Sensor-Based Gait Classification in Idiopathic Toe Walking Adolescents. Biomed. Sci. Instrum 55, 178–185.
Kondragunta, J., Jaiswal, A., and Hirtz, G. (2019). “Estimation of Gait Parameters from 3D Pose for Elderly Care,” in Proceedings of the 2019 6th International Conference on Biomedical and Bioinformatics Engineering, 66–72.
König, A., Crispim-Junior, C. F., Covella, A. G., Bremond, F., Derreumaux, A., Bensadoun, G., et al. (2015). Ecological Assessment of Autonomy in Instrumental Activities of Daily Living in Dementia Patients by the Means of an Automatic Video Monitoring System. Front Aging Neurosci. 7, 98. doi:10.3389/fnagi.2015.00098
Kotti, M., Duffell, L. D., Faisal, A. A., and Mcgregor, A. H. (2017). Detecting Knee Osteoarthritis and its Discriminating Parameters Using Random Forests. Med. Eng. Phys. 43, 19–29. doi:10.1016/j.medengphy.2017.02.004
Krizhevsky, A., Sutskever, I., and Hinton, G. E. (2012). Imagenet Classification with Deep Convolutional Neural Networks. Adv. Neural Inf. Process. Syst. 25, 1097–1105.
Kusakunniran, W. (2020). Review of Gait Recognition Approaches and Their Challenges on View Changes. IET biom. 9, 238–250. doi:10.1049/iet-bmt.2020.0103
Kwon, B., and Lee, S. (2021). Joint Swing Energy for Skeleton-Based Gender Classification. IEEE Access 9, 28334–28348. doi:10.1109/access.2021.3058745
Lan, G., Ma, D., Xu, W., Hassan, M., and Hu, W. (2020). Capacitor-based Activity Sensing for Kinetic-Powered Wearable IoTs. ACM Trans. Internet Things 1, 2. doi:10.1145/3362124
Lan, G., Ma, D., Xu, W., Hassan, M., and Hu, W. (2017). “CapSense: Capacitor-Based Activity Sensing for Kinetic Energy Harvesting Powered Wearable Devices,” in Proceedings of the 14th EAI International Conference on Mobile and Ubiquitous Systems: Computing, Networking and Services (Melbourne, VIC, Australia: Association for Computing Machinery).
Lasselin, J., Sundelin, T., Wayne, P. M., Olsson, M. J., Paues Göranson, S., Axelsson, J., et al. (2020). Biological Motion during Inflammation in Humans. Brain Behav. Immun. 84, 147–153. doi:10.1016/j.bbi.2019.11.019
Lee, M. H., Siewiorek, D. P., Smailagic, A., Bernardino, A., and Badia, S. B. I. (2019). “Learning to Assess the Quality of Stroke Rehabilitation Exercises,” in Proceedings of the 24th International Conference on Intelligent User Interfaces, 218–228.
Li, R., Agu, E., Balakrishnan, G., Herman, D., Abrantes, A., Stein, M., et al. (2019). “WeedGait: Unobtrusive Smartphone Sensing of Marijuana-Induced Gait Impairment by Fusing Gait Cycle Segmentation and Neural Networks,” in 2019 IEEE Healthcare Innovations and Point of Care Technologies, (HI-POCT), 91–94.
Li, W., Mahadevan, V., and Vasconcelos, N. (2013). Anomaly Detection and Localization in Crowded Scenes. IEEE Trans. Pattern Anal Mach Intell. 36, 18–32. doi:10.1109/TPAMI.2013.111
Li, W., Zhao, R., and Wang, X. (2012). “Human Reidentification with Transferred Metric Learning,” in Asian conference on computer vision (Berlin, Germany: Springer), 31–44.
Li, W., Zhao, R., Xiao, T., and Wang, X. (2014). “DeepReID: Deep Filter Pairing Neural Network for Person Re-identification,” in 2014 IEEE Conference on Computer Vision and Pattern Recognition, 152–159. doi:10.1109/cvpr.2014.27
Li, Y., Hu, H., Zhu, Z., and Zhou, G. (2020). SCANet: Sensor-Based Continuous Authentication with Two-Stream Convolutional Neural Networks. ACM Trans. Sen. Netw. 16, 29. doi:10.1145/3397179
Liang, Y., Samtani, S., Guo, B., and Yu, Z. (2020). Behavioral Biometrics for Continuous Authentication in the Internet-Of-Things Era: An Artificial Intelligence Perspective. IEEE Internet Things J. 7, 9128–9143. doi:10.1109/jiot.2020.3004077
Lianyang Ma, L., Xiaokang Yang, X., and Dacheng Tao, D. (2014a). Person Re-identification over Camera Networks Using Multi-Task Distance Metric Learning. IEEE Trans. Image Process. 23, 3656–3670. doi:10.1109/tip.2014.2331755
Liu, H., Xiao, Z., Fan, B., Zeng, H., Zhang, Y., and Jiang, G. (2021a). PrGCN: Probability Prediction with Graph Convolutional Network for Person Re-identification. Neurocomputing 423, 57–70. doi:10.1016/j.neucom.2020.10.019
Liu, J., Zha, Z.-J., Chen, X., Wang, Z., and Zhang, Y. (2019a). Dense 3D-Convolutional Neural Network for Person Re-identification in Videos. ACM Trans. Multimedia Comput. Commun. Appl. 15, 8. doi:10.1145/3231741
Liu, K., Ma, B., Zhang, W., and Huang, R. (2015). “A Spatio-Temporal Appearance Representation for Viceo-Based Pedestrian Re-identification,” in Proceedings of the IEEE International Conference on Computer Vision, 3810–3818.
Liu, W., Zhang, C., Ma, H., and Li, S. (2018). Learning Efficient Spatial-Temporal Gait Features with Deep Learning for Human Identification. Neuroinform 16, 457–471. doi:10.1007/s12021-018-9362-4
Liu, X., Sun, B., Zhang, Z., Wang, Y., Tang, H., and Zhu, T. (2019b). Gait Can Reveal Sleep Quality with Machine Learning Models. PLoS One 14, e0223012. doi:10.1371/journal.pone.0223012
Liu, Y., Zeng, Y., Pu, J., Shan, H., He, P., and Zhang, J. (2021b). “Selfgait: A Spatiotemporal Representation Learning Method for Self-Supervised Gait Recognition,” in ICASSP 2021-2021 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) (Toronto, ON, Canada: IEEE), 2570–2574.
Llorente-Vidrio, D., Pérez-San Lázaro, R., Ballesteros, M., Salgado, I., Cruz-Ortiz, D., and Chairez, I. (2020). Event Driven Sliding Mode Control of a Lower Limb Exoskeleton Based on a Continuous Neural Network Electromyographic Signal Classifier. Mechatronics 72, 102451. doi:10.1016/j.mechatronics.2020.102451
Loper, M., Mahmood, N., and Black, M. J. (2014). MoSh: Motion and Shape Capture from Sparse Markers. ACM Trans. Graph. 33, 220. doi:10.1145/2661229.2661273
Loper, M., Mahmood, N., Romero, J., Pons-Moll, G., and Black, M. J. (2015). Smpl. ACM Trans. Graph. 34, 1–16. doi:10.1145/2816795.2818013
Low, J.-H., Chee, P.-S., Lim, E.-H., and Ganesan, V. (2020). Design of a Wireless Smart Insole Using Stretchable Microfluidic Sensor for Gait Monitoring. Smart Mater. Struct. 29, 065003. doi:10.1088/1361-665x/ab802c
Loy, C. C., Hospedales, T. M., Xiang, T., and Gong, S. (2012). “Stream-based Joint Exploration-Exploitation Active Learning,” in 2012 IEEE Conference on Computer Vision and Pattern Recognition, 1560–1567. doi:10.1109/cvpr.2012.6247847
Loy, C. C., Xiang, T., and Gong, S. (2009). Xi'an, China: IEEE, 1988–1995. Multi-camera Activity Correlation Analysis2009 IEEE Conference on Computer Vision and Pattern Recognition
Lozano Domínguez, J. M., and Mateo Sanguino Tde, J. (2021). Walking Secure: Safe Routing Planning Algorithm and Pedestrian’s Crossing Intention Detector Based on Fuzzy Logic App. Sensors (Basel) 21.
Lu, C., Shi, J., and Jia, J. (2013). “Abnormal Event Detection at 150 Fps in Matlab,” in Proceedings of the IEEE international conference on computer vision, 2720–2727.
Lu, H., Shao, W., Ngai, E., Hu, X., and Hu, B. (2020). “A New Skeletal Representation Based on Gait for Depression Detection,” in 2020 IEEE International Conference on E-health Networking, Application & Services (HEALTHCOM), 1–6.
Lu, J., Wang, G., and Moulin, P. (2014). Human Identity and Gender Recognition from Gait Sequences with Arbitrary Walking Directions. IEEE Trans.Inform.Forensic Secur. 9, 51–61. doi:10.1109/tifs.2013.2291969
Lu, L., Megahed, F. M., Sesek, R. F., and Cavuoto, L. A. (2017). A Survey of the Prevalence of Fatigue, its Precursors and Individual Coping Mechanisms Among U.S. Manufacturing Workers. Appl. Ergon. 65, 139–151. doi:10.1016/j.apergo.2017.06.004
Lu, W., Chen, Y., Wang, J., and Qin, X. (2021b). Cross-domain Activity Recognition via Substructural Optimal Transport. Neurocomputing 454, 65–75. doi:10.1016/j.neucom.2021.04.124
Luo, C., Wu, J., Li, J., Wang, J., Xu, W., Ming, Z., et al. (2020). Gait Recognition as a Service for Unobtrusive User Identification in Smart Spaces. ACM Trans. Internet Things 1, 1–21. doi:10.1145/3375799
Ma, D., Lan, G., Xu, W., Hassan, M., and Hu, W. (2018). “SEHS: Simultaneous Energy Harvesting and Sensing Using Piezoelectric Energy Harvester,” in 2018 IEEE/ACM Third International Conference on Internet-of-Things Design and Implementation (IoTDI) (Orlando, FL, USA: IEEE), 201–212. doi:10.1109/iotdi.2018.00028
Ma, X., Wang, H., Xue, B., Zhou, M., Ji, B., and Li, Y. (2014b). Depth-Based Human Fall Detection via Shape Features and Improved Extreme Learning Machine. IEEE J. Biomed. Health Inform. 18, 1915–1922. doi:10.1109/jbhi.2014.2304357
Maghded, H. S., Ghafoor, K. Z., Sadiq, A. S., Curran, K., Rawat, D. B., and Rabie, K. (2020). “A Novel AI-Enabled Framework to Diagnose Coronavirus COVID-19 Using Smartphone Embedded Sensors: Design Study,” in 2020 IEEE 21st International Conference on Information Reuse and Integration for Data Science (IRI) (Las Vegas, NV, USA: IEEE), 180–187.
Mahesh, B. (2020). Machine Learning Algorithms-A Review. Int. J. Sci. Res. (Ijsr) [Internet] 9, 381–386.
Mannini, A., Genovese, V., and Sabatin, A. M. (2014). Online Decoding of Hidden Markov Models for Gait Event Detection Using Foot-Mounted Gyroscopes. IEEE J. Biomed. Health Inform. 18, 1122–1130. doi:10.1109/jbhi.2013.2293887
Mannini, A., and Sabatini, A. M (2012). Gait Phase Detection and Discrimination Between Walking–Jogging Activities Using Hidden Markov Models Applied to Foot Motion Data From a Gyroscope. Gait Postur 36, 657–661.
Martinho-Corbishley, D., Nixon, M. S., and Carter, J. N. (2019). Super-Fine Attributes with Crowd Prototyping. IEEE Trans. Pattern Anal. Mach. Intell. 41, 1486–1500. doi:10.1109/tpami.2018.2836900
Mathis, A., Mamidanna, P., Cury, K. M., Abe, T., Murthy, V. N., Mathis, M. W., et al. (2018). DeepLabCut: Markerless Pose Estimation of User-Defined Body Parts with Deep Learning. Nat. Neurosci. 21, 1281–1289. doi:10.1038/s41593-018-0209-y
Meena, T., and Sarawadekar, K. (2020). “Gender Recognition Using In-Built Inertial Sensors of Smartphone,” in 2020 IEEE REGION 10 CONFERENCE (TENCON), 462–467.
Mei, J., Desrosiers, C., and Frasnelli, J. (2021). Machine Learning for the Diagnosis of Parkinson's Disease: A Review of Literature. Front Aging Neurosci. 13, 633752. doi:10.3389/fnagi.2021.633752
Michalopoulos, N., Raptis, G. E., Katsini, C., Chrysikos, T., Mamalakis, M., and Vigotsky, A. (2016). “A Personalised Monitoring and Recommendation Framework for Kinetic Dysfunctions: The Trendelenburg Gait,” in Proceedings of the 20th Pan-Hellenic Conference on Informatics (Patras, Greece: Association for Computing Machinery).
Morbidoni, C., Cucchiarelli, A., Fioretti, S., and Di Nardo, F. (2019). A Deep Learning Approach to EMG-Based Classification of Gait Phases during Level Ground Walking. Electronics 8, 894. doi:10.3390/electronics8080894
Mori, R., Furukawa, T., Makino, Y., and Shinoda, H. (2020). “Measurement of Disturbance-Induced Fall Behavior and Prediction Using Neural Network,” in 2020 IEEE International Conference on Systems, Man, and Cybernetics (SMC), 2142–2149.
Munaro, M., Basso, A., Fossati, A., Van Gool, L., and Menegatti, E. (2014). “3D Reconstruction of Freely Moving Persons for Re-identification with a Depth Sensor,” in 2014 IEEE international conference on robotics and automation (ICRA) (Hong Kong, China: IEEE), 4512–4519.
Najafi, B., and Mishra, R. (2021). Harnessing Digital Health Technologies to Remotely Manage Diabetic Foot Syndrome: A Narrative Review. Medicina (Kaunas) 57, 377. doi:10.3390/medicina57040377
Nambiar, A., Bernardino, A., and Nascimento, J. C. (2019). Gait-based Person Re-identification. ACM Comput. Surv. 52, 1–34. doi:10.1145/3243043
Neverova, N., Wolf, C., Lacey, G., Fridman, L., Chandra, D., Barbello, B., et al. (2016). Learning Human Identity from Motion Patterns. IEEE Access 4, 1810–1820. doi:10.1109/access.2016.2557846
Ngo, T. T., Makihara, Y., Nagahara, H., Mukaigawa, Y., and Yagi, Y. (2015). Similar Gait Action Recognition Using an Inertial Sensor. Pattern Recognition 48, 1289–1301. doi:10.1016/j.patcog.2014.10.012
Ngo, T. T., Makihara, Y., Nagahara, H., Mukaigawa, Y., and Yagi, Y. (2014). The Largest Inertial Sensor-Based Gait Database and Performance Evaluation of Gait-Based Personal Authentication. Pattern Recognition 47, 228–237. doi:10.1016/j.patcog.2013.06.028
Nguyen, M. D., Mun, K.-R., Jung, D., Han, J., Park, M., Kim, J., et al. (2020). “IMU-based Spectrogram Approach with Deep Convolutional Neural Networks for Gait Classification,” in 2020 IEEE International Conference on Consumer Electronics (ICCE) (Las Vegas, NV, USA: IEEE), 1–6. doi:10.1109/icce46568.2020.9042999
Nunes, J. F., Moreira, P. M., and Tavares, J. M. R. (2019). “Benchmark RGB-D Gait Datasets: A Systematic Review,” in ECCOMAS Thematic Conference on Computational Vision and Medical Image Processing (Berlin, Germany: Springer), 366–372.
Ojetola, O., Gaura, E., and Brusey, J. (2015). “Data Set for Fall Events and Daily Activities from Inertial Sensors,” in Proceedings of the 6th ACM multimedia systems conference, 243–248.
Onodera, A. N., Gavião Neto, W. P., Roveri, M. I., Oliveira, W. R., and Sacco, I. C. (2017). Immediate Effects of EVA Midsole Resilience and Upper Shoe Structure on Running Biomechanics: a Machine Learning Approach. PeerJ 5, e3026. doi:10.7717/peerj.3026
Özdemir, A. T., and Barshan, B. (2014). Detecting Falls with Wearable Sensors Using Machine Learning Techniques. Sensors (Basel) 14, 10691–10708.
Papamichail, M. D., Chatzidimitriou, K. C., Karanikiotis, T., Oikonomou, N.-C. I., Symeonidis, A. L., and Saripalle, S. K. (2019). BrainRun: A Behavioral Biometrics Dataset towards Continuous Implicit Authentication. Data 4, 60. doi:10.3390/data4020060
Parisi, L., Ravichandran, N., and Lanzillotta, M. (2020). Artificial Intelligence for Clinical Gait Diagnostics of Knee Osteoarthritis: An Evidence-Based Review and Analysis.
Pech, M., Vrchota, J., and Bednář, J. (2021). Predictive Maintenance and Intelligent Sensors in Smart Factory: Review. Sensors (Basel) 21. doi:10.3390/s21041470
Peng, X. B., Abbeel, P., Levine, S., and Panne, M. V. D. (2018a). DeepMimic: Example-Guided Deep Reinforcement Learning of Physics-Based Character Skills. ACM Trans. Graph. 37, 143. doi:10.1145/3197517.3201311
Peng, X. B., Kanazawa, A., Malik, J., Abbeel, P., and Levine, S. (2018b). SFV: Reinforcement Learning of Physical Skills from Videos. ACM Trans. Graph. 37, 178. doi:10.1145/3197517.3201311
Pepa, L., Capecci, M., Andrenelli, E., Ciabattoni, L., Spalazzi, L., and Ceravolo, M. G. (2020). A Fuzzy Logic System for the home Assessment of Freezing of Gait in Subjects with Parkinsons Disease. Expert Syst. Appl. 147, 113197. doi:10.1016/j.eswa.2020.113197
Perera, A. G., Law, Y. W., and Chahl, J. (2019). Drone-action: An Outdoor Recorded Drone Video Dataset for Action Recognition. Drones 3, 82. doi:10.3390/drones3040082
Pérez-Ibarra, J. C., Siqueira, A. A., Terra, M. H., and Krebs, H. I. (2020). “Hybrid Simulated Annealing and Genetic Algorithm for Optimization of a Rule-Based Algorithm for Detection of Gait Events in Impaired Subjects,” in 2020 IEEE/ASME International Conference on Advanced Intelligent Mechatronics (AIM) (Boston, MA, USA: IEEE), 1167–1171.
Phinyomark, A., Hettinga, B. A., Osis, S., and Ferber, R. (2015). Do intermediate- and Higher-Order Principal Components Contain Useful Information to Detect Subtle Changes in Lower Extremity Biomechanics during Running? Hum. Movement Sci. 44, 91–101. doi:10.1016/j.humov.2015.08.018
Poniszewska-Maranda, A., Kaczmarek, D., Kryvinska, N., and Xhafa, F. (2019). Studying Usability of AI in the IoT Systems/paradigm through Embedding NN Techniques into mobile Smart Service System. Computing 101, 1661–1685. doi:10.1007/s00607-018-0680-z
Pouyanfar, S., Sadiq, S., Yan, Y., Tian, H., Tao, Y., Reyes, M. P., et al. (2018). A Survey on Deep Learning: Algorithms, Techniques, and Applications. ACM Comput. Surv. 51, 92. doi:10.1145/3150226
Prasanth, H., Caban, M., Keller, U., Courtine, G., Ijspeert, A., Vallery, H., et al. (2021). Wearable Sensor-Based Real-Time Gait Detection: a Systematic Review. Sensors 21, 2727. doi:10.3390/s21082727
Procházka, A., Vyšata, O., Vališ, M., Ťupa, O., Schätz, M., and Mařík, V. (2015). Bayesian Classification and Analysis of Gait Disorders Using Image and Depth Sensors of Microsoft Kinect. Digital Signal Process. 47, 169–177.
Pulido-Valdeolivas, I., Gómez-Andrés, D., Martín-Gonzalo, J. A., Rodríguez-Andonaegui, I., López-López, J., Pascual-Pascual, S. I., et al. (2018). Gait Phenotypes in Paediatric Hereditary Spastic Paraplegia Revealed by Dynamic Time Warping Analysis and Random Forests. PLoS One 13, e0192345. doi:10.1371/journal.pone.0192345
Qin, Z., Huang, G., Xiong, H., Qin, Z., and Choo, K. K. R. (2019). A Fuzzy Authentication System Based on Neural Network Learning and Extreme Value Statistics. IEEE Trans. Fuzzy Syst. 29, 549–559.
Rabbi, M., Ali, S., Choudhury, T., and Berke, E. (2011). Passive and In-Situ Assessment of Mental and Physical Well-Being Using Mobile Sensors.
Raghavendra, R., Bue, A., and Cristani, M. (2006). Unusual Crowd Activity Dataset of University of Minnesota.
Reiss, A., and Stricker, D. (2012). “Introducing a New Benchmarked Dataset for Activity Monitoring,” in 2012 16th international symposium on wearable computers (Newcastle, UK: IEEE), 108–109. doi:10.1109/iswc.2012.13
Renner, S. W., Qiao, Y., Gmelin, T., Santanasto, A. J., Boudreau, R. M., Walston, J. D., et al. (2021). Association of Fatigue, Inflammation, and Physical Activity on Gait Speed: the Long Life Family Study. Aging Clin. Exp. Res. 2021, 1–8. doi:10.1007/s40520-021-01923-x
Ristani, E., Solera, F., Zou, R., Cucchiara, R., and Tomasi, C. (2016). “Performance Measures and a Data Set for Multi-Target, Multi-Camera Tracking,” in Computer Vision – ECCV 2016 Workshops. Editors G. Hua, and H. Jégou (Berlin, Germany: Springer International Publishing), 17–35.
Robertson, P., Frassl, M., Angermann, M., Doniec, M., Julian, B. J., Puyol, M. G., et al. (2013). “Simultaneous Localization and Mapping for Pedestrians Using Distortions of the Local Magnetic Field Intensity in Large Indoor Environments,” in International Conference on Indoor Positioning and Indoor Navigation, 1–10.
Rogez, G., Rihan, J., Guerrero, J. J., and Orrite, C. (2014). Monocular 3-D Gait Tracking in Surveillance Scenes. IEEE Trans. Cybern. 44, 894–909. doi:10.1109/tcyb.2013.2275731
Roth, P. M., Hirzer, M., Köstinger, M., Beleznai, C., and Bischof, H. (2014). “Mahalanobis Distance Learning for Person Re-identification,” in Person Re-identification (Berlin, Germany: Springer), 247–267. doi:10.1007/978-1-4471-6296-4_12
Rudenko, A., Palmieri, L., Herman, M., Kitani, K. M., Gavrila, D. M., and Arras, K. O. (2020). Human Motion Trajectory Prediction: A Survey. Int. J. Robotics Res. 39, 895–935. doi:10.1177/0278364920917446
Rum, L., Sten, O., Vendrame, E., Belluscio, V., Camomilla, V., Vannozzi, G., et al. (2021). Wearable Sensors in Sports for Persons with Disability: A Systematic Review. Sensors (Basel) 21, 1858. doi:10.3390/s21051858
Russakovsky, O., Deng, J., Su, H., Krause, J., Satheesh, S., Ma, S., et al. (2015). ImageNet Large Scale Visual Recognition Challenge. Int. J. Comput. Vis. 115, 211–252. doi:10.1007/s11263-015-0816-y
Russell, B., Mcdaid, A., Toscano, W., and Hume, P. (2021). Moving the Lab into the Mountains: A Pilot Study of Human Activity Recognition in Unstructured Environments. Sensors (Basel) 21, 654. doi:10.3390/s21020654
Saghafi, M. A., Hussain, A., Zaman, H. B., and Saad, M. H. M. (2014). Review of Person Re‐identification Techniques. IET Computer Vis. 8, 455–474. doi:10.1049/iet-cvi.2013.0180
San-Segundo, R., Blunck, H., Moreno-Pimentel, J., Stisen, A., and Gil-Martín, M. (2018). Robust Human Activity Recognition Using Smartwatches and Smartphones. Eng. Appl. Artif. Intelligence 72, 190–202. doi:10.1016/j.engappai.2018.04.002
Sandhu, M. M., Khalifa, S., Geissdoerfer, K., Jurdak, R., and Portmann, M. (2021). “SolAR: Energy Positive Human Activity Recognition Using Solar Cells,” in 2021 IEEE International Conference on Pervasive Computing and Communications (PerCom), 1–10.
Santoyo-Ramón, J. A., Casilari, E., and Cano-García, J. M. (2018). Analysis of a Smartphone-Based Architecture with Multiple Mobility Sensors for Fall Detection with Supervised Learning. Sensors (Basel) 18, 1155. doi:10.3390/s18041155
Sarkar, S., Phillips, P. J., Liu, Z., Vega, I. R., Grother, P., and Bowyer, K. W. (2005). The Humanid Gait challenge Problem: Data Sets, Performance, and Analysis. IEEE Trans. Pattern Anal. Mach. Intell. 27, 162–177. doi:10.1109/tpami.2005.39
Sato, K., Nagashima, Y., Mano, T., Iwata, A., and Toda, T. (2019). Quantifying normal and Parkinsonian Gait Features from home Movies: Practical Application of a Deep Learning-Based 2D Pose Estimator. PloS one 14, e0223549. doi:10.1371/journal.pone.0223549
Schwartz, W. R., and Davis, L. S. (2009). “Learning Discriminative Appearance-Based Models Using Partial Least Squares,” in 2009 XXII Brazilian symposium on computer graphics and image processing (Rio de Janeiro, Brazil: IEEE), 322–329. doi:10.1109/sibgrapi.2009.42
Sedighi Maman, Z., Chen, Y.-J., Baghdadi, A., Lombardo, S., Cavuoto, L. A., and Megahed, F. M. (2020). A Data Analytic Framework for Physical Fatigue Management Using Wearable Sensors. Expert Syst. Appl. 155, 113405. doi:10.1016/j.eswa.2020.113405
Semwal, V. B., Raj, M., and Nandi, G. C. (2015). Biometric Gait Identification Based on a Multilayer Perceptron. Robotics Autonomous Syst. 65, 65–75. doi:10.1016/j.robot.2014.11.010
Sessoms, P. H., Gottshall, K. R., Collins, J.-D., Markham, A. E., Service, K. A., and Reini, S. A. (2015). Improvements in Gait Speed and Weight Shift of Persons with Traumatic Brain Injury and Vestibular Dysfunction Using a Virtual Reality Computer-Assisted Rehabilitation Environment. Mil. Med. 180, 143–149. doi:10.7205/milmed-d-14-00385
Seyfioglu, M. S., Ozbayoglu, A. M., and Gurbuz, S. Z. (2018). Deep Convolutional Autoencoder for Radar-Based Classification of Similar Aided and Unaided Human Activities. IEEE Trans. Aerosp. Electron. Syst. 54, 1709–1723. doi:10.1109/taes.2018.2799758
Shetty, S., and Rao, Y. S. (2016). “SVM Based Machine Learning Approach to Identify Parkinson's Disease Using Gait Analysis,” in 2016 International Conference on Inventive Computation Technologies (ICICT), 1–5.
Shi, B., Chen, X., Yue, Z., Yin, S., Weng, Q., Zhang, X., et al. (2019). Wearable Ankle Robots in Post-stroke Rehabilitation of Gait: A Systematic Review. Front Neurorobot 13, 63. doi:10.3389/fnbot.2019.00063
Shioiri, K., Saho, K., Fujimoto, M., and Kobayashi, Y. (2021). “Radar-based Gait Classification of Elderly Non-fallers and Multiple Fallers Using Machine Learning,” in 2021 IEEE 3rd Global Conference on Life Sciences and Technologies (LifeTech), 399–400.
Sigcha, L., Pavón, I., Costa, N., Costa, S., Gago, M., Arezes, P., et al. (2021). Automatic Resting Tremor Assessment in Parkinson's Disease Using Smartwatches and Multitask Convolutional Neural Networks. Sensors (Basel) 21, 291. doi:10.3390/s21010291
Singh, J. P., Jain, S., Arora, S., and Singh, U. P. (2018). Vision-Based Gait Recognition: A Survey. IEEE Access 6, 70497–70527. doi:10.1109/access.2018.2879896
Slijepcevic, D., Zeppelzauer, M., Schwab, C., Raberger, A.-M., Breiteneder, C., and Horsak, B. (2020). Input Representations and Classification Strategies for Automated Human Gait Analysis. Gait & Posture 76, 198–203. doi:10.1016/j.gaitpost.2019.10.021
Song, C., Huang, Y., Huang, Y., Jia, N., and Wang, L. (2019). Gaitnet: An End-To-End Network for Gait Based Human Identification. Pattern recognition 96, 106988. doi:10.1016/j.patcog.2019.106988
Song, C., Huang, Y., Ouyang, W., and Wang, L. (2018). “Mask-Guided Contrastive Attention Model for Person Re-identification,” in 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, 1179–1188. doi:10.1109/cvpr.2018.00129
Stoeve, M., Schuldhaus, D., Gamp, A., Zwick, C., and Eskofier, B. M. (2021). From the Laboratory to the Field: IMU-Based Shot and Pass Detection in Football Training and Game Scenarios Using Deep Learning. Sensors (Basel) 21, 71. doi:10.3390/s21093071
Strath, S. J., Kate, R. J., Keenan, K. G., Welch, W. A., and Swartz, A. M. (2015). Ngram Time Series Model to Predict Activity Type and Energy Cost from Wrist, Hip and Ankle Accelerometers: Implications of Age. Physiol. Meas. 36, 2335–2351. doi:10.1088/0967-3334/36/11/2335
Struss, P., Reiser, R., and Kreuzpointner, F. (2018). A Qualitative Gait Model for Diagnosis of Juvenile Idiopathic Arthritis.
Sturma, A., Schuhfried, O., Hasenoehrl, T., Ambrozy, C., Salminger, S., Hruby, L. A., et al. (2019). The Long-Term Effects of an Implantable Drop Foot Stimulator on Gait in Hemiparetic Patients. PLoS One 14, e0214991. doi:10.1371/journal.pone.0214991
Sun, Q., Liu, H., and Harada, T. (2017). Online Growing Neural Gas for Anomaly Detection in Changing Surveillance Scenes. Pattern Recognition 64, 187–201. doi:10.1016/j.patcog.2016.09.016
Sundholm, M., Cheng, J., Zhou, B., Sethi, A., and Lukowicz, P. (2014). “Smart-mat: Recognizing and Counting Gym Exercises with Low-Cost Resistive Pressure Sensing Matrix,” in Proceedings of the 2014 ACM international joint conference on pervasive and ubiquitous computing, 373–382.
Sutkowska, E., Sutkowski, K., Sokołowski, M., Franek, E., and Dragan, S. (2019). Distribution of the Highest Plantar Pressure Regions in Patients with Diabetes and its Association with Peripheral Neuropathy, Gender, Age, and BMI: One centre Study. J. Diabetes Res. 2019, 7395769. doi:10.1155/2019/7395769
Taborri, J., Keogh, J., Kos, A., Santuz, A., Umek, A., Urbanczyk, C., et al. (2020). Sport Biomechanics Applications Using Inertial, Force, and EMG Sensors: A Literature Overview. Appl. Bionics Biomech. 2020, 2041549. doi:10.1155/2020/2041549
Taborri, J., Molinaro, L., Santospagnuolo, A., Vetrano, M., Vulpiani, M. C., and Rossi, S. (2021). A Machine-Learning Approach to Measure the Anterior Cruciate Ligament Injury Risk in Female Basketball Players. Sensors (Basel) 21, 3141. doi:10.3390/s21093141
Taborri, J., Rossi, S., Palermo, E., Patanè, F., and Cappa, P. (2014). A Novel HMM Distributed Classifier for the Detection of Gait Phases by Means of a Wearable Inertial Sensor Network. Sensors 14, 16212–16234. doi:10.3390/s140916212
Takemura, N., Makihara, Y., Muramatsu, D., Echigo, T., and Yagi, Y. (2018). Multi-view Large Population Gait Dataset and its Performance Evaluation for Cross-View Gait Recognition. IPSJ Trans. Computer Vis. Appl. 10, 1–14. doi:10.1186/s41074-018-0039-6
Taralunga, D. D., and Florea, B. C. (2021). A Blockchain-Enabled Framework for mHealth Systems. Sensors (Basel) 21, 2828. doi:10.3390/s21082828
Totaro, M., Poliero, T., Mondini, A., Lucarotti, C., Cairoli, G., Ortiz, J., et al. (2017). Soft Smart Garments for Lower Limb Joint Position Analysis. Sensors (Basel) 17, 2314. doi:10.3390/s17102314
Trewin, S., Basson, S., Muller, M., Branham, S., Treviranus, J., Gruen, D., et al. (2019). Considerations for AI Fairness for People with Disabilities. AI Matters 5, 40–63. doi:10.1145/3362077.3362086
Vaith, A., Taetz, B., and Bleser, G. (2020). “Uncertainty Based Active Learning with Deep Neural Networks for Inertial Gait Analysis,” in 2020 IEEE 23rd International Conference on Information Fusion (FUSION), 1–8. doi:10.23919/fusion45008.2020.9190449
Van Kasteren, T. L. M., Englebienne, G., and Kröse, B. J. A. (2011). “Human Activity Recognition from Wireless Sensor Network Data: Benchmark and Software,” in Activity Recognition in Pervasive Intelligent Environments (Berlin, Germany: Springer), 165–186. doi:10.2991/978-94-91216-05-3_8
Vandersmissen, B., Knudde, N., Jalalvand, A., Couckuyt, I., Bourdoux, A., De Neve, W., et al. (2018). Indoor Person Identification Using a Low-Power FMCW Radar. IEEE Trans. Geosci. Remote Sensing 56, 3941–3952. doi:10.1109/tgrs.2018.2816812
Vera-Rodriguez, R., Mason, J. S., Fierrez, J., and Ortega-Garcia, J. (2012). Comparative Analysis and Fusion of Spatiotemporal Information for Footstep Recognition. IEEE Trans. Pattern Anal Mach Intell. 35, 823–834. doi:10.1109/TPAMI.2012.164
Wahid, F., Begg, R. K., Hass, C. J., Halgamuge, S., and Ackland, D. C. (2015). Classification of Parkinson's Disease Gait Using Spatial-Temporal Gait Features. IEEE J. Biomed. Health Inform. 19, 1794–1802. doi:10.1109/jbhi.2015.2450232
Wakefield, Jane. (2020). Wearable to Spot Alzheimer’s Being Developed. BBC News. Available at: https://www.bbc.com/news/technology-51474115 (Accessed February 13, 2020).
Wan, C., Wang, L., and Phoha, V. V. (2018a). A Survey on Gait Recognition. ACM Comput. Surv. (Csur) 51, 1–35.
Wan, S., Liang, Y., Zhang, Y., and Guizani, M. (2018b). Deep Multi-Layer Perceptron Classifier for Behavior Analysis to Estimate Parkinson's Disease Severity Using Smartphones. IEEE Access 6, 36825–36833. doi:10.1109/access.2018.2851382
Wang, C., Zhang, J., Wang, L., Pu, J., and Yuan, X. (2011). Human Identification Using Temporal Information Preserving Gait Template. IEEE Trans. Pattern Anal Mach Intell. 34, 2164–2176. doi:10.1109/TPAMI.2011.260
Wang, J., Hu, A., Liu, C., and Li, X. (2015). A Floor-Map-Aided WiFi/Pseudo-Odometry Integration Algorithm for an Indoor Positioning System. Sensors 15, 7096–7124. doi:10.3390/s150407096
Wang, T., Gong, S., Zhu, X., and Wang, S. (2016). Person Re-identification by Discriminative Selection in Video Ranking. IEEE Trans. Pattern Anal. Mach. Intell. 38, 2501–2514. doi:10.1109/tpami.2016.2522418
Wang, T., Gong, S., Zhu, X., and Wang, S. (2014). “Person Re-identification by Video Ranking,” in European Conference on Computer Vision (Berlin, Germany: Springer), 688–703.
Wang, Z., Hu, R., Chen, C., Yu, Y., Jiang, J., Liang, C., et al. (2018). Person Reidentification via Discrepancy Matrix and Matrix Metric. IEEE Trans. Cybern. 48, 3006–3020. doi:10.1109/tcyb.2017.2755044
Wannenburg, J., and Malekian, R. (2017). Physical Activity Recognition From Smartphone Accelerometer Data for User Context Awareness Sensing. IEEE Trans. Syst. Man Cybern.: Syst. 47, 3142–3149.
Wazzeh, A., Birdal, R. G., and Sertbaş, A. (2019). “Human Gait Based Gender Detection Using Light CNN with Max Feature Map,” in 2019 4th International Conference on Computer Science and Engineering (UBMK), 1–4. doi:10.1109/ubmk.2019.8907174
Wei-Shi Zheng, W., Shaogang Gong, S., and Tao Xiang, T. (2013). Reidentification by Relative Distance Comparison. IEEE Trans. Pattern Anal. Mach. Intell. 35, 653–668. doi:10.1109/tpami.2012.138
Weiss, G. M., Yoneda, K., and Hayajneh, T. (2019). Smartphone and Smartwatch-Based Biometrics Using Activities of Daily Living. IEEE Access 7, 133190–133202. doi:10.1109/access.2019.2940729
Wu, C., Song, Y., and Zhang, Y. (2019a). Multi-view Gait Recognition Using NMF and 2DLDA. Multimed Tools Appl. 78, 35789–35811. doi:10.1007/s11042-019-08153-4
Wu, G., Wang, J., Zhang, Y., and Jiang, S. (2018). A Continuous Identity Authentication Scheme Based on Physiological and Behavioral Characteristics. Sensors (Basel) 18, 179. doi:10.3390/s18010179
Wu, L., Wang, Y., Gao, J., and Li, X. (2019b). Where-and-When to Look: Deep Siamese Attention Networks for Video-Based Person Re-identification. IEEE Trans. Multimedia 21, 1412–1424. doi:10.1109/tmm.2018.2877886
Wu, Q., Zhang, Y. D., Tao, W., and Amin, M. G. (2015a). Radar‐based Fall Detection Based on Doppler Time-Frequency Signatures for Assisted Living. IET Radar, Sonar & Navigation 9, 164–172. doi:10.1049/iet-rsn.2014.0250
Wu, Z., Cheng, T., and Wang, Z. L. (2020). Self-Powered Sensors and Systems Based on Nanogenerators. Sensors (Basel) 20, 2925. doi:10.3390/s20102925
Wu, Z., Huang, Y., and Wang, L. (2015b). Learning Representative Deep Features for Image Set Analysis. IEEE Trans. Multimedia 17, 1960–1968. doi:10.1109/tmm.2015.2477681
Wu, Z., Huang, Y., Wang, L., Wang, X., and Tan, T. (2017). A Comprehensive Study on Cross-View Gait Based Human Identification with Deep CNNs. IEEE Trans. Pattern Anal. Mach. Intell. 39, 209–226. doi:10.1109/tpami.2016.2545669
Wu, Z., Li, Y., and Radke, R. J. (2015c). Viewpoint Invariant Human Re-identification in Camera Networks Using Pose Priors and Subject-Discriminative Features. IEEE Trans. Pattern Anal. Mach. Intell. 37, 1095–1108. doi:10.1109/tpami.2014.2360373
Xi, X., Tang, M., Miran, S. M., and Luo, Z. (2017). Evaluation of Feature Extraction and Recognition for Activity Monitoring and Fall Detection Based on Wearable sEMG Sensors. Sensors (Basel) 17, 1229. doi:10.3390/s17061229
Xiong, M., Chen, D., Chen, J., Chen, J., Shi, B., Liang, C., et al. (2019). Person Re-identification with Multiple Similarity Probabilities Using Deep Metric Learning for Efficient Smart Security Applications. J. Parallel Distributed Comput. 132, 230–241. doi:10.1016/j.jpdc.2017.11.009
Xu, W., Lan, G., Lin, Q., Khalifa, S., Bergmann, N., Hassan, M., et al. (2017). “KEH-gait: Towards a Mobile Healthcare User Authentication System by Kinetic Energy Harvesting,” in The Network and Distributed System Security Symposium, 4.
Xu, W., Lan, G., Lin, Q., Khalifa, S., Hassan, M., Bergmann, N., et al. (2018). Keh-gait: Using Kinetic Energy Harvesting for Gait-Based User Authentication Systems. IEEE Trans. Mobile Comput. 18, 139–152.
Xu, W., Xue, W., Lin, Q., Lan, G., Feng, X., Wei, B., et al. (2021). PrivGait: An Energy Harvesting-Based Privacy-Preserving User Identification System by Gait Analysis. IEEE Internet Things J. 99, 1. doi:10.1109/jiot.2021.3089618
Yadollahpour, A., Nourozi, J., Mirbagheri, S. A., Simancas-Acevedo, E., and Trejo-Macotela, F. R. (2018). Designing and Implementing an ANFIS Based Medical Decision Support System to Predict Chronic Kidney Disease Progression. Front Physiol. 9, 1753. doi:10.3389/fphys.2018.01753
Yang, J., and Yin, Y. (2021). Novel Soft Smart Shoes for Motion Intent Learning of Lower Limbs Using LSTM with a Convolutional Autoencoder. IEEE Sensors J. 21, 1906–1917. doi:10.1109/jsen.2020.3019053
Yang, J., Zou, H., Jiang, H., and Xie, L. (2018). Device-Free Occupant Activity Sensing Using WiFi-Enabled IoT Devices for Smart Homes. IEEE Internet Things J. 5, 3991–4002. doi:10.1109/jiot.2018.2849655
Yang, Q., Peng, G., Nguyen, D. T., Qi, X., Zhou, G., Sitová, Z., et al. (2014). “A Multimodal Data Set for Evaluating Continuous Authentication Performance in Smartphones,” in Proceedings of the 12th ACM Conference on Embedded Network Sensor Systems, 358–359.
Ye, Q., Xia, Y., and Yao, Z. (2018). Classification of Gait Patterns in Patients with Neurodegenerative Disease Using Adaptive Neuro-Fuzzy Inference System. Comput. Math. Methods Med. 2018, 1252. doi:10.1155/2018/9831252
Yu, S., Chen, H., and Brown, R. A. (2018). Hidden Markov Model-Based Fall Detection with Motion Sensor Orientation Calibration: A Case for Real-Life Home Monitoring. IEEE J. Biomed. Health Inform. 22, 1847–1853. doi:10.1109/jbhi.2017.2782079
Yu, S., Tan, D., and Tan, T. (2006). “A Framework for Evaluating the Effect of View Angle, Clothing and Carrying Condition on Gait Recognition,” in 18th International Conference on Pattern Recognition (ICPR'06) (Hong Kong, China: IEEE), 441–444.
Zeng, W., Ismail, S. A., and Pappas, E. (2020). Detecting the Presence of Anterior Cruciate Ligament Injury Based on Gait Dynamics Disparity and Neural Networks. Artif. Intell. Rev. 53, 3153–3176. doi:10.1007/s10462-019-09758-9
Zhang, H., Deng, K., Li, H., Albin, R. L., and Guan, Y. (2020a). Deep Learning Identifies Digital Biomarkers for Self-Reported Parkinson's Disease. Patterns 1, 100042. doi:10.1016/j.patter.2020.100042
Zhang, H., Liu, J., Li, K., Tan, H., and Wang, G. (2020b). Gait Learning Based Authentication for Intelligent Things. IEEE Trans. Veh. Technol. 69, 4450–4459. doi:10.1109/tvt.2020.2977418
Zhang, J., Lockhart, T. E., and Soangra, R. (2014a). Classifying Lower Extremity Muscle Fatigue during Walking Using Machine Learning and Inertial Sensors. Ann. Biomed. Eng. 42, 600–612. doi:10.1007/s10439-013-0917-0
Zhang, L., Zhang, L., Tao, D., and Du, B. (2015). A Sparse and Discriminative Tensor to Vector Projection for Human Gait Feature Representation. Signal Process. 106, 245–252. doi:10.1016/j.sigpro.2014.08.005
Zhang, M., and Sawchuk, A. A. (2012). “USC-HAD: a Daily Activity Dataset for Ubiquitous Activity Recognition Using Wearable Sensors,” in Proceedings of the 2012 ACM conference on ubiquitous computing, 1036–1043.
Zhang, Q., Wang, D., Zhao, R., Deng, Y., and Yu, Y. (2019a). “ShopEye: Fusing RFID and Smartwatch for Multi-Relation Excavation in Physical Stores,” in Proceedings of the 24th International Conference on Intelligent User Interfaces (Marina del Ray, California: Association for Computing Machinery).
Zhang, R., Wu, L., Yang, Y., Wu, W., Chen, Y., and Xu, M. (2020c). Multi-camera Multi-Player Tracking with Deep Player Identification in Sports Video. Pattern Recognition 102, 107260. doi:10.1016/j.patcog.2020.107260
Zhang, S., Wang, Y., and Li, A. (2019b). “Gait-Based Age Estimation with Deep Convolutional Neural Network,” in 2019 International Conference on Biometrics (ICB), 1–8. doi:10.1109/icb45273.2019.8987240
Zhang, W., Hu, S., Liu, K., and Zha, Z. (2019c). Learning Compact Appearance Representation for Video-Based Person Re-identification. IEEE Trans. Circuits Syst. Video Technol. 29, 2442–2452. doi:10.1109/tcsvt.2018.2865749
Zhang, Y., Pan, G., Jia, K., Lu, M., Wang, Y., and Wu, Z. (2014b). Accelerometer-based Gait Recognition by Sparse Representation of Signature Points with Clusters. IEEE Trans. Cybern 45, 1864–1875. doi:10.1109/TCYB.2014.2361287
Zhang, Y., Huang, Y., Yu, S., and Wang, L. (2020d). Cross-View Gait Recognition by Discriminative Feature Learning. IEEE Trans. Image Process. 29, 1001–1015. doi:10.1109/tip.2019.2926208
Zhang, Y., Li, S., Nolan, K. J., and Zanotto, D. (2020e). “Reinforcement Learning Assist-As-Needed Control for Robot Assisted Gait Training,” in 2020 8th IEEE RAS/EMBS International Conference for Biomedical Robotics and Biomechatronics (BioRob), 785–790.
Zhang, Z., He, T., Zhu, M., Sun, Z., Shi, Q., Zhu, J., et al. (2020f). Deep Learning-Enabled Triboelectric Smart Socks for IoT-Based Gait Analysis and VR Applications. npj Flexible Electronics 4, 1–12. doi:10.1038/s41528-020-00092-7
Zhang, Z. X., He, T. Y. Y., Zhu, M. L., Sun, Z. D., Shi, Q. F., Zhu, J. X., et al. (2020g). Deep Learning-Enabled Triboelectric Smart Socks for IoT-Based Gait Analysis and VR Applications. Npj Flexible Electronics 4, 29. doi:10.1038/s41528-020-00092-7
Zhao, N., Zhang, Z., Wang, Y., Wang, J., Li, B., Zhu, T., et al. (2019). See Your Mental State from Your Walk: Recognizing Anxiety and Depression through Kinect-Recorded Gait Data. PLoS one 14, e0216591. doi:10.1371/journal.pone.0216591
Zhao, R., Oyang, W., and Wang, X. (2017). Person Re-identification by Saliency Learning. IEEE Trans. Pattern Anal. Mach. Intell. 39, 356–370. doi:10.1109/tpami.2016.2544310
Zheng, L., Bie, Z., Sun, Y., Wang, J., Su, C., Wang, S., et al. (2016). “Mars: A Video Benchmark for Large-Scale Person Re-identification,” in European Conference on Computer Vision (Springer), 868–884.
Zheng, L., Shen, L., Tian, L., Wang, S., Wang, J., and Tian, Q. (2015). “Scalable Person Re-identification: A Benchmark,” in 2015 IEEE International Conference on Computer Vision (ICCV), 1116–1124.
Zheng, W.-S., Gong, S., and Xiang, T. (2009). “Associating Groups of People,” in British Machine Vision Conference, 1–11.
Zhong, Y., and Deng, Y. (2014). “Sensor Orientation Invariant mobile Gait Biometrics,” in IEEE International Joint Conference on Biometrics, 1–8.
Zhou, B., Song, B., Hassan, M. M., and Alamri, A. (2018). Multilinear Rank Support Tensor Machine for Crowd Density Estimation. Eng. Appl. Artif. Intelligence 72, 382–392. doi:10.1016/j.engappai.2018.04.011
Zhou, Y., Romijnders, R., Hansen, C., Campen, J. V., Maetzler, W., Hortobágyi, T., et al. (2020). The Detection of Age Groups by Dynamic Gait Outcomes Using Machine Learning Approaches. Sci. Rep. 10, 4426. doi:10.1038/s41598-020-61423-2
Zhu, Y., Xu, W., Luo, G., Wang, H., Yang, J., and Lu, W. (2020). Random Forest Enhancement Using Improved Artificial Fish Swarm for the Medial Knee Contact Force Prediction. Artif. intelligence Med. 103, 101811. doi:10.1016/j.artmed.2020.101811
Zou, H., Zhou, Y., Yang, J., Gu, W., Xie, L., and Spanos, C. J. (2018). “Wifi-based Human Identification via Convex Tensor Shapelet Learning,” in Thirty-Second AAAI Conference on Artificial Intelligence.
Keywords: review, human gait analysis, biometrics, machine learning, artificial intelligence
Citation: Harris EJ, Khoo I-H and Demircan E (2022) A Survey of Human Gait-Based Artificial Intelligence Applications. Front. Robot. AI 8:749274. doi: 10.3389/frobt.2021.749274
Received: 29 July 2021; Accepted: 01 November 2021;
Published: 03 January 2022.
Edited by:
Noman Naseer, Air University, PakistanReviewed by:
Emna Fendri, Faculty of Science of Sfax, TunisiaUsman Ghafoor, Institute of Space Technology, Pakistan
Copyright © 2022 Harris, Khoo and Demircan. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Emel Demircan, emel.demircan@csulb.edu