- 1Division of Automatic Control, Department of Electrical Engineering, The Institute of Technology, Linköping University, Linkoping, Sweden
- 2Eriksholm Research Centre, Oticon A/S, Snekkersten, Denmark
- 3Department of Electrical Engineering, Technical University of Denmark, Ørsteds Plads, Lyngby, Denmark
- 4Oticon A/S, Smorum, Denmark
Introduction: By means of adding more sensor technology, modern hearing aids (HAs) strive to become better, more personalized, and self-adaptive devices that can handle environmental changes and cope with the day-to-day fitness of the users. The latest HA technology available in the market already combines sound analysis with motion activity classification based on accelerometers to adjust settings. While there is a lot of research in activity tracking using accelerometers in sports applications and consumer electronics, there is not yet much in hearing research.
Objective: This study investigates the feasibility of activity tracking with ear-level accelerometers and how it compares to waist-mounted accelerometers, which is a more common measurement location.
Method: The activity classification methods in this study are based on supervised learning. The experimental set up consisted of 21 subjects, equipped with two XSens MTw Awinda at ear-level and one at waist-level, performing nine different activities.
Results: The highest accuracy on our experimental data as obtained with the combination of Bagging and Classification tree techniques. The total accuracy over all activities and users was 84% (ear-level), 90% (waist-level), and 91% (ear-level + waist-level). Most prominently, the classes, namely, standing, jogging, laying (on one side), laying (face-down), and walking all have an accuracy of above 90%. Furthermore, estimated ear-level step-detection accuracy was 95% in walking and 90% in jogging.
Conclusion: It is demonstrated that several activities can be classified, using ear-level accelerometers, with an accuracy that is on par with waist-level. It is indicated that step-detection accuracy is comparable to a high-performance wrist device. These findings are encouraging for the development of activity applications in hearing healthcare.
1. Introduction
A strong trend in modern hearing aid (HA) development and research is the inclusion of more sensing technologies. The driver behind this is the wish for better, more personalized, and self-adaptive (1–3) devices that can handle environmental changes (4–7) and cope with day-to-day fitness of the users. Current HAs usually try to analyze the soundscape and adjust the settings according to a formula. However, recent HAs have advanced further by combining sound analysis with motion activity classification based on accelerometers to adjust settings with the aim of a better user experience. A few other possible uses of accelerometers in HAs are as follows: fall detection (8) to alert caretakers; tap detection for user interfacing (9); and health monitoring based on physical activity (10). The backbone of the above-mentioned applications is accurate and robust activity tracking that can determine and distinguish between several relevant activities, e.g., standing, sitting, walking, running, and more. While a lot of research in activity tracking and classification using accelerometers has been in sports applications (11–14) and general consumer electronics (15–17), such as smart-watches and cell phones, the hearing research body is small. The results in this contribution is based on the work in Balzi (18). The background to this study is that accelerometers are, or are to appear, in hearing devices and that it is of fundamental interest to investigate their usefulness in the activity tracking. The key objective of this study is to investigate the feasibility of activity tracking with ear-level accelerometers and how it compares to waist-mounted accelerometers, which is the more common measurement location in sports and healthcare. The activity classification method is based on supervised learning on experimental data from 21 subjects. The scope of the investigation is limited to 21, normal hearing, healthy subjects, and 9 activities.
2. Method
This section outlines the relevant details of the activity tracking methodology based on accelerometer data and machine learning.
2.1. Accelerometer Measurements
It is assumed that the sensors are mounted rigidly onto the users and that any kind of mounting play is negligible. It is further assumed that sensor axes are orthogonal, that the sensitivities are known and linear in the working span, and that sensor biases are negligible. The assumed inertial (fixed) coordinate frame, with axes (XYZ), is a local, right-handed, Euclidean frame with the Z-axis parallel to the local gravity vector. The data from a tri-axial accelerometer are then
where is the body referenced measurement for the sensor axes (xyz), R is a rotation matrix relating the orientation of the inertial frame and the body frame, is the acceleration in an inertial frame, the local gravity vector g ≈ [0, 0, 9.81]Tm/s2 is assumed constant, and the noise, e, is assumed Gaussian distributed with the same standard deviation (SD), σe, in each axis, . The measured forces can be divided in to static forces, such as, constant acceleration and gravity, and the dynamic forces that are due to motion of changing rate, e.g., nodding and shaking. Note that even in the ideal case without noise, it is not possible to solve (Equation 1) for ai and other data, e.g., a magnetometer or a high-grade gyroscope is needed to resolve the rotation R, see Titterton et al. (19) for details. In most situations, the human body accelerations are small compared to gravity, and it is, therefore, possible to estimate the inclination, i.e., each sensor axis angle with respect to the gravity vector, which is related to the orientation in roll and pitch.
2.2. Accelerometer Features
In machine learning, feature extraction is pre-processing of data with the intention of increasing the overall performance of classifiers. The underlying idea is that certain transformations can yield more information, higher independence, and give larger margins for class separability. Feature selection is very much application dependent and usually require domain knowledge, though computationally expensive automated methods exist, [see, e.g., (20)] for an overview. The selected features described below are inspired by the work of Masse et al. (21), Gjoreski et al. (22), and Hua et al. (23) and have been adapted with this application in mind. The features are defined from 13 metrics of which 10 are applied to each axis, resulting in a total of 33 features as described below.
2.2.1. Tilt Angles
The tilt angles, sometimes referred to as inclination, are defined as
where
and is indicative of each axis angle with respect to the local gravity vector. Errors in the tilt angles arise from the presence of motion and noise.
2.2.2. Acceleration Vector Change
The acceleration vector change (AVC) is a motion-sensitive metric defined by the mean of the absolute value of the differences in the acceleration vector length (Equation 3), and the mean is calculated in a window with size M as
where i denotes samples the ith sample at a given, fixed, sampling frequency fs with the sampling interval Ts = 1/fs.
2.2.3. Signal Magnitude Area
The signal magnitude area (SMA) is defined over a window with size M as
and it is a measure of the magnitude.
2.2.4. Mean and SD
The mean (Equation 6) and SD (Equation 7) are computed for each axis over a window with size M as
2.2.5. Root Mean Square
Similar to the SD (Equation 7), the root mean square (RMS) is computed for each axis over a window of size M as
2.2.6. Minimum and Maximum
Minimum (MIN) and maximum (MAX) values per axis over a window with size M are defined by
and
respectively.
2.2.7. Median
The median is the center value of a size-ordered sample, and it is not skewed by large or small values as the mean is. The median, can e.g., be used to detect burst noise outliers in data and is computed for each axis over a window of size M as
2.2.8. Median Absolute Deviation
The median absolute deviation (MAD) is a measure of sample variability around the median. The MAD is computed for each axis over a window of size M as
where the MEDIAN from Equation (11) is used.
2.2.9. Skewness
The skewness (SKW) is the third standardized moment of a sample and is a measure of the asymmetry of a distribution about the mean. Using the previously defined μk (Equation 6) and σk (Equation 7), the sample SKW is computed for each axis over a window of size M as
Note that other approximations for sample skewness are possible (24).
2.2.10. Counts per Second
Counts per second (counts/s) is a widely used measure in the activity tracking. The computation of counts/s is proprietary of ActiGraph LLC and is usually carried out with ActiGraph accelerometer devices. However, Crouter et al. (25) details on how to derive this measure on a standard accelerometer and for the work here our Matlab implementation is based on Brønd et al. (26).
2.3. Machine Learning
The activity classification is based on a supervised learning to train the classifier. Three classification methods are considered in this study, K-nearest-neighbor (KNN), linear-discriminant-analysis (LDA), and decision tree (DT). Further improvement of the classification can be obtained using ensemble learning methods such as Boosting and Bootstrap aggregation (Bagging), and variations thereof.
2.3.1. Classifiers
In supervised learning, a set of training instances with corresponding class labels is given, and a classifier is trained and used to predict the class of an unseen instance, [see, e.g., (27)] for details. The N samples of training data x and class labels y were ordered in pairs {(x1, y1), …, (xN, yN)} such that the i-th feature vector corresponds to the binary class label vector . For the case with a single tri-axial accelerometer and the features described in section 2.2, the feature vector dimension, p, is 33 per sample while the class label vector dimension c is 9.
The three well-known, but rather different, supervised classifiers are considered, and the choice to use these was based on the availability of good implementations. The classifiers are as follows:
• The KNN classifier (28, 29) is here based on the Euclidean distance between the test- and training samples. However, other distance measures can be used.
• The LDA, or Fisher's discriminant (30), is a statistical method to find linear combinations in the feature space to separate the classes, and it is carried out by solving a generalized eigenvalue problem.
• DT has a flow chart-like structure where the root corresponds to feature inputs, the branches of the descending test-nodes represents the outcome of the test, and each leaf-node represents a class label, [see, e.g., (31)].
2.3.2. Ensemble Training
Ensemble training is used to increase the predictive classification (or regression) performance by learning a combination of several classifiers. The two main categories used here are Bagging (31) and Boosting (32) with a few selected variations. In Bagging, the classifiers are trained in parallel on randomly sampled training data while in Boosting the training is carried out sequentially as the classifiers and data are weighted according to their importance. The first, and most well-known, Boosting algorithm is called AdaBoost (short for Adaptive Boosting) and was originally formulated in Freund and Schapire (33).
3. Experiment
Experimental data were collected from 21 voluntary subjects performing a series of tasks representative of the stipulated activities. A total of three tri-axial accelerometers were used. Experimental data were collected jointly in the projects (18, 34).
3.1. Subjects
Prior to the experiments, the subjects were informed about the experiment procedure and the use of data before deciding on their participation with oral consent. Data were stored and labeled anonymously. The 21 subjects had an age range between 24 and 60 years old, , both women and men with a height bracket of 1.55−195 m, . None of the subjects had any reported health issues. The subjects did not receive any compensation.
3.2. Data
For all subjects, two of the accelerometers were placed on each side of the head at ear-level, see Figure 1, to mimic HA sensors and the third accelerometer was placed at waist-level using an elastic strap, see Figure 2, as it is a more common region for activity measurements and it is also representative of an in-pocket smartphone.
The accelerometers are XSens MTw Awinda produced by XSens Technologies B.V. and are battery powered, wireless devices, inertial sensors containing tri-axial accelerometers and gyroscopes and also tri-axial magnetometers enabling accurate orientation estimation when the device is stationary. The data are collected wirelessly using the MT Manager software by XSens on a PC laptop running Microsoft Windows 10 at a sampling rate, fs, of 100 Hz that was decided to be fast enough for the intended activities. Data are manually labeled based on visual inspection during the experiment to match the activities in section 3.4.
3.3. Task
The experiment was carried out in a room with a soft carpet at Oticon main offices, Smørum, Denmark. For tasks involving lying down and falling, a mattress was used. Each of the 21 subjects from whom the data have been gathered was asked to perform 6 different tasks while wearing all three accelerometers, and between each task all the data from the accelerometers were saved and anonymously cataloged. Except for the accelerometer data, from each subject, only gender, age, and height were collected. To every subject, the same specific information about which actions to be carried out was given by reading out loud from a manuscript, and no restrictions were communicated regarding how to carry out the various exercises with the intention of increasing the possibility of movement variability in the activities. With more in-class variation used for training, the classifier is less prone to over-fitting at the expense of higher probability of between-class overlap. The mean test duration was about 22 min, including pauses, and generated about 13–14 min of data per subject.
3.4. Physical Activities
The choice of activities to track is a trade-off between how clearly activities can be discerned from each other, the likelihood of activities being present in the daily routines of the subjects, and the intended use of activity tracking. A typical scenario, not addressed here, is that HA users often remove the HAs when lying down to rest. Hence, for a HA applications, in-ear detection, e.g., using accelerometers could be useful. The fidelity of activity categories is chosen as either resting or moving, and no intensity or within-class variation is considered.
The resting activities are as follows:
Act1 :Standing in a still position.
Act2 :Sitting
Act3 :Lying face-up (LFU)
Act4 :Lying face-down (LFD)
Act5 :Lying side (LS) on either left or right side
and the moving activities are as follows:
Act6 :Walking, on the floor
Act7 :Jogging, moderate pace in circles/square
Act8 :Falling on whichever side, some subjects could not simulate a perfect falling motion and, therefore, have been asked to perform a fast transition from standing to Lying down, at top of their capability.
Act9 :Transitioning (TRN) all instances not being any of the other activities, e.g., going from one activity to another.
Note that there is no specific class for head motion as it was predicted being difficult to correctly label and that there is already significant head motion within all the moving activities 3.4 to 3.4. With e.g., a waist-level accelerometer, it may be possible to separate head motions from general body motions, but it is beyond this study.
4. Results
As an initial step, 200 min of the dataset in Anguita et al. (35) was used on all combinations of classfiers and ensemble training methods described in Section 2.3. The data are open, pre-labeled, in-pocket cell phone, and it has six activity classes: walking; walking stairs up; walking stairs down; sitting; standing; and laying. For the use here, all the walking classes were considered the same. The best predictive classification accuracy was obtained using the DT and Bagging, and it was furthermore also the case for our experimental data. Consequently, all following results are obtained using DT and Bagging.
4.1. Pre-processing
The data from the 21 subjects were randomly partitioned in to two groups with all activities present in both groups and with 70% used for training and 30% used for testing (validation). Training with cross-validation is carried out for as many times as there are trees, i.e., 100–500 depending on the setup. The tree depth is 480 with 13,121 nodes. Features are computed for each sample at 100 Hzwith the window size to one sample, M = 1 for the features AVC and SMA, while M = fs = 100, centered at the current sample, for the other applicable features. All results are obtained using the Machine Learning Toolbox in Matlab 2021b and with dependencies to the Optimization Toolbox for certain classifiers.
4.2. Classification
The main performance target here is the accuracy of predicted class labels in data not used for training as it is a common measure in supervised learning, [see, e.g., (11, 12, 14, 36, 37)]. The accuracy is defined as the number of correctly classified labels divided by the total number of labels and simply states how much of the data not used for training that is correctly classified. In Table 1, the classification result using both ear-level accelerometers is illustrated based on 500 Decision trees and Bagging trained with a learning rate of 0.1 and showing an overall predicted accuracy of 84.4%. Most prominently, the classes, namely, standing 3.4, jogging 3.4, laying side 3.4, lying face-down 3.4, and, walking 3.4, all have an accuracy of above 90%. The lowest scoring activities are as follows: falling 3.4, sitting 3.4, and transitioning 3.4. The falling activity is often confused with transitioning, which, in turn, generally is confounded with all other activities. The overall accuracy is more than 90% without the sitting activity.
Table 1. Confusion matrix of the predicted accuracy with Bagging and Decision Tree using both ear-level accelerometers.
4.3. Feature Evaluation
From a computational perspective, it is good to minimize the number of features needed and, therefore, the relative importance of features for each activity is analyzed using 100 Decision trees and Bagging, with a learning rate of 0.1. In Table 2, the total accuracy per feature (or pairs in some cases) for each activity is considered where accelerations only on the top row are considered that base model and the contribution of each additional feature and activity are below. Notably, the tilt angles have only a marginal positive effect on the sitting activity and mostly negative effect on all other activities. Other features with little importance are RMS and counts/s. In Table 3, the increase (or regression) per feature and activity, compared to the base level accelertions only, is shown. The overall most important features are: mean, SD, MIN, MAX, median, and MAD. Furthermore, in Table 4 a summary of the best and worst activity per feature is shown and, as expected, the best ranking features are important for several activities.
Table 2. Total accuracy per feature (or pairs of features) per activity compared to just using accelerations (ACC) only.
Table 4. Best- and worst-case per feature accuracy using the features listed in the left column based on Table 3.
4.4. Sensor Combinations
One of the main motivations of this study is to compare the feasibility of ear-level activity tracking compared to waist-level sensoring. For all sensor combinations and data types the same features and training were computed to get the comparable classification results. In Table 5, a confusion matrix shows the results of using only the waist-level accelerometer, giving an overall accuracy of 89.6% In Table 6, a confusion matrix showing the result from two ear-level accelerometers and the waist-level accelerometer is used together with 500 Decision trees, giving an overall accuracy of 91.6%. One of the reasons for the improvement is that the addition of the waist-level accelerometer makes the sitting activity easier to distinguish with 94.4% correct compared to 52.5% when using only ear-level accelerometers and this performance increase also shows on the standing activity as the previously discussed confusion is decreased. In Table 7, the overall accuracy using the various combinations of sensors is shown; for instance, 25 Hzaccelerometer data from a wrist-worn Garmin Vivosmart 4 are used with only a 56.3% accuracy. In Table 7, 100 Decision trees compared to the previous 500 were chosen to save computations and the accuracy decrease is negligible. The gyroscope and orientation data are obtained from the XSens devices. Note that the orientation data are adapted for stationary orientations and, therefore, of low-pass characteristics and potentially not well suited for all aspects of this application. It can be noted that the waist-level accelerometer alone is rather efficient and, as noted before, the combination with the ear-level accelerometers gives even better performance.
Table 6. Confusion matrix using Bagging and Decision Tree using both ear-level accelerometers, and the waist accelerometer.
4.5. Step Detection
Another concrete measure that can be useful for activity tracking is step detection, which was also analyzed in Acker (10). For the walking and jogging activities, ear-level step detection was computed based on Bai et al. (38) and Abadleh etal. (39) using the AVC feature resulting in 95% and a 90% accuracy, respectively, using both ear-level accelerometers. This can be compared with the highly optimized Garmin Forerunner 35 giving a 99% (walking) and 95% (jogging) accuracy, respectively.
5. Discussion
The main objective of this study is to compare ear- and waist-level activity tracking performance. Therefore, it is not fundamental to have features and classifiers that could outperform the works of others and such a comparison is beyond this study. While not directly comparing to other methods, the overall ear-level activity classification results are encouraging in the proposed setup.
As noted, it is difficult to separate falling and transitioning with ear-level data only. Possible explanations are that the selected features are not sensitive enough to distinguish between falling and transitioning and that transitioning is too general [or complex as Dernbach et al. (11)] and can, for instance, be confounded with general head movements. More controlled falling experiments, such as, Burwinkel and Xu (8) could provide useful insights. The sitting activity was confounded with the standing activity as they are typically rather similar, and the only potential differences at ear-level between the two may be in postural sway that should be clearer for standing subjects. This difficulty was also found in Parkka et al. (40), and an accelerometer below the waist is particularly useful here. At waist-level it is easier to distinguish sitting and standing and it is possibly explained by the change in the tilt angles for seated subjects.
Designing features sensitive to particular classes is an engineering task requiring expertise. As noted in section 4.3, some features are not that well-suited for any of the activities and would benefit from further tuning, such as, other pre-processing and different window sizes, or should otherwise be omitted. A feature that is sensitive to postural sway (low-frequency component) could potentially support separating standing and sitting at ear-level.
Sensor combinations can improve the results, [see, e.g., (14, 40)], and here the combination of ear- and waist-level data is the overall best. On another positive note, the accuracy difference between one and two ear-level devices is small and this is a good news for HA applications, as single-sided hearing compensation is common. The wrist data, here from the Garmin device, may be difficult in general as arms may do many types of motions not specifically relating to the activities.
The use of gyroscope data is common, [see, e.g., (41–43)], and was expected to improve the results in general. However, all sensor types were processed in the same fashion as accelerometer data with the same features and are a possible explanation of the poor performance of many of the additional data types in Table 7.
Wrist data were poor for activity tracking in the setup here, but the output of the proprietary algorithms on these types of devices suggests that a lot more can be achieved on ear-level devices too. The step detection is almost on par with the commercial Garmin device, for the short durations considered here, and these typically utilize additional sensors, e.g., magnetometer and gyroscope, and their algorithms can be considered state-of-the-art.
6. Conclusions
We investigated the feasibility of ear-level accelerometers for activity tracking in comparison to waist-level accelerometers. Many activities can be classified with an accuracy that is on par with a waist-level accelerometer, and this is particularly encouraging for the development of activity applications in hearing healthcare. Furthermore, we indicate that step-detection accuracy is comparable to a high-performance wrist device. It is also shown that higher predictive performance can be obtained when combining ear- and waist-level accelerometer data, and this could potentially assist in isolating head motion from full body activities, opening for a higher granularity of activity classes.
Noteworthy limitations in this study were as follows: the modest number of test subjects (21); the number of activities (9) per test subject; that the manual data labeling may have errors; and efforts spent on feature design and learning methods. Also, data with a more control and a clearer reference, e.g., a motion capture system, could provide valuable insight at the expense of more costly and complex experiments.
Future directions should consider further feature design with, e.g., multi-tapers and various transforms. The design should start with time-frequency analysis of activities for guidance. Experiments on a larger, more diverse population, with additional knowledge on head motion/orientation throughout, can open up for higher performance and other activity classes. The classification methods could be further improved considering the state-of-the-art in Deep Learning as initially explored by Ronao and Cho (37) and Hammerla et al. (44).
Data Availability Statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics Statement
Ethical review and approval was not required for the study on human participants in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.
Author Contributions
The majority of manuscript preparation was carried out by MS with assistance from all authors. GB generated results. GB, TB, and MS developed the methodology. GB and EJ carried out the experiments and data collection.
Funding
This work was financially supported by the Swedish Research Council (Vetenskapsrådet, VR 2017-06092 Mekanismer och behandling vid åldersrelaterad hörselnedsättning).
Conflict of Interest
MS, TB, and SR-G are employed by Oticon A/S.
The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
The authors are grateful for the kind support from François Patou at Oticon Medical, Elvira Fischer, and Martha Shiell at Oticon.
References
1. Johansen B, Flet-Berliac YPR, Korzepa MJ, Sandholm P, Pontoppidan NH, Petersen MK, et al. Hearables in hearing care: discovering usage patterns through IoT devices. In: Antona M, Stephanidis C, editors. Universal Access in Human-Computer Interaction. Human and Technological Environments. Cham: Springer International Publishing (2017). p. 39–49.
2. Korzepa MJ, Johansen B, Petersen MK, Larsen J, Larsen JE, Pontoppidan NH. Modeling user intents as context in smartphone-connected hearing aids. In: Adjunct Publication of the 26th Conference on User Modeling, Adaptation and Personalization. UMAP '18. New York, NY: Association for Computing Machinery (2018). p. 151–5.
3. Wang EK, Liu H, Wang G, Ye Y, Wu TY, Chen CM. Context recognition for adaptive hearing-aids. In: 2015 IEEE 13th International Conference on Industrial Informatics (INDIN). Cambridge: IEEE (2015). p. 1102–7.
4. Insignia. Insio NX (2021). Available online at: https://www.signia.net/en/hearing-aids/other/insio-nx-bluetooth/; https://www.signia.net/en/hearing-aids/other/insio-nx-bluetooth/.
5. Starkey Livio AI Hearing Aids (2021). Available online at: https://www.starkey.com/hearing-aids/livio-artificial-intelligence-hearing-aids. Available from: https://www.starkey.com/hearing-aids/livio-artificial-intelligence-hearing-aids.
6. Oticon More Hearing Aids (2021). Available online at: https://www.oticon.com/solutions/more-hearing-aids.
7. GN Resound, Resound One (2021). Available online at: https://www.resound.com/en-us/hearing-aids/resound-hearing-aids/resound-one; https://www.resound.com/en-us/hearing-aids/resound-hearing-aids/resound-one.
8. Burwinkel JR, Xu B. Preliminary examination of the accuracy of a fall detection device embedded into hearing instruments. Am Acad Audiol. (2020) 31:393–403. doi: 10.3766/jaaa.19056
9. Starkey. Livio Ai Tap Control Feature (2019). Available online at: https://www.starkey.co.uk/blog/articles/2019/04/Livio-AI-hearing-aids-tap-control-featur.
10. Acker K. The Ear is the New Wrist: Livio AI's Sensors Lead in Step Accuracy (2018). Available online at: https://starkeypro.com/pdfs/white-papers/The_Ear_Is_The_New_Wrist.pdf.
11. Dernbach S, Das B, Krishnan NC, Thomas BL, Cook DJ. Simple and complex activity recognition through smart phones. In: 2012 Eighth International Conference on Intelligent Environments. Guanajuato: IEEE (2012). p. 214–21.
12. Jobanputra C, Bavishi J, Doshi N. Human activity recognition: a survey. Procedia Comput Sci. (2019) 155:698–703. doi: 10.1016/j.procs.2019.08.100
13. Brezmes T, Gorricho JL, Cotrina J. Activity recognition from accelerometer data on a mobile phone. In: Proceedings of the 10th International Work-Conference on Artificial Neural Networks: Part II: Distributed Computing, Artificial Intelligence, Bioinformatics, Soft Computing, and Ambient Assisted Living. IWANN '09. Berlin; Heidelberg: Springer-Verlag;. (2009). p. 796–9.
14. Ermes M, Pärkkâ J, Mäntyjärvi J, Korhonen I. Detection of daily activities and sports with wearable sensors in controlled and uncontrolled conditions. IEEE Trans Inf Technol Biomed. (2008) 12:20–26. doi: 10.1109/TITB.2007.899496
15. Kwapisz JR, Weiss GM, Moore SA. Activity recognition using cell phone accelerometers. ACM SigKDD Explor Newsl. (2011) 12:74–82. doi: 10.1145/1964897.1964918
16. Atallah L, Lo B, King R, Yang GZ. Sensor positioning for activity recognition using wearable accelerometers. IEEE Trans Biomed Circ Syst. (2011) 5:320–9. doi: 10.1109/TBCAS.2011.2160540
17. Khan AM, Lee YK, Lee SY, Kim TS. Human activity recognition via an accelerometer-enabled-smartphone using kernel discriminant analysis. In: 2010 5th International Conference on Future Information Technology. Busan: IEEE (2010). p. 1–6.
18. Balzi G. Activity Tracking Using Ear-Level Accelerometers Through Machine Learning. Technical University of Denmark, Department of Electrical Engineering (2021).
19. Titterton DH, Weston JL. Strapdown inertial navigation technology. In: IEE Radar, Sonar, Navigation and Avionics Series. Stevenage: Peter Peregrinus Ltd. (1997).
20. Bengio Y, Courville A, Vincent P. Representation learning: a review and new perspectives. IEEE Trans Pattern Anal Mach Intell. (2013) 35:1798–828. doi: 10.1109/TPAMI.2013.50
21. Masse LC, Fuemmeler BF, Anderson CB, Matthews CE, Trost SG, Catellier DJ, et al. Accelerometer data reduction: a comparison of four reduction algorithms on select outcome variables. Med Sci Sports Exerc. (2005) 37:S544. doi: 10.1249/01.mss.0000185674.09066.8a
22. Gjoreski H, Lustrek M, Gams M. Accelerometer placement for posture recognition and fall detection. In: 2011 Seventh International Conference on Intelligent Environments. Nottingham: IEEE (2011). p. 47–54.
23. Hua A, Quicksall Z, Di C, Motl R, LaCroix AZ, Schatz B, et al. Accelerometer-based predictive models of fall risk in older women: a pilot study. NPJ Digit Med. (2018) 1:1–8. doi: 10.1038/s41746-018-0033-5
24. Joanes DN, Gill CA. Comparing measures of sample skewness and kurtosis. J R Stat Soc Ser D. (1998) 47:183–9. doi: 10.1111/1467-9884.00122
25. Crouter SE, Clowers KG, Bassett Jr DR. A novel method for using accelerometer data to predict energy expenditure. J Appl Physiol. (2006) 100:1324–31. doi: 10.1152/japplphysiol.00818.2005
26. Brønd JC, Andersen LB, Arvidsson D. Generating ActiGraph counts from raw acceleration recorded by an alternative monitor. Med Sci Sports Exerc. (2017) 49:2351–60. doi: 10.1249/MSS.0000000000001344
27. Bishop C. Pattern Recognition and Machine Learning. Springer (2006). Available online at: https://www.microsoft.com/en-us/research/publication/pattern-recognition-machine-learning/.
28. Fix E, Hodges JL. Nonparametric discrimination: consistency properties. In: USAF School of Aviation Medicine, Randolph Field, Texas Universal City, TX (1951). p. 21–49.
29. Cover T, Hart P. Nearest neighbor pattern classification. IEEE Trans Inf Theory. (1967) 13:21–27. doi: 10.1109/TIT.1967.1053964
30. Fisher RA. The use of multiple measurements in taxonomic problems. Ann Eugen. (1936) 7:179–88. doi: 10.1111/j.1469-1809.1936.tb02137.x
31. Breiman L, Friedman JH, Olshen RA, Stone CJ. Classification and Regression Trees. Boca Raton, FL: Chapman & Hall (1984).
32. Schapire RE. The strength of weak learnability. Mach Learn. (1990) 5:197–227. doi: 10.1007/BF00116037
33. Freund Y, Schapire RE. A decision-theoretic generalization of on-line learning and an application to boosting. J Comput Syst Sci. (1997) 55:119–39. doi: 10.1006/jcss.1997.1504
34. Lindegaard Jensen E. Head Motion Estimation Using Ear-Worn Binaural Accelerometers. Technical University of Denmark, Department of Electrical Engineering (2021).
35. Anguita D, Ghio A, Oneto L, Parra X, Reyes-Ortiz JL. A public domain dataset for human activity recognition using smartphones. In: 21th European Symposium on Artificial Neural Networks, Computational Intelligence and Machine Learning, ESANN. Belgium: Bruges (2013). p. 437–42.
36. Khan AM, Lee YK, Lee SY. Accelerometer's position free human activity recognition using a hierarchical recognition model. In: The 12th IEEE International Conference on e-Health Networking, Applications and Services. Lyon: IEEE (2010). p. 296–301. doi: 10.1109/HEALTH.2010.5556553
37. Ronao CA, Cho SB. Human activity recognition with smartphone sensors using deep learning neural networks. Expert Syst Appl. (2016) 59:235–44. doi: 10.1016/j.eswa.2016.04.032
38. Bai YW, Yu CH, Wu SC. Using a three-axis accelerometer and GPS module in a smart phone to measure walking steps and distance. In: 2014 IEEE 27th Canadian Conference on Electrical and Computer Engineering (CCECE). Toronto, ON: IEEE (2014). p. 1–6.
39. Abadleh A, Al-Hawari E, Alkafaween E, Al-Sawalqah H. Step detection algorithm for accurate distance estimation using dynamic step length. In: 2017 18th IEEE International Conference on Mobile Data Management (MDM). Daejeon: IEEE (2017). p. 324–7.
40. Parkka J, Ermes M, Korpipaa P, Mantyjarvi J, Peltola J, Korhonen I. Activity classification using realistic data from wearable sensors. IEEE Trans Inf Technol Biomed. (2006) 10:119–28. doi: 10.1109/TITB.2005.856863
41. Anik MAI, Hassan M, Mahmud H, Hasan MK. Activity recognition of a badminton game through accelerometer and gyroscope. in 201619th International Conference on Computer and Information Technology (ICCIT). Dhaka: IEEE (2016). p. 213–7.
42. Zhang Y, Markovic S, Sapir I, Wagenaar RC, Little TD. Continuous functional activity monitoring based on wearable tri-axial accelerometer and gyroscope. In: 2011 5th International Conference on Pervasive Computing Technologies for Healthcare (PervasiveHealth) and Workshops. Dublin: IEEE (2011). p. 370–3.
43. Shoaib M, Bosch S, Incel OD, Scholten H, Havinga PJ. Fusion of smartphone motion sensors for physical activity recognition. Sensors. (2014) 14:10146–76. doi: 10.3390/s140610146
Keywords: activity tracking, accelerometer, classification, machine learning, supervised learning, hearing aids, hearing healthcare
Citation: Skoglund MA, Balzi G, Jensen EL, Bhuiyan TA and Rotger-Griful S (2021) Activity Tracking Using Ear-Level Accelerometers. Front. Digit. Health 3:724714. doi: 10.3389/fdgth.2021.724714
Received: 14 June 2021; Accepted: 17 August 2021;
Published: 17 September 2021.
Edited by:
Jing Chen, Peking University, ChinaReviewed by:
Parisis Gallos, National and Kapodistrian University of Athens, GreeceIvan Miguel Pires, Universidade da Beira Interior, Portugal
Copyright © 2021 Skoglund, Balzi, Jensen, Bhuiyan and Rotger-Griful. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Martin A. Skoglund, martin.skoglund@liu.se