- 1Bioengineering Laboratory, Department of Electrical, Computer and Biomedical Engineering, University of Pavia, Pavia, Italy
- 2Custom Computing and Programmable Systems Laboratory, Department of Electrical, Computer and Biomedical Engineering, University of Pavia, Pavia, Italy
Ambient Assisted Living is a concept that focuses on using technology to support and enhance the quality of life and well-being of frail or elderly individuals in both indoor and outdoor environments. It aims at empowering individuals to maintain their independence and autonomy while ensuring their safety and providing assistance when needed. Human Activity Recognition is widely regarded as the most popular methodology within the field of Ambient Assisted Living. Human Activity Recognition involves automatically detecting and classifying the activities performed by individuals using sensor-based systems. Researchers have employed various methodologies, utilizing wearable and/or non-wearable sensors, and employing algorithms ranging from simple threshold-based techniques to more advanced deep learning approaches. In this review, literature from the past decade is critically examined, specifically exploring the technological aspects of Human Activity Recognition in Ambient Assisted Living. An exhaustive analysis of the methodologies adopted, highlighting their strengths and weaknesses is provided. Finally, challenges encountered in the field of Human Activity Recognition for Ambient Assisted Living are thoroughly discussed. These challenges encompass issues related to data collection, model training, real-time performance, generalizability, and user acceptance. Miniaturization, unobtrusiveness, energy harvesting and communication efficiency will be the crucial factors for new wearable solutions.
1. Introduction
Ambient Assisted Living (AAL) refers to the use of Information and Communication Technologies (ICT), assistive devices, and sensor network technologies to support, monitor and enhance the quality of life for individuals, particularly older adults, or people with disabilities, within their daily living and working environment. The primary goal of AAL is to provide individuals with increased independence, autonomy, and safety by incorporating technological solutions into their surroundings. These solutions can assist individuals in various activities of daily living, such as managing their health, monitoring their safety, and improving their social interactions (Blackman et al., 2016; Stodczyk, 2020). One significant aspect of AAL is subject monitoring, which involves the continuous and unobtrusive tracking of an individual’s activities, health parameters, and environment to ensure their safety and provide timely assistance when needed. Subject monitoring utilizes various sensors to collect data and analyze patterns, enabling caregivers and healthcare professionals to gain valuable insights into an individual’s daily routines, health conditions, and potential risks. The choice of sensors can be made among two main groups: wearable and non-wearable sensors. The first one could be incorporated on clothing or worn by the user like accessories. Non-wearable sensors, on the other hand, are strategically placed on furniture, appliances, walls, doors, and other objects throughout the home. By integrating both types of sensors, through a so-called hybrid approach, a comprehensive monitoring solution can be created to effectively and efficiently monitor the subject (Calvaresi et al., 2017; Clapés et al., 2018).
The work of Aleksic et al. proposed a subdivision of AAL systems for subject monitoring into four distinct generations (see Figure 1) based on technological variations, highlighting the application of ICT, stand-alone assistive devices, and technologies for indoor environments within individuals’ daily living and working environments (Al-Shaqi et al., 2016). These AAL systems actively encourage healthy lifestyles, contribute to disease prevention through personalized risk assessment and continuous monitoring, and primarily cater to frail individuals, by offering continuous support and actively promoting their independent and healthy living (Blackman et al., 2016; Calvaresi et al., 2017; Stodczyk, 2020; Cicirelli et al., 2021):
• First Generation of AAL Systems: the first generation of AAL systems primarily consists of alert and alarm systems using pendant or button devices worn by the monitored individuals. In the event of a dangerous situation, the individual would activate the button or pendant to send an alarm signal to a call center or caregiver. Examples of such solutions include the Salvalavita Beghelli1 and LifeAlert.2 While these devices offer several benefits, they also have specific limitations. For instance, individuals may be physically or mentally incapacitated, making them unable to trigger the alarm. Additionally, there are issues with individuals forgetting to wear or recharge the device.
• Second Generation of AAL Systems: the second generation of AAL systems involves more technologically advanced devices, installed in indoor spaces, incorporating sensors capable of automatically detecting dangerous conditions and triggering appropriate responses without relying on user activation. However, a weakness associated with this generation is that some users may perceive it as intrusive.
• Third Generation of AAL Systems: the third generation of AAL systems expands through advancements in ICT, introducing a more comprehensive concept of AAL. These systems encompass sensors designed to detect potentially dangerous situations and proactively prevent adverse scenarios, actuators providing support to the assisted individuals, and smart interfaces delivering information, assistance, and encouragement. The aim is to create minimally intrusive home setups comprising multiple sensors, actuators, and computing systems. These systems not only monitor the home environment but also track vital signs, changes in habits and activity patterns of frail individuals, and facilitate the execution of daily living activities (Mainetti et al., 2016).
• Fourth Generation of AAL Systems: the fourth generation of AAL systems incorporates Artificial Intelligence (AI) algorithms for data analysis within AAL solutions. These intelligent systems have the ability to learn from data and evolve over time, offering personalized assistance and support. The co-design approach is embraced, involving end-users, caregivers and stakeholders to create users-centered and inclusive solutions (Siriwardhana et al., 2019; Bansal et al., 2021; Sophia et al., 2021; Gulati and Kaur, 2022; Rupasinghe and Maduranga, 2022; Warunsin and Phairoh, 2022).
1.1. Related works
By combining AAL with subject monitoring, it becomes possible to create personalized and proactive care solutions, promoting independent living while offering a safety net for vulnerable individuals. Subject monitoring can be conducted in both outdoor and indoor environments. Outdoor environments expose frail people to various risks, such as falls, extreme temperatures, and potential wandering or confusion in individuals with early symptoms of dementia. In this context, AAL systems aim to provide support to frail individuals by facilitating route checking, anomalous behavior recognition, motion activity evaluation, and other relevant functionalities (Fernandes et al., 2020; Lee, 2021). Since wearables are the only devices that can be employed outdoors, they acquire a fundamental importance. Smartwatches and smart wristbands are the most commonly used devices, while Inertial Measurement Units (IMUs) are typically the sensors chosen (the same used in indoor scenarios) (Bhargava et al., 2017; Iadarola et al., 2022). In literature, other solutions have been proposed that address alternative or complementary approaches (Sokullu et al., 2019; Kenfack Ngankam et al., 2020; Rupasinghe and Maduranga, 2022): for example, Global Navigation Satellite System (GNSS) sensors for outdoor localization (Junior et al., 2023); instrumented insoles (commercial or customized) as an aid for gait detection and consequently for pointing out unsteady walking or falls (Cristiani et al., 2014; Sivakumar et al., 2018) and sensors mounted on the body of a walker as a low-cost solution for people with limited mobility (Ding et al., 2022). Conversely, indoor scenarios present frail individuals with risks closely associated with their living spaces.
Notably, the indoor environment has been identified as a significant contributing element to falls (Lee, 2021), which are attributed to factors such as uneven or slippery floor surfaces (including carpets and mats), tripping obstacles, inadequate lighting, poorly designed or maintained stairs without handrails and unsuitable furniture. These criticalities increase the likelihood of tripping, falling, or slipping for frail individuals. Additional hazards arise from the absence of safety or preventive devices, such as night lights and grab rails (Carter et al., 1997; Lee, 2021). Indeed, the requirements of monitored subjects can vary significantly across different indoor scenarios. In private homes, where individuals live alone or with a caregiver, the primary focus of monitoring is on preventing domestic accidents and delivering essential healthcare services. On the other hand, in retirement residences where multiple individuals share common spaces, subject monitoring systems are primarily designed to facilitate group activities and controlled physical exercises (Nastac et al., 2019; Cicirelli et al., 2021).
Adapting the monitoring approach to suit the specific needs and dynamics of each indoor setting is crucial. This ensures that monitored subjects receive personalized support and care tailored to their circumstances. Additionally, it is important to recognize that AAL systems cannot remain static, as people’s needs and habits evolve over time, along with the parameters that need to be observed. Consequently, data analysis methodologies must account for the evolving nature of these systems, allowing for the possibility of dynamically weighting or customizing certain parameters over others (Cicirelli et al., 2021). By embracing flexibility and adaptability, AAL systems can continue to provide effective and relevant support to individuals in various indoor environments.
The selection of appropriate sensors considers multiple factors, including the specific objectives of the AAL system, sensor cost, intrusiveness, user acceptability, and privacy concerns. However, more complex sensor networks, comprising environmental sensors, object sensors, cameras, and wearable sensors constitute the foundation of indoor AAL. The living facilities may be equipped with an array of interconnected sensors and actuators, enabling remote control and capable of detecting various environmental parameters such as door openings and room brightness. These sensors are strategically deployed to monitor the daily activities of individuals, ensuring security and safety. The selection of appropriate sensors considers multiple factors, including the specific objectives of the AAL system, sensor cost, intrusiveness, user acceptability, and privacy concerns. Communication protocols, such as ZigBee, Bluetooth, ZWave, USB, Ethernet, among others, are utilized to interconnect sensors, actuators, and smart devices throughout the environment (Tazari et al., 2011; Lloret et al., 2015; Babangida et al., 2022). Typically, raw or pre-processed data from sensors are transmitted to a collection center, either local or remote, where they undergo integration and analysis using robust algorithms (Plentz and De Pieri, 2018). A thorough and reliable data analysis becomes crucial in indoor scenarios equipped with automatic dangerous situation detection or capable of requesting help triggering alarms to third parties.
In the field of AAL, Human Activity Recognition (HAR) has emerged as a valuable tool with multifaceted utility. Within the AAL domain, HAR presents a range of solutions aimed at enhancing the quality of life of frail individuals (elderly and/or disabled people) and maintaining improved health and independence. Additionally, it provides also support to caregivers and medical professionals. HAR has garnered substantial interest as a prominent field of study in recent times. HAR methodologies are devised with the objective of autonomously detecting and classifying individuals’ routine activities within defined contexts. Depending on the task and the employed technologies, two main methodologies are commonly adopted. The first approach relies on a threshold analysis method, which can suffice for triggering alerts when detecting dangerous events (Zdravevski et al., 2017; Al Machot et al., 2018a,b; Cicirelli et al., 2021). The second and more recent approach (see Figure 1, fourth generation), employs Artificial Intelligence (AI) solutions such as Machine Learning (ML) and Deep Learning (DL) algorithms for HAR (Aggarwal and Ryoo, 2011; Wang J. et al., 2019).
Among the various possible applications, human activities can be classified into four distinct groups based on the involvement of various body parts (Jegham et al., 2020; Minh Dang et al., 2020):
• gestures involve basic actions carried out by different parts of the human body, including hand gestures like the “okay” gesture or “thumbs up” gesture;
• actions refer to a collection of fundamental movements executed by an individual, such as walking, standing, sitting, running, and other similar activities;
• interactions encompass not only activities involving interactions between two individuals, but they can also involve the relationship between a person and an object. Examples of interactions include playing a guitar or hugging another person;
• group of activities are the most intricate category as they primarily involve a combination of gestures, actions, and interactions. Examples of group activities include group meetings or group walking, where multiple individuals engage in coordinated actions and interactions.
Detailed framework of the HAR process will be discussed in the upcoming section. Finally, the aim of this work is to provide an overview of recent literature of AAL systems focusing on HAR from a technological perspective, tackling emerging evidences, challenges and future directions (Stavropoulos et al., 2020).
Papers have been revised by searching published works in the last 10 years in the following databases: IEEE Xplore, PubMed, Scopus. A series of keywords have been used, alone or in combination: ambient assisted living, ambient assisted living technology, healthy ageing, human activity recognition, ambient sensors, wearable sensors, wearable technology, activity monitoring, machine learning, deep learning, frail person.
The rest of the paper is organized as follows: Section 2 describes the Monitoring Solutions, focusing on the main components of an AAL system and detailing wearable and non-wearable sensors. Section 3 presents the processing chain, outlining the most relevant methodologies for data acquisition, processing and analysis. Section 4 proposes a critical discussion that also addresses future directions of this research field.
2. Monitoring solutions
Monitoring solutions in AAL systems can be developed using a variety of technologies, depending on the specific application domains and requirements. Specifically, in the field of HAR different solutions can be employed (Cicirelli et al., 2021; Ranieri et al., 2021). The upcoming session will provide a detailed description of wearable and non-wearable sensor solutions for HAR within the context of AAL in indoor environments.
2.1. Wearable sensors
Wearable devices typically consist of small sensors that can be integrated into cloths, rings, shirts, watches or other garments and devices. Such sensors gather body and context information to be locally processed or directly transmitted, mainly wirelessly through appropriate communication protocols, to a central processing unit of an AAL system (Kumari et al., 2017).
In the last few years, most wearable devices have been miniaturized and have optimized their power consumption (Cicirelli et al., 2021). Wearable devices, especially fitness trackers, wristbands and smartwatches (Andò et al., 2016; Lutze, 2018; Xie et al., 2020), have various built-in/integrated sensors such as accelerometers, gyroscopes and orientation sensors. Smartphones represent an additional solution for their characteristics of cost effectiveness and high number of embedded sensors (Ramanujam et al., 2021). Moreover, smartphones’ embedded sensors can be used alone or in combination with other wearable technology to evaluate posture and activities and to prevent falls, together with biological and behavioral monitoring (Haghi et al., 2019; Badgujar and Pillai, 2020; Nouredanesh et al., 2020). In this review, since users take their smartphone with them almost everywhere, even though they are not always in direct contact with the body, they are considered wearable devices at the same level as wristwatches, rings, glasses and necklaces, as opposed to environmental sensors and cameras. Apps for recording the device’s sensors data can be run on all commercial operating systems (Android or iOS) and they can be combined with commercial smartwatches, self-developed smart bands or devices like Shimmer nodes (Sanchez-Comas et al., 2020). However, not all applications provide an integration with smartphones; instead, many studies considered custom-developed solutions of electronic components. A significant number of works developed technologies in the laboratory, whereas fewer studies used commercial devices. Inertial sensors are the most common wearable elements used for HAR in AAL; in some cases, accelerometers (Tay et al., 2023), gyroscopes, magnetometers, temperature and object sensors may be applied, mainly worn on the waist or the hip (Pierleoni et al., 2019; Sarabia-Jácome et al., 2020).
Identification of the user’s position can be obtained even with passive RFID sensors or Bluetooth Low Energy (BLE) technology, that paved the way to the Bluetooth Smart for wearable devices (Ciabattoni et al., 2019; Paolini et al., 2019; Bilbao-Jayo et al., 2021). Also, wearables usually commercialized for fitness purposes have been evaluated for elderly wellbeing in the AAL domain (Piwek et al., 2016; Seneviratne et al., 2017; Cedillo et al., 2018; Surendran et al., 2018; Stavropoulos et al., 2020).
The majority of HAR systems in AAL are dedicated to the identification and management of falls, as confirmed by literature (Dhiman and Vishwakarma, 2019). In such systems, wearables are one of the key elements due to their mobility, portability, cost and availability and several studies have been conducted, mostly using a single wearable device (Wang X. et al., 2020). Even in this context, inertial sensors represent a large percentage of the research, whereas only a minority deployed other solutions (Bourke et al., 2007). Although Shimmer nodes, smartphones and smart watches often contain sensors like magnetometers, such elements were not normally used to detect falls; indeed, the only sensors embedded in smartphones used for this purpose are accelerometers and gyroscopes (Shi et al., 2016; Islam et al., 2017; Medrano et al., 2017; Chen et al., 2018).
Even combining multiple sensors into a single framework can provide valuable data for meaningful and complex predictions, thus achieving a more versatile, robust and trustworthy wearable system for HAR purposes. Moreover, commercial tools are widely used, such as Samsung Galaxy Gear Live (Faye et al., 2017), Microsoft Band 2 (Garcia-Ceja et al., 2018) and Intel Basis Peak (Kang and Larkin, 2017), as well as other alternatives like Empatica E3 (Clapés et al., 2018), Fitbit (Kang and Larkin, 2017), and Google Glass (Shewell et al., 2017; Clapés et al., 2018).
It is worth noticing that most of the time, wearable technology alone would be sufficient to assess activity recognition in indoor environments and AAL systems. This is an important advantage, combined with their low cost, portability and unobtrusiveness. However, a hybrid approach combining wearable and non-wearable sensors often overcomes possible drawbacks due to users not wearing the device correctly, thus leading to low quality signals, not being comfortable in putting on a wearable that could excessively stand out, forgetting to charge it or even to wear it continuously.
2.2. Non-wearable sensors
Non-wearable sensor solutions for HAR encompass devices or systems that can detect and analyze human activities without needing to be directly attached to the body. These solutions play a crucial role in the functionality and effectiveness of AAL systems. Operating in a passive manner, these sensors autonomously monitor room occupants without the need for manual intervention. This eliminates the need for users to carry additional devices during their daily activities.
One example of a non-wearable sensor solution is radio-frequency-based systems, as demonstrated by (Fan et al., 2020). These systems utilize the analysis of radiofrequency (RF) signal reflections to monitor various activities performed by individuals. In this study, Radio Frequency sensors were embedded directly in the floor to capture the everyday activities of residents (Fan et al., 2020). As part of radiofrequency-based systems, radar and ultra-broadband technologies, as well as automotive-derived solutions, also emerged as interesting approaches for human activity recognition (Guendel et al., 2021; Senigagliesi et al., 2022).
Furthermore, contemporary monitoring and behavior analysis tasks can be facilitated by diverse image-based technologies.
Nowadays, low-cost cameras have emerged as viable options for monitoring individuals’ daily activities, ensuring their well-being (Geman and Costin, 2014; Cebanov et al., 2019; Minh Dang et al., 2020; Ryselis et al., 2020; Bouchabou et al., 2021; Chen et al., 2021; Raeis et al., 2021). These devices enable continuous monitoring of individuals without requiring their active involvement. These include RGB, Depth, and RGB-D cameras, as well as IR (infrared) array sensors, known as thermal cameras.
RGB cameras, being widely available and affordable, provide information about the shape, color, and texture of the scene (Zerrouki et al., 2018). However, they have some drawbacks such as a limited field of view, complex calibration procedures, sensitivity to environmental variations (e.g., lighting conditions, type of illumination and cluttered background) and privacy concerns. To address this latter issue, depth cameras offer distance information from the sensor to elements in the scene, capturing detailed spatial information while maintaining heightened privacy protection (Jayaraj et al., 2019; Beddiar et al., 2020). Depth sensors also exhibit superior resilience to variations in illumination, color, and texture compared to RGB devices. However, noisy measurements can occasionally affect accurate object or subject identification, necessitating data processing and refinement. In recent years, low-cost devices integrating RGB and depth sensors, such as Kinect and Intel RealSense systems, have been employed as environmental sensors in AAL systems. Another alternative, IR array cameras, measure thermal energy emitted by the human body or other objects (Mashiyama et al., 2015; Spasova et al., 2016; Karayaneva et al., 2023). These low-resolution IR arrays offer advantages such as privacy preservation, low power consumption, insensitivity to ambient lighting variations, operation in complete darkness, fast response time, easy deployment, and straightforward image processing.
All these devices suffer from subject occlusion, which occurs when certain body parts of the subject are hidden or obscured by other objects or body parts within the room, leading to incomplete or inaccurate tracking of the subject’s movements. To overcome the occlusion limitation, a practical solution is to employ a multiple camera setup that covers various areas of the room from different viewpoints. By using multiple cameras, the chances of occlusion can be reduced, as different cameras capture different perspectives of the scene. This approach allows for a more comprehensive view of the subject’s activities and improves the accuracy of tracking. However, it is important to note that using multiple cameras increases HAR systems’ costs and requires synchronization among them to ensure proper coordination and alignment of the captured data. Synchronizing the cameras enables the seamless integration of the captured images or depth data, allowing for a more complete understanding of the subject’s movements and activities.
In the context of monitoring human activities, sensors can be also embedded in everyday objects within the environment. Contrary to wearable sensors, which directly measure human activities, these sensors enable the detection of movements and activities through the usage of specific objects, providing valuable insights into the daily lives of individuals (Bassoli et al., 2017; Rafferty et al., 2017). For instance, sensors can be integrated into furniture items such as carpets, beds, fridges, and more, allowing for unobtrusive monitoring of daily living activities. Presence statistics of users in different spaces can be gathered by monitoring the sensors embedded in furniture. Power meters can be employed to track appliance usage (Bianchi et al., 2019), such as monitoring TV sets, while smart pill box devices assist in checking medication intake (Keum et al., 2019). Roland et al. proposed the installation of an accelerometer attached to a smart drinking cup to efficiently identify the user’s drinking movement (Roland et al., 2018). Bassoli et al. installed sensors directly on the furniture of the subject’s house for HAR. Pressure pads are used to monitor bed or chair occupancy, while sensors inside the fridge provide indirect information about feeding habits (Marenzi et al., 2012, 2013a,b; Bassoli et al., 2017). Chaccour et al. developed a smart carpet with piezoresistive pressure sensors to detect falls of the inhabitant (Chaccour et al., 2015; Singh et al., 2020).
Everyday object sensors offer a less invasive and privacy-friendly alternative to cameras, as they are designed to specifically recognize human activities related to the intended use of the object. By focusing on that, these sensors can provide valuable insights and functionality while minimizing potential privacy concerns. Everyday object sensors can detect interactions with household appliances or devices, such as opening a refrigerator, without capturing or storing detailed visual information of individuals. Yue et al. proposed an RF-based system that accurately monitors sleep postures overnight in the user’s own house. By analyzing RF reflections and distinguishing them from other signals, the system can identify different sleeping postures such as supine, left side, right side, and prone (Yue et al., 2020). This approach can help alleviate privacy concerns while still enabling the development of innovative and convenient technologies that enhance user experiences in a responsible and respectful manner.
3. HAR processing chain
Irrespective of the specific type of human activity being classified, the process of HAR typically adheres to a standard framework comprising several distinct phases. These phases are summarized in each block of Figure 2 and detailed in the following paragraphs, keeping the same names and order as in the Figure 2.
3.1. Data acquisition and pre-processing
Sensors and devices are characterized by outputs that are either punctual values (i.e., pixels in an image) or temporal series (i.e., position, acceleration). HAR applications could rely on raw data or need further pre-processing steps to enhance the signal quality (i.e., noise reduction, data normalization, segmentation) and/or to obtain derived data. For instance, from data captured by RGB-D cameras it is possible to estimate the position over time of specific points on the human body, often corresponding to anatomical points of repère or joints with respect to a specific coordinate system typically intrinsic to the device. This data processing stage is called skeleton tracking. Moreover, joint angles can be obtained starting from body joint positions (Jegham et al., 2020). Notably, in the context of AAL, it is common practice to store and analyze the skeleton tracking data instead of RGB or depth data to prioritize privacy preservation (Gasparrini et al., 2014; Tu et al., 2018; Li and Sun, 2021; Srimath et al., 2021). The pre-processing stage aims at noise reduction, data normalization and segmentation.
3.1.1. Noise reduction
Raw, noisy signals adversely affect signal segmentation and possible further computing steps (Deep et al., 2020). Noise can be removed in several ways: linear, averaging low-pass and high-pass filters (Amin et al., 2016). In some specific situations (e.g., inertial data and skeletal data), Kalman filters, moving average or adaptive filters are useful to correct distorted signals (Jeong et al., 2017; Beddiar et al., 2020; Ahad, 2021). The choice of an appropriate filter is critical, since it affects the quality of the filtered signals and of the successive steps of the processing chain. The Signal-to-Noise Ratio (SNR) is a crucial parameter in the identification of the most appropriate filter, together with the correlation coefficient (R) between the filtered and the reference signals, the cut-off frequency, the waveform delay, the filter size and the window length. The choice of the most suitable filter has to be a compromise among all these parameters. The components of the target signal that fall within the stop band of the filter are lost. At the same time filtering a signal introduces a delay (waveform delay), i.e., the output signal is shifted in time with respect to the input. This factor plays a critical role, especially when the responsiveness of the application is mandatory. The filtering of the signal at the same time as its acquisition is important in order not to accumulate delay between the filtered and raw signal. In other words, if the complete processing chain is designed to identify a potentially dangerous situation, the time shift should be compatible with this task, to ensure a prompt detection. Finally, a filter operates by allowing a specific range of frequencies to pass through. For instance, since frequency range of human activity is usually about 0–20 Hz (Antonsson and Mann, 1985; Grossman et al., 1988), the cut-off frequency is usually equal to at least twice such value (Castro et al., 2016; Wang et al., 2017; Ma et al., 2018; Minh Dang et al., 2020).
3.1.2. Data normalization
Data normalization is characterized by scaling or transforming the acquired data. This step is often necessary in HAR scenarios where data originated from different types of sensors and/or from people with various anthropometric characteristics. According to the type of data (e.g., RGB or depth, temporal or skeleton data) several normalization methods can be implemented (Pires et al., 2020; Ahad, 2021). Common normalization methods are min-max, mean, standardization and scaling to unit length. The first method scales the data in their maximum and minimum range: the minimum value is subtracted from each data point and the result is divided by the data range. In the second one the mean of all data samples is subtracted from the data vector, and the result is divided by the difference between the maximum and minimum samples. In the standardization method, the mean value of all data samples is subtracted from the data vector, and the result is divided by the standard deviation value. Finally, the last normalization technique scales all the data with respect to the sum of all elements of the data vector (Patro and Sahu, 2015; Mistry and Inden, 2018; Narkhede, 2019; Ahad, 2021).
When considering skeletal tracking data instead, there are also two other types of normalization methods. The first one is the Bounding-box normalization (referring to the border’s coordinates that enclose the subject), in which all skeleton 3D joints coordinates are normalized using the maximum side-length of the bounding box of the skeleton (Cippitelli et al., 2016; Liu et al., 2020). In the second method data are normalized by dividing the 3D coordinates of the skeleton with respect to the length of a specific body segment (i.e., head, neck, torso and so on) or by the subject height. For example (Cippitelli et al., 2016), scaled joint position dividing each value by the Euclidean distance between the neck and torso joints.
3.1.3. Data segmentation
Data segmentation is strongly related to the type of data. When dealing with temporal sequences, it consists of partitioning the data into time windows. Otherwise, when RGB or depth images are analyzed, the segmentation involves the separation of the selected target subject in the scene from the background. The data subdivision in time windows is principally done to overcome the limitations due to the difference between the duration of the action and the sampling rate imposed by the data acquisition device (Minh Dang et al., 2020; Ahad, 2021). The data segmentation can be categorized into Fixed size Non-overlapping Sliding Window (FNSW), Fixed-size Overlapping Sliding Window (FOSW) or Event-Driven Sliding Window (EDSW) (Bulling et al., 2014; Ahad, 2021). In FNSW and FOSW segmentations, the number of samples included in each window is fixed. The difference is that in FOSW consecutive analysis blocks of the time data are overlapped by the designated percentage of the time record, while in FNSW there is no overlapping. Concerning EDSW, it differs from the others since the number of samples included in the time window is not fixed (i.e., it features a variable window size) (Devanne et al., 2019). Generally, the window size has to always be carefully established to comprehend an adequate number of samples, at the same time avoiding prolonged execution times. When the HAR process is part of a system aiming at monitoring a person in the AAL environment, in which the promptness of the recognition is mandatory, smaller window segmentation is suitable (Buzzelli et al., 2020). It also reduces the complexity and the computational time of the HAR process. The overlapping technique can handle the transition between human activities more accurately, i.e., the transitions between sitting and standing postures, or between walking and running (Torti et al., 2019; Buzzelli et al., 2020; Khan and Ghani, 2021; Guerra et al., 2022, 2023). In case of RGB image, or depth image, the segmentation process is implemented using two different approaches: namely, the background subtraction and the foreground extraction. The first one consists in the extraction of the body silhouette in an image sequence captured from a static camera by comparing each incoming frame with a background model. A crucial step of this technique is to obtain a stable and accurate background model. The second one is recommended when the images are acquired by a moving camera, and it consists in the computation of the difference between consecutive image frames. The foreground extraction is more challenging than the background subtraction because, in addition to the motion of the target object, it also needs to consider the motion of the camera and the change of background (Ke et al., 2013; Babaee et al., 2018; Minh Dang et al., 2020; Mliki et al., 2020).
3.2. Feature extraction and selection
The feature extraction procedure consists in the definition of a set of parameters able to discriminate the activities to be classified. Based on the given data nature and characteristics, the features can be divided into several categories: time-domain, frequency-domain and kinematic features (Dhiman and Vishwakarma, 2019; Sousa Lima et al., 2019; Ahad, 2020, 2021). The time-domain features are usually defined to describe the data amplitude variation and distribution over time (for instance mean, variance and kurtosis). On the other hand, the frequency-domain features show the distribution of signal energy (i.e., Fast Fourier Transform, entropy and power spectral density). Kinematic features include all the characteristics of the subject’s movements, acceleration and posture (joints positions and angles). The kinematic features describe geometric relations between body joints (Müller et al., 2005; Guerra et al., 2020).
In the case of images, usually global and local features are computed. The first ones describe the image frames as a whole, providing different types of information (spatial, temporal, frequency) (Ke et al., 2013). Local features extract information around a set of interest points or describe a selected image region, through techniques like histograms of oriented gradients (Aly and Sayed, 2019).
After feature extraction, the relevant ones are selected to achieve dimensionality reduction by finding the smallest subset of features which efficiently defines the data for the given problem (Blum and Langley, 1997; Chandrashekar and Sahin, 2014; Jindal and Kumar, 2017; Ayesha et al., 2020). It can be accomplished using different methods, such as filter, wrapper, embedded, and the more recent hybrid approach (Blum and Langley, 1997; Minh Dang et al., 2020; Zebari et al., 2020). Filter methods measure the relevance of features using statistical standards for evaluating a subset, they process data before the classification occurs and are independent from the latter. The features are ordered according to the ranking of importance (computed with suitable score metrics) and those below a certain threshold are removed. Among the different algorithms, the most used are: ReliefF, statistical techniques such as Principal Component Analysis, Independent Component Analysis, Neighborhood Component Analysis and Correlation Based filter (Suto et al., 2016; Alzahrani et al., 2019; Siddiqi and Alsirhani, 2022). Wrapper method selects the optimal features subset evaluating alternative sets by running the classification algorithm on the training data. It employs the classifier estimated accuracy as its metric (Bhavan and Aggarwal, 2018; Zebari et al., 2020). The most used iterative algorithms are the Recursive Feature Elimination with Support Vector Machine, the Sequential Feature Selection algorithm and the Genetic Algorithm (Liu and Shao, 2013; Guerra et al., 2022). Compared to the filter method, the wrapper method achieves better performance and higher accuracy, nevertheless it increases computing complexity due to the need to recall the learning process for each feature set considered (Jindal and Kumar, 2017; Zebari et al., 2020). In the embedded method, as the name suggests, the selection occurs within the learning algorithm. The most common are the tree-based algorithms like, for example, the Random Forest and the Decision Tree. Embedded methods can be used in multiclass and regression problems and compared to a wrapper method, it is computationally more effective while retaining similar performance (Negin et al., 2013). Finally, the hybrid approach combines filter and wrapper methods to achieve the benefits of both. Usually, the filter technique is first applied to reduce the search space and then, a wrapper model is used to acquire the best subset (Peng et al., 2010).
3.3. Dataset construction
Dataset construction concerns the process that divides the acquired data into training, validation and test sets. Generally, a set of data is required to train the classification model and a set of validation data is used to evaluate the performance of the model during training epochs for fine tuning the hyperparameters and to estimate if the model does not overfit, i.e., when a statistical model fits exactly against its training data at the expenses of its generalization abilities. Finally, test data, different from those involved in the training set, are used to evaluate the performance of the model (Bouchabou et al., 2021). The data contained in the training, validation and test could be described by labels also called classes. As will be stressed in the next Section, this operation is of crucial importance for the classification algorithms. In HAR tasks, the classes represent the type of activity to be recognized (i.e., walking, sitting, lying down, and so on) (Sathya and Abraham, 2013; Sindhu Meena and Suriya, 2020).
In HAR, three methods have been used to divide the data into training, validation and test set. In the first one, called cross-subject, the subjects are divided into two groups. The data of the first group are used for the training phase, whereas those of the second one are involved in the validation and test phase (Khan and Ghani, 2021). The cross-subject method aims at guiding the learning process of the AI model so that it becomes as robust as possible, in order to adapt it to the heterogeneity of the subjects. The second splitting method is characterized by randomly dividing the whole dataset based on a percentage such as 70-30%, 80-20%, and so on. The larger portion is fed for training the model where the other portion is kept for validation and test (Ahad, 2021). This is the most used splitting criteria in the general problems of classification algorithms and have been reported in HAR with success. Alternatively, when data are acquired by multiple cameras with different points of view, a cross-view method can be used. In this case the data coming from one or more cameras are used for the training phase and those of the remaining ones for the validation and test phase (Wang L. et al., 2020).
3.4. Classification
The most critical step of HAR systems relies on classification algorithms. In the literature, two main categories can be identified, namely threshold-based and AI methods. Furthermore, AI algorithms can be divided into ML and DL techniques. In the following, these methods are analyzed, highlighting the main advantages and issues in their application for HAR data analysis.
3.4.1. Threshold-based methods
Threshold-based methods are the first introduced in the literature and they typically do not need feature extraction and selection. They are based on comparing the acquired values with a pre-defined threshold (Fetzer et al., 2018). If a signal value exceeds the threshold, then the algorithm identifies the targeted situation. More advanced threshold-based algorithms adopt adaptive thresholds (Madhu et al., 2021) or apply it on statistical indicators extracted from the original signals (Cola et al., 2017). Moreover, data fusion is employed when considering multiple sensors. A particular strategy uses only one of them to make a final decision: this is called partial fusion. An example is a fall detection system that employs a tri-axial accelerometer and an RGB-D camera: only when the measured signal exceeds a threshold, the camera is activated to capture the ongoing event (Kwolek and Kepski, 2014). One of the most important advantages of threshold-based algorithms is the low computational complexity. This allows the deployment of these algorithms directly on a small computation unit which typically also manages the data acquisition and pre-processing. Indeed, this solution is widely adopted for wearable devices which do not rely on external centralized processing (Jung et al., 2015; Cola et al., 2017). Concerning non-wearables, the preferred strategy is to send all the acquired data to a central host, which applies the threshold-based algorithm (Andò et al., 2016). The major drawback of these methods is the threshold selection since it depends on the monitored subject. Indeed, inter-subject movements show high variability and even the same person can perform a certain movement in different ways (intra-subject variability) depending on a multitude of factors, such as injuries or illness (Jegham et al., 2020). This affects the classification performance.
3.4.2. Machine learning methods
In the last decade, ML methods have been widely explored for HAR since they can automatically extract high level features and produce more affordable results than threshold-based approaches. In particular, the best results have been achieved by Support Vector Machines (SVMs), Artificial Neural Networks (ANNs), K-Nearest-Neighbours (KNN) and Complex Trees (Oniga and Suto, 2014; Hemmatpour et al., 2017; Su et al., 2017). SVMs and KNN rely on the concept of instances. First, they create sets of example data, in which each set is related to a specific class. Then the distances between the new data and the example data sets are computed. The aim is to find the example set with the minimum distance from the new data. Finally, the class of the minimum distance set is given to the new data. ANNs are based on the structure of biological neural networks. They are composed of elementary computational units, which perform a weighted sum of the inputs and apply a nonlinear function. These are organized into layers as Multi-Layer Perceptrons (MLP) and are used to map input data into output classes. Finally, Complex Trees build a decision-making diagram with a tree structure. The tree structure is based on the attribute values of the input data. The classification is obtained following the tree structure until a leaf is reached. Later, single algorithms have been combined in the so-called ensemble learning with different strategies such as boosting, stacking, bagging and majority voting to enhance the classification quality (Hasan et al., 2022).
These algorithms have been used both for data acquired by wearable and non-wearables devices. In both cases, the data are acquired and sent to a central unit for the classification step (Sheikh and Jilani, 2023). Therefore, the main drawback is related to data transmission since the communication rate should be high enough to guarantee continuous monitoring. Moreover, this represents a critical issue for wearable devices since wireless communication is the main power consuming process.
Concerning classification performance, it is affected by the quality of the dataset. AI methods need to be trained on significant examples, which should be balanced among the different classes conceived in the target application. It is worth noticing that an unbalanced training set negatively affects the model performance. A recent trend is to exploit data augmentation strategies to create synthetic data both to increase the training set size and to balance it (Um et al., 2017; Mathur et al., 2018; Steven Eyobu and Han, 2018). The main drawback is related to the choice of the augmentation techniques, since synthetic data can differ too much from the real ones.
3.4.3. Deep learning methods
DL emerged in recent years as the most powerful AI tool to automatically extract high level features and perform affordable classifications. The development of DL models has been enabled by the computing power offered by the technological evolution of devices such as multi-core processors and graphic processing units. Among the different DL methods used in HAR Convolutional Neural Networks (CNNs), Long Short-Term Memory (LSTM) cells and Gated Recurrent Units (GRUs) (Torti et al., 2019; Ronald et al., 2021; Poulose et al., 2022; Guerra et al., 2023; Sonawani and Patil, 2023) emerged as suitable solutions. CNNs are mostly used to deal with visual data, since they roughly imitate human vision. The data is processed as a grid-like topology through the convolutional operator. They can be used both for images and for time series. On the other hand, LSTM and GRUs are mainly used for time series analysis. In fact, their main feature is to learn time dependencies.
These methods have been initially used for data coming from non-wearable devices, especially when vision-based sensors were employed. However, they are gaining popularity also for wearable devices (Torti et al., 2019; Goh et al., 2021; Luwe et al., 2022).
3.5. Computing power constraints
An important aspect related to HAR for frail people is the time needed by the AAL system to detect possible dangerous situations. This time is strictly related to the computational complexity of the adopted algorithms and to the computing power of the devices which perform the processing chain.
Threshold-based methods are the lowest computing power demanding techniques, since they are based on simple comparisons with a fixed value. Even in the case of threshold computed at runtime and/or applied to statistical indicators, the computational complexity can be easily managed by standard microcontrollers and does not have a critical impact on processing times.
AI methods are characterized by a higher computational complexity than threshold-based techniques. In particular, DL methods feature the highest computational complexity, not only for their training, which is run off-line and may require multiple CPU and/or GPU, but also for their implementation as a classifier once all the weights are determined. Therefore, their processing is typically performed by a centralized unit for both wearable and non-wearable devices. However, recently some researchers have designed DL algorithms on low power devices suitable for wearable applications.
Recent and very complex solutions exploited ensemble learning also with DL algorithms (Kumar and Suresh, 2023) enhancing the classification performance of single techniques.
The increasing popularity of AI solutions has led to the development of various software tools, libraries, and frameworks that facilitate the implementation of these algorithms on devices with limited resources. An example of such a tool is TensorFlow Lite for microcontroller3 (TFLM), an open-source library designed to enable the implementation of AI methods (i.e., ANNs, LSTM, CNN and so on) on a wide variety of MCUs and Digital Signal Processors (DSPs). TFLM allows the execution of pre-trained algorithms developed using TensorFlow for on-device inference. Another prominent solution is X-Cube-AI, a software tool developed by STMicroelectronics.4 It offers a comprehensive environment for generating and optimizing AI algorithms developed using popular ML and AI frameworks such as TensorFlow,5 Keras,6 or PyTorch.7 X-Cube-AI is tailored for deployment on the STM32 family of MCUs, empowering developers to leverage familiar frameworks and simplify the integration of AI capabilities into their applications. In addition to these tools, there are cloud-based platforms like Edge Impulse, which provides a flexible environment for the development of AI models. Edge Impulse8 supports various embedded platforms, including MCUs and smartphones, enabling developers to create AI models that cater to diverse hardware constraints. NanoEdge AI Studio9 is another valuable tool that supports both learning and inference directly inside the MCUs. Notably, it offers the advantage of automatically selecting the most suitable machine learning libraries based on the provided data (Shumba et al., 2022).
Inference needs to be performed under real-time constraint, especially when a potentially dangerous condition needs to be identified. This means that the classification should be strictly performed prior to a fixed temporal deadline, which is defined by the acquisition time window. In other words, the system acquires a window of data and its pre-processing and classification should be performed before the end of acquisition of the following time window. The computational complexity of the algorithms and the computing units included in the system play a critical role in the real-time compliance (Avram and Pop, 2023; de la Cal et al., 2023; Gonçalves et al., 2023; Saliba et al., 2023; Zeng et al., 2023). Delayed or sluggish processing can hinder the effectiveness of AAL systems in providing timely assistance, which is crucial for ensuring the safety and well-being of individuals. Efficient algorithms and optimized implementations are necessary to overcome these constraints and enable real-time processing on resource-limited platforms. The computational unit controls the sampling and acquisition of data: usually this is performed by Commercial-Off-The-Shelf (COTS) low power and low cost microcontrollers, supporting interfaces and data transfer protocols (SPI, I2C…). When high speed, flexibility and control over the elements of the architecture are required, Field Programmable Gate Arrays (FPGAs) and Application-Specific Integrated Circuits (ASICs) can be used, with increased costs and higher time for development and/or production. The computational unit is equipped with initial signal conditioning and processing algorithms together with specific classification methods that perform data analysis for local decision making, real-time response and forwarding processed data to successive layers. Characteristics that can influence all these choices include power consumption, computational and storage capacity, complexity and results accuracy of the algorithms, privacy concerns and latency requirements. Pre-trained models using computationally efficient algorithms may be used for anomaly detection and the results can produce warnings or propose a course of action. Alternatively, a fraction of an ANN can perform partial on-device data processing, to forward only intermediate data, thus ensuring also the privacy of the user (Zhang et al., 2023). Lastly, after the elaboration and analysis of data, results, inferences, or alarms are packaged and forwarded to the upper layers for further processing or management (Shumba et al., 2022).
4. Discussion
This review aims at providing an overview on the application of HAR process in the context of AAL systems, underlining their potential to support independent living for frail individuals.
Recent advances in AAL technologies and the reduced cost of sensors have encouraged the development of technological environments to enable frail people to live healthier and more independent lives and to support caregivers, medical personnel, thus limiting hospitalization, promoting personalized therapy and enhancing wellbeing. To provide such services, an AAL system has to be able to understand the daily activities of its residents. In general, the choice of technologies used for HAR purposes can include wearable sensors (IMUs, smartwatches, smartphones…), non-wearable sensors (environmental devices, objects, and cameras), or a combination of the two (hybrid approach). The sensor selection can be made especially depending on the individual’s needs. Wearable sensors have numerous advantages, including their small size, low power consumption, direct acquisition of information on the subject’s activity and full respect of the subject’s privacy. At the same time, they also have some drawbacks. For example, they need to be worn by the subjects and to operate for long time periods. The latter could represent a significant problem for the monitored subject and for the battery life of the devices. Also, to fully capture the 3D motion associated with a human action, a single sensor may not be adequate. It may be necessary to utilize multiple configurations, thus increasing the intrusiveness of the devices worn by the subject (Wang Y. et al., 2019; Beddiar et al., 2020; Qiu et al., 2022). On the other hand, environmental and camera sensors offer the key advantage of being unobtrusive, as they do not require the individual to wear them. However, they also face certain challenges. One major issue is their reliance on infrastructure, which can limit their effectiveness in identifying specific movements or activities. Additionally, the utilization of environmental sensors is less frequent compared to wearable sensors due to higher costs and setup difficulties. Furthermore, similar to the wearable approach, this solution may not always be feasible as it requires users to interact with tagged objects or to remain within the environment where the sensors are installed. In particular, cameras suffer from drawbacks like occlusion, multiple-view variations and privacy issues (Minh Dang et al., 2020). A possible solution to handle occlusion relies on a multiple camera setup, even though these devices need to be synchronized among each other. Cameras are often perceived as the most intrusive technologies in terms of the privacy of the monitored individuals. The solution to this drawback may be RGB-D cameras, like the Kinect V2, which, through a data elaboration, are able to extract the “skeleton” of the subject from the depth image, i.e., the subject is represented as a set of body segments and joints, bypassing the need for using the RGB image for HAR purposes. These tools increase the person’s acceptance towards the assistive technology, since they ensure privacy preservation (Gasparrini et al., 2014). Among non-wearable devices, object sensors are the least invasive and the most respectful of users’ privacy, as they focus on recognizing human activities related only to the intended use of the object. For example, a smart cup recognizes drinking actions, and sensors embedded in cushions or beds identify specific sleeping postures.
To overcome the wearable and non-wearable limits a possible solution could be represented by a hybrid approach. Sensor fusion provides a more robust approach since multiple sensors may complement each other with their specific signals, producing a reliable system (Wang X. et al., 2020). Moreover, the likelihood of having missing data is progressively balanced out by increasing the number of sensors in the system. However, critical issues in AAL services stem from the integration of multiple technological types, mainly environmental and wearable sensors (Calvaresi et al., 2017). For instance, hybrid systems are characterized by sensors with different sampling frequencies. Thus, synchronization and interpolation of acquired data is mandatory for better correlation of output devices. Furthermore, some challenges can still be present; for example, security and privacy requirements need to be taken into consideration.
Independently from the adopted monitored system, HAR requires affordable processing chains to classify the target human activities. Typically, HAR should be performed meeting the real-time constraint, especially when frail people are monitored. While threshold algorithms are efficient for real-time processing, they may struggle with handling complex activities or adapting to dynamic contexts. These limits are overcome by AI methods, at the price of an increased computational complexity, which negatively impacts on classification time. For this reason, the optimal solution requires a suitable trade-off between classification quality and processing time to ensure real-time compliance. Common factors that determine real-time compliancy of a method are the computational complexity and the processing power of the system. It must be stressed that the computational complexity alone is not sufficient to determine if a method is real-time compliant. Indeed, the response time strictly depends on the processing power of the device. It is not trivial that very different processing units perform the same computation with significantly different times. Therefore, the choice of a suitable processing element covers a critical role in the real-time compliancy of a system. Moreover, datasets derived from real situations are not always available or sufficiently precise; in some cases, only simulated conditions are present, but this greatly compromises the results (Casilari and Silva, 2022). At the same time, prompt detection of dangerous conditions cannot be provided in some contexts.
Table 1 summarizes the type of activity, sensors, input data, datasets, main approaches and potential applicability in a real-life scenario, considered in the most relevant works in the field of Human Activity Recognition, already cited in the previous sections. It categorizes the papers into three main classes on the basis of their adopted technologies: wearable, hybrid and non-wearable solutions. In each row, among the labelled information previously reported, the type of activity is related to the target application of each proposed HAR system. Indeed, most of the works aim at recognizing falls and/or Activities of Daily Living (ADL), i.e., lying down, walking, sitting and so on. The input data type is labelled “Dynamic” in case of time series or “Static” otherwise. Concerning the datasets, publicly available ones present the corresponding reference paper, whereas custom Datasets report the number of involved subjects and the amount of performed activities (if one or more parameters are not available, the acronym N.A. is indicated). The “Real life scenario” column is related to the system applicability outside the laboratory setting. It is labelled with “YES,” in case of data acquisitions performed in an environment which considered not only ideal conditions (i.e., the subject is always directly in front of the camera or the falls are simulated only by young people).
5. Conclusion
In conclusion, it is significant to look toward the future of AAL systems, giving importance to HAR indoors and outdoors. The outdoor environment offers numerous activities that can contribute to preventing functional decline in frail individuals. However, monitoring outdoor activities presents different challenges, as non-wearable sensors may not be suitable and safety risks are increased. Therefore, wearables become crucial in this context. While user acceptance can sometimes be a challenging requirement to meet, the ease of use and unobtrusiveness of wearables greatly overcome this drawback. Wearable sensors enable continuous monitoring of indoor and outdoor activities, allowing for a more comprehensive assessment of an individual’s daily life. By incorporating wearable technology into AAL systems, it becomes possible to gather valuable data and insights to support preventive measures and promote healthy lifestyles among frail individuals.
As research and technological advancements continue, it is important to explore and optimize the use of wearable sensors in AAL systems, considering the specific constraints and requirements posed by outdoor monitoring. Future research trends in wearables design could be an enhanced miniaturization of the sensors used nowadays leading to better unobtrusiveness and the possibility of integrating these sensors inside even smaller devices or directly into clothes. An example is the fast development of Micro Electro-Mechanical Systems (MEMS) which enabled the optimization of several sensors-based applications. Moreover, power efficiency will represent a crucial issue, since it impacts both on communication and battery life. Improvements on battery technology should also be coupled with energy harvesting methods to partially recharge the device. On the other hand, the technological evolution of processors and microcontrollers could enable the adoption of state of the art classification methods, overcoming the actual limitations on the models size related to the available amount of memory and of computational power. Finally, communication technologies such as 6G could represent the ideal technology to transmit data between acquisition points and data collection centers. By doing so, the effectiveness of these systems can be enhanced in supporting independent living, improving safety, and preventing functional decline in the target population.
Author contributions
BG: Conceptualization, Investigation, Writing - original draft. ET: Conceptualization, Investigation, Writing – original draft. EM: Conceptualization, Investigation, Writing – original draft. MS: Writing – review & editing. SR: Writing – review & editing. FL: Writing – review & editing. GD: Writing – review & editing.
Funding
The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declared that they were an editorial board member of Frontiers, at the time of submission. This had no impact on the peer review process and the final decision.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Footnotes
1. ^https://www.beghelli.it/salvalavita
3. ^https://www.tensorflow.org/lite/microcontrollers
4. ^https://www.st.com/content/st_com/en.html
5. ^https://www.tensorflow.org
8. ^https://www.edgeimpulse.com
9. ^https://www.st.com/en/development-tools/nanoedgeaistudio.html
References
Aggarwal, J. K., and Ryoo, M. S. (2011). Human activity analysis. ACM Comput. Surv. 43, 1–43. doi: 10.1145/1922649.1922653
Ahad, M. A. R. (2020). “Vision and sensor-based human activity recognition: challenges ahead” in Advancements in instrumentation and control in applied system applications (Pennsylvania: IGI Global), 17–35.
Al Machot, F., Mosa, A. H., Ali, M., and Kyamakya, K. (2018a). Activity recognition in sensor data streams for active and assisted living environments. IEEE Trans. Circuits Syst. Video Technol. 28, 2933–2945. doi: 10.1109/TCSVT.2017.2764868
Al Machot, F., Ranasinghe, S., Plattner, J., and Jnoub, N. (2018b). Human activity recognition based on real life scenarios. 2018 IEEE international conference on pervasive computing and communications workshops (PerCom Workshops), Athens, Greece 3–8.
Al-Shaqi, R., Mourshed, M., and Rezgui, Y. (2016). Progress in ambient assisted systems for independent living by the elderly. Springerplus 5:624. doi: 10.1186/s40064-016-2272-8
Aly, S., and Sayed, A. (2019). Human action recognition using bag of global and local Zernike moment features. Multimed. Tools Appl. 78, 24923–24953. doi: 10.1007/s11042-019-7674-5
Alzahrani, M. S., Jarraya, S. K., Ben-Abdallah, H., and Ali, M. S. (2019). Comprehensive evaluation of skeleton features-based fall detection from Microsoft Kinect v2. SIViP 13, 1431–1439. doi: 10.1007/s11760-019-01490-9
Amin, M. G., Zhang, Y. D., Ahmad, F., and Ho, K. C. D. (2016). Radar signal processing for elderly fall detection: the future for in-home monitoring. IEEE Signal Process. Mag. 33, 71–80. doi: 10.1109/MSP.2015.2502784
Andò, B., Baglio, S., Lombardo, C. O., and Marletta, V. (2016). A multisensor data-fusion approach for ADL and fall classification. IEEE Trans. Instrum. Meas. 65, 1960–1967. doi: 10.1109/TIM.2016.2552678
Antonsson, E. K., and Mann, R. W. (1985). The frequency content of gait. J. Biomech. 18, 39–47. doi: 10.1016/0021-9290(85)90043-0
Avram, M.-R., and Pop, F. (2023). Real-time running workouts monitoring using Cloud–Edge computing. Neural Comput. Applic. 35, 13803–13822. doi: 10.1007/s00521-021-06675-3
Ayesha, S., Hanif, M. K., and Talib, R. (2020). Overview and comparative study of dimensionality reduction techniques for high dimensional data. Informat. Fusion 59, 44–58. doi: 10.1016/j.inffus.2020.01.005
Aziz, O., Musngi, M., Park, E. J., Mori, G., and Robinovitch, S. N. (2017). A comparison of accuracy of fall detection algorithms (threshold-based vs. machine learning) using waist-mounted tri-axial accelerometer signals from a comprehensive set of falls and non-fall trials. Med. Biol. Eng. Comput. 55, 45–55. doi: 10.1007/s11517-016-1504-y
Babaee, M., Dinh, D. T., and Rigoll, G. (2018). A deep convolutional neural network for video sequence background subtraction. Pattern Recogn. 76, 635–649. doi: 10.1016/j.patcog.2017.09.040
Babangida, L., Perumal, T., Mustapha, N., and Yaakob, R. (2022). Internet of things (IoT) based activity recognition strategies in smart homes: a review. IEEE Sensors J. 22, 8327–8336. doi: 10.1109/JSEN.2022.3161797
Badgujar, S., and Pillai, A. S. (2020). Fall detection for elderly people using machine learning. 2020 11th international conference on computing, communication and networking technologies (ICCCNT), Kharagpur, India 1–4.
Bansal, D., Khanna, K., Chhikara, R., Dua, R. K., and Malhotra, R. (2021). “A systematic literature review of deep learning for detecting dementia” in Proceedings of the second international conference on information management and machine intelligence. Lecture notes in networks and systems. eds. D. Goyal, A. K. Gupta, V. Piuri, M. Ganzha, and M. Paprzycki, vol. 166 (Singapore: Springer)
Bassoli, M., Bianchi, V., De Munari, I., and Ciampolini, P. (2017). An IoT approach for an AAL Wi-fi-based monitoring system. IEEE Trans. Instrum. Meas. 66, 3200–3209. doi: 10.1109/TIM.2017.2753458
Beddiar, D. R., Nini, B., Sabokrou, M., and Hadid, A. (2020). Vision-based human activity recognition: a survey. Multimed. Tools Appl. 79, 30509–30555. doi: 10.1007/s11042-020-09004-3
Bhargava, K., McManus, G., and Ivanov, S. (2017). Fog-centric localization for ambient assisted living. 2017 international conference on engineering, technology and innovation (ICE/ITMC), 1424–1430. Madeira, Portugal
Bhavan, A., and Aggarwal, S. (2018). Stacked generalization with wrapper-based feature selection for human activity recognition. 2018 IEEE symposium series on computational intelligence (SSCI), 1064–1068. Bangalore, India
Bianchi, V., Ciampolini, P., and De Munari, I. (2019). RSSI-based indoor localization and identification for ZigBee wireless sensor networks in smart homes. IEEE Trans. Instrum. Meas. 68, 566–575. doi: 10.1109/TIM.2018.2851675
Bilbao-Jayo, A., Almeida, A., Sergi, I., Montanaro, T., Fasano, L., Emaldi, M., et al. (2021). Behavior modeling for a Beacon-based indoor location system. Sensors 21:4839. doi: 10.3390/s21144839
Blackman, S., Matlo, C., Bobrovitskiy, C., Waldoch, A., Fang, M. L., Jackson, P., et al. (2016). Ambient assisted living technologies for aging well: a scoping review. J. Intell. Syst. 25, 55–69. doi: 10.1515/jisys-2014-0136
Blum, A. L., and Langley, P. (1997). Selection of relevant features and examples in machine learning. Artif. Intell. 97, 245–271. doi: 10.1016/S0004-3702(97)00063-5
Bouchabou, D., Nguyen, S. M., Lohr, C., LeDuc, B., and Kanellos, I. (2021). A survey of human activity recognition in smart homes based on IoT sensors algorithms: taxonomies, challenges, and opportunities with deep learning. Sensors 21:6037. doi: 10.3390/s21186037
Bourke, A. K., O’Brien, J. V., and Lyons, G. M. (2007). Evaluation of a threshold-based tri-axial accelerometer fall detection algorithm. Gait Posture 26, 194–199. doi: 10.1016/j.gaitpost.2006.09.012
Bulling, A., Blanke, U., and Schiele, B. (2014). A tutorial on human activity recognition using body-worn inertial sensors. ACM Comput. Surv. 46, 1–33. doi: 10.1145/2499621
Buzzelli, M., Albé, A., and Ciocca, G. (2020). A vision-based system for monitoring elderly people at home. Appl. Sci. 10:374. doi: 10.3390/app10010374
Calvaresi, D., Cesarini, D., Sernani, P., Marinoni, M., Dragoni, A. F., and Sturm, A. (2017). Exploring the ambient assisted living domain: a systematic review. J. Ambient. Intell. Humaniz. Comput. 8, 239–257. doi: 10.1007/s12652-016-0374-3
Carter, S. E., Campbell, E. M., Sanson-Fisher, R. W., Redman, S., and Gillespie, W. J. (1997). Environmental hazards in the homes of older people. Age Ageing 26, 195–202. doi: 10.1093/ageing/26.3.195
Casilari, E., Santoyo-Ramón, J. A., and Cano-García, J. M. (2017). UMAFall: a multisensor dataset for the research on automatic fall detection. Proc. Comput. Sci. 110, 32–39. doi: 10.1016/j.procs.2017.06.110
Casilari, E., and Silva, C. A. (2022). An analytical comparison of datasets of real-world and simulated falls intended for the evaluation of wearable fall alerting systems. Measurement 202:111843. doi: 10.1016/j.measurement.2022.111843
Castro, H. F., Correia, V., Sowade, E., Mitra, K. Y., Rocha, J. G., Baumann, R. R., et al. (2016). All-inkjet-printed low-pass filters with adjustable cutoff frequency consisting of resistors, inductors and transistors for sensor applications. Org. Electron. 38, 205–212. doi: 10.1016/j.orgel.2016.08.025
Cebanov, E., Dobre, C., Gradinaru, A., Ciobanu, R.-I., and Stanciu, V.-D. (2019). Activity recognition for ambient assisted living using off-the-shelf motion sensing input devices. 2019 Global IoT Summit (GIoTS), Aarhus, Denmark 1–6.
Cedillo, P., Sanchez, C., Campos, K., and Bermeo, A. (2018). A Systematic Literature Review on Devices and Systems for Ambient Assisted Living: Solutions and Trends from Different User Perspectives. 2018 International Conference on eDemocracy and eGovernment (ICEDEG). 59–66. doi: 10.1109/ICEDEG.2018.8372367
Chaccour, K., Darazi, R., Hajjam el Hassans, A., and Andres, E. (2015). “Smart carpet using differential piezoresistive pressure sensors for elderly fall detection” in 2015 IEEE 11th international conference on wireless and mobile computing, networking and communications (WiMob), 225–229. doi: 10.1109/WiMOB.2015.7347965
Chandrashekar, G., and Sahin, F. (2014). A survey on feature selection methods. Comput. Electr. Engin. 40, 16–28. doi: 10.1016/j.compeleceng.2013.11.024
Chen, K.-H., Hsu, Y.-W., Yang, J.-J., and Jaw, F.-S. (2018). Evaluating the specifications of built-in accelerometers in smartphones on fall detection performance. Instrumentat. Sci. Technol. 46, 194–206. doi: 10.1080/10739149.2017.1363054
Chen, J., Huang, X., Jiang, H., and Miao, X. (2021). Low-cost and device-free human activity recognition based on hierarchical learning model. Sensors 21:2359. doi: 10.3390/s21072359
Chen, C., Jafari, R., and Kehtarnavaz, N. (2015). UTD-MHAD: a multimodal dataset for human action recognition utilizing a depth camera and a wearable inertial sensor. 2015 IEEE international conference on image processing (ICIP), 168–172 Quebec City, QC
Ciabattoni, L., Foresi, G., Monteriù, A., Pepa, L., Pagnotta, D. P., Spalazzi, L., et al. (2019). Real time indoor localization integrating a model based pedestrian dead reckoning on smartphone and BLE beacons. J. Ambient. Intell. Humaniz. Comput. 10, 1–12. doi: 10.1007/s12652-017-0579-0
Cicirelli, G., Marani, R., Petitti, A., Milella, A., and D’Orazio, T. (2021). Ambient assisted living: a review of technologies, methodologies and future perspectives for healthy aging of population. Sensors 21:3549. doi: 10.3390/s21103549
Cippitelli, E., Gasparrini, S., Gambi, E., and Spinsante, S. (2016). A human activity Recognition system using skeleton data from RGBD sensors. Comput. Intell. Neurosci. 2016, 1–14. doi: 10.1155/2016/4351435
Clapés, A., Pardo, À., Pujol Vila, O., and Escalera, S. (2018). Action detection fusing multiple Kinects and a WIMU: an application to in-home assistive technology for the elderly. Mach. Vis. Appl. 29, 765–788. doi: 10.1007/s00138-018-0931-1
Cola, G., Avvenuti, M., Piazza, P., and Vecchio, A. (2017). Fall detection using a head-worn barometer. Proceedings of the international conference on wireless mobile communication and healthcare (MCH’ 17), Springer, Cham (pp. 217–224)
Cristiani, A. M., Bertolotti, G. M., Marenzi, E., and Ramat, S. (2014). An instrumented insole for long term monitoring movement, comfort, and ergonomics. IEEE Sensors J. 14, 1564–1572. doi: 10.1109/JSEN.2014.2299063
Das, S., Dai, R., Koperski, M., Minciullo, L., Garattoni, L., Bremond, F., et al. (2019). Toyota smarthome: real-world activities of daily living. 2019 IEEE/CVF international conference on computer vision (ICCV), Seoul, Korea 833–842.
de la Cal, E. A., Fáñez, M., Villar, M., Villar, J. R., and González, V. M. (2023). A low-power HAR method for fall and high-intensity ADLs identification using wrist-worn accelerometer devices. Logic J. IGPL 31, 375–389. doi: 10.1093/jigpal/jzac025
Deep, S., Zheng, X., Karmakar, C., Yu, D., Hamey, L. G. C., and Jin, J. (2020). A survey on anomalous behavior detection for elderly care using dense-sensing networks. IEEE Communicat. Surv. Tutor. 22, 352–370. doi: 10.1109/COMST.2019.2948204
Devanne, M., Papadakis, P., and Nguyen, S. M. (2019). Recognition of activities of daily living via hierarchical long-short term memory networks. 2019 IEEE international conference on systems, man and cybernetics (SMC), Bari, Italy 3318–3324
Dhiman, C., and Vishwakarma, D. K. (2019). A review of state-of-the-art techniques for abnormal human activity recognition. Eng. Appl. Artif. Intell. 77, 21–45. doi: 10.1016/j.engappai.2018.08.014
Ding, D.-M., Wang, Y.-G., Zhang, W., and Chen, Q. (2022). Fall detection system on smart Walker based on multisensor data fusion and SPRT method. IEEE Access 10, 80932–80948. doi: 10.1109/ACCESS.2022.3195674
Fan, L., Li, T., Yuan, Y., and Katabi, D. (2020). In-home daily-life captioning using radio signals. Computer vision – ECCV 2020: 16th European conference, Glasgow, UK, 105–123.
Faye, S., Bronzi, W., Tahirou, I., and Engel, T. (2017). Characterizing user mobility using mobile sensing systems. Int. J. Distrib. Sens. Netw. 13:155014771772631. doi: 10.1177/1550147717726310
Fernandes, C. D., Depari, A., Sisinni, E., Ferrari, P., Flammini, A., Rinaldi, S., et al. (2020). Hybrid indoor and outdoor localization for elderly care applications with LoRaWAN. 2020 IEEE international symposium on medical measurements and applications (MeMeA), 1–6. Bari, Italy
Fetzer, T., Ebner, F., Bullmann, M., Deinzer, F., and Grzegorzek, M. (2018). Smartphone-based indoor localization within a 13th century historic building. Sensors 18:4095. doi: 10.3390/s18124095
Gaglio, S., Re, G. L., and Morana, M. (2015). Human activity Recognition process using 3-D posture data. IEEE Trans. Hum.-Mach. Syst. 45, 586–597. doi: 10.1109/THMS.2014.2377111
Garcia-Ceja, E., Galván-Tejada, C. E., and Brena, R. (2018). Multi-view stacking for activity recognition with sound and accelerometer data. Inf. Fusion 40, 45–56. doi: 10.1016/j.inffus.2017.06.004
Garcia-Gonzalez, D., Rivero, D., Fernandez-Blanco, E., and Luaces, M. R. (2020). A public domain dataset for real-life human activity recognition using smartphone sensors. Sensors 20:2200. doi: 10.3390/s20082200
Gasparrini, S., Cippitelli, E., Spinsante, S., and Gambi, E. (2014). Depth cameras in AAL environments: technology and real-world applications. Assistive technologies for physical and cognitive disabilities Pennsylvania, United States: IGI Global 22–41.
Geman, O., and Costin, H. (2014). Automatic assessing of tremor severity using nonlinear dynamics, artificial neural networks and neuro-fuzzy classifier. Adv. Electr. Comput. Eng. 14, 133–138. doi: 10.4316/AECE.2014.01020
Goh, J. X., Lim, K. M., and Lee, C. P. (2021). 1D convolutional neural network with long short-term memory for human activity recognition. 2021 IEEE international conference on artificial intelligence in engineering and technology (IICAIET), Kota Kinabalu, Malaysia. 1–6.
Gonçalves, C., Lopes, J. M., Moccia, S., Berardini, D., Migliorelli, L., and Santos, C. P. (2023). Deep learning-based approaches for human motion decoding in smart walkers for rehabilitation. Expert Syst. Appl. 228:120288. doi: 10.1016/j.eswa.2023.120288
Grossman, G. E., Leigh, R. J., Abel, L. A., Lanska, D. J., and Thurston, S. E. (1988). Frequency and velocity of rotational head perturbations during locomotion. Exp. Brain Res. 70, 470–476. doi: 10.1007/BF00247595
Guendel, R. G., Unterhorst, M., Gambi, E., Fioranelli, F., and Yarovoy, A. (2021). Continuous human activity recognition for arbitrary directions with distributed radars. 2021 IEEE radar conference (RadarConf21). Atlanta, GA, USA. 1–6.
Guerra, B. M. V., Ramat, S., Beltrami, G., and Schmid, M. (2020). Automatic pose Recognition for monitoring dangerous situations in ambient-assisted living. Front. Bioeng. Biotechnol. 8:415. doi: 10.3389/fbioe.2020.00415
Guerra, B. M. V., Ramat, S., Beltrami, G., and Schmid, M. (2023). Recurrent network solutions for human posture recognition based on Kinect skeletal data. Sensors 23:5260. doi: 10.3390/s23115260
Guerra, B. M. V., Schmid, M., Beltrami, G., and Ramat, S. (2022). Neural networks for automatic posture recognition in ambient-assisted living. Sensors 22:2609. doi: 10.3390/s22072609
Gulati, N., and Kaur, P. D. (2022). FriendCare-AAL: a robust social IoT based alert generation system for ambient assisted living. J. Ambient. Intell. Humaniz. Comput. 13, 1735–1762. doi: 10.1007/s12652-021-03236-3
Haghi, M., Geissler, A., Fleischer, H., Stoll, N., and Thurow, K. (2019). Ubiqsense: a personal wearable in ambient parameters monitoring based on IoT platform. 2019 international conference on sensing and instrumentation in IoT era (ISSI), Lisbon, Portugal. 1–6
Hasan, T., Bin Karim, M. F., Mahadi, M. K., Nishat, M. M., and Faisal, F. (2022). Employment of ensemble machine learning methods for human activity Recognition. J. Healthcare Engin. 2022, 1–18. doi: 10.1155/2022/6963891
Hemmatpour, M., Ferrero, R., Montrucchio, B., and Rebaudengo, M. (2017). “A neural network model based on co-occurrence matrix for fall prediction” in Wireless mobile communication and healthcare. MobiHealth 2016. Lecture notes of the institute for computer sciences, social informatics and telecommunications engineering. eds. P. Perego, G. Andreoni, and G. Rizzo, vol. 192 (Cham: Springer), 241–248.
Iadarola, G., Disha, D., De Santis, A., Spinsante, S., and Gambi, E. (2022). Global positioning system measurements: comparison of IoT wearable devices. 2022 IEEE 9th international workshop on metrology for AeroSpace (MetroAeroSpace), 213–218. Pisa, Italy
Islam, Z. Z., Tazwar, S. M., Islam, M. Z., Serikawa, S., and Ahad, M. A. R.. (2017). Automatic fall detection system of unsupervised elderly people using smartphone. Proceedings of the 5th IIAE international conference on industrial application engineering 2017, 418–424.
Jayaraj, R. L., Azimullah, S., Beiram, R., Jalal, F. Y., and Rosenberg, G. A. (2019). Neuroinflammation: friend and foe for ischemic stroke. J. Neuroinflammation 16:142. doi: 10.1186/s12974-019-1516-2
Jegham, I., Ben Khalifa, A., Alouani, I., and Mahjoub, M. A. (2020). Vision-based human action recognition: An overview and real world challenges. Foren. Sci. Internat. 32:200901. doi: 10.1016/j.fsidi.2019.200901
Jeong, S. Y., Choi, M. S., Kang, S. J., and Lee, H.-W. (2017). Homecare kit platform supporting multiple bio-signals acquisition and analysis in daily-life. 2017 IEEE international conference on big data and smart computing (BigComp), Jeju, Korea 64–68.
Jindal, P., and Kumar, D. (2017). A review on dimensionality reduction techniques. Int. J. Comput. Applicat. 173, 42–46. doi: 10.5120/ijca2017915260
Jung, S., Hong, S., Kim, J., Lee, S., Hyeon, T., Lee, M., et al. (2015). Wearable fall detector using integrated sensors and energy devices. Sci. Rep. 5:17081. doi: 10.1038/srep17081
Junior, A. P., Díez, L. E., Bahillo, A., and Eyobu, O. S. (2023). Remote pedestrian localization systems for resource-constrained environments: a systematic review. IEEE Access 11, 36865–36889. doi: 10.1109/ACCESS.2023.3266957
Kang, J., and Larkin, H. (2017). Application of an emergency alarm system for physiological sensors utilizing smart devices. Technologies 5:26. doi: 10.3390/technologies5020026
Karayaneva, Y., Sharifzadeh, S., Jing, Y., and Tan, B. (2023). Human activity recognition for AI-enabled healthcare using low-resolution infrared sensor data. Sensors 23:478. doi: 10.3390/s23010478
Ke, S.-R., Thuc, H., Lee, Y.-J., Hwang, J.-N., Yoo, J.-H., and Choi, K.-H. (2013). A review on video-based human activity recognition. Computers 2, 88–131. doi: 10.3390/computers2020088
Kenfack Ngankam, H., Pigot, H., Lorrain, D., Viens, I., and Giroux, S. (2020). Context awareness architecture for ambient-assisted living applications: case study of nighttime wandering. J. Rehabilit. Assist. Technol. Eng. 7:205566831988786. doi: 10.1177/2055668319887864
Keum, S. S., Lee, C. H., and Kang, S. J. (2019). Device to device collaboration architecture for real- time identification of user and abnormal activities in home. 2019 29th international telecommunication networks and applications conference (ITNAC), 1–3
Khan, N. S., and Ghani, M. S. (2021). A survey of deep learning based models for human activity recognition. Wirel. Pers. Commun. 120, 1593–1635. doi: 10.1007/s11277-021-08525-w
Kumar, P., and Suresh, S. (2023). Deep-HAR: an ensemble deep learning model for recognizing the simple, complex, and heterogeneous human activities. Multimed. Tools Appl. doi: 10.1007/s11042-023-14492-0
Kumari, P., Mathew, L., and Syal, P. (2017). Increasing trend of wearables and multimodal interface for human activity monitoring: a review. Biosens. Bioelectron. 90, 298–307. doi: 10.1016/j.bios.2016.12.001
Kwapisz, J. R., Weiss, G. M., and Moore, S. A. (2011). Activity recognition using cell phone accelerometers. ACM SIGKDD Explorat. Newsl. 12, 74–82. doi: 10.1145/1964897.1964918
Kwolek, B., and Kepski, M. (2014). Human fall detection on embedded platform using depth maps and wireless accelerometer. Comput. Methods Prog. Biomed. 117, 489–501. doi: 10.1016/j.cmpb.2014.09.005
Lee, S. (2021). Falls associated with indoor and outdoor environmental hazards among community-dwelling older adults between men and women. BMC Geriatr. 21:547. doi: 10.1186/s12877-021-02499-x
Li, M., and Sun, Q. (2021). 3D skeletal human action recognition using a CNN fusion model. Math. Probl. Eng. 2021, 1–11. doi: 10.1155/2021/6650632
Li, W., Zhang, Z., and Liu, Z. (2010). Action recognition based on a bag of 3D points. 2010 IEEE computer society conference on computer vision and pattern recognition – workshops San Francisco, CA, USA, 9–14
Liu, M., Liu, H., and Chen, C. (2017). Enhanced skeleton visualization for view invariant human action recognition. Pattern Recogn. 68, 346–362. doi: 10.1016/j.patcog.2017.02.030
Liu, L., and Shao, L. (2013). Learning discriminative representations from RGB-D video data. Proceedings of the twenty-third international joint conference on artificial intelligence, Beijing, China, AAAI Press 1493–1500.
Liu, J., Wang, Y., Liu, Y., Xiang, S., and Pan, C. (2020). 3D PostureNet: a unified framework for skeleton-based posture recognition. Pattern Recogn. Lett. 140, 143–149. doi: 10.1016/j.patrec.2020.09.029
Lloret, J., Canovas, A., Sendra, S., and Parra, L. (2015). A smart communication architecture for ambient assisted living. IEEE Commun. Mag. 53, 26–33. doi: 10.1109/MCOM.2015.7010512
Lutze, R. (2018). Practicality of smartwatch apps for supporting elderly people – a comprehensive survey. 2018 IEEE international conference on engineering, technology and innovation (ICE/ITMC), Stuttgart, Germany 1–7
Luwe, Y. J., Lee, C. P., and Lim, K. M. (2022). Wearable sensor-based human activity recognition with hybrid deep learning model. Informatics 9:56. doi: 10.3390/informatics9030056
Ma, C., Wang, A., Chen, G., and Xu, C. (2018). Hand joints-based gesture recognition for noisy dataset using nested interval unscented Kalman filter with LSTM network. Vis. Comput. 34, 1053–1063. doi: 10.1007/s00371-018-1556-0
Madhu, B., Mukherjee, A., Islam, M. Z., Mamun-Al-Imran, G. M., Roy, R., and Ali, L. E. (2021). “Depth motion map based human action recognition using adaptive threshold technique” in 2021 5th international conference on electrical information and communication technology (EICT), Khulna, Bangladesh, 1–6.
Mainetti, L., Manco, L., Patrono, L., Secco, A., Sergi, I., and Vergallo, R. (2016). An ambient assisted living system for elderly assistance applications. 2016 IEEE 27th annual international symposium on personal, indoor, and mobile radio communications (PIMRC), Valencia, Spain 1–6
Malekzadeh, M., Clegg, R. G., Cavallaro, A., and Haddadi, H. (2018). Protecting sensory data against sensitive inferences. Proceedings of the 1st workshop on privacy by design in distributed systems, Porto, Portugal 1–6
Marenzi, E., Bertolotti, G. M., Cristiani, A., De Donno, F., Leporati, F., Danese, G., et al. (2013a). Identification and evaluation of parameters for the prevention of pressure ulcers in hospitalized patients. IEEE EuroCon 2013. Zagreb, Croatia
Marenzi, E., Bertolotti, G. M., Leporati, F., and Danese, G. (2013b). Capacitive sensors matrix for interface pressure measurement in clinical, ergonomic and automotive environments. Proceedings – 16th Euromicro Conference on Digital System Design, DSD 2013. Los Alamitos, CA, USA
Marenzi, E., Lombardi, R., Bertolotti, G. M., Cristiani, A., and Cabras, B. (2012). Design and development of a novel capacitive sensor matrix for measuring pressure distribution. 2012 IEEE sensors applications symposium, SAS 2012 – proceedings. Brescia, Italy
Mashiyama, S., Hong, J., and Ohtsuki, T. (2015). Activity recognition using low resolution infrared array sensor. 2015 IEEE international conference on communications (ICC), London, UK 495–500
Mathur, A., Zhang, T., Bhattacharya, S., Velickovic, P., Joffe, L., Lane, N. D., et al. (2018). Using deep data augmentation training to address software and hardware heterogeneities in wearable and smartphone sensing devices. 2018 17th ACM/IEEE international conference on information processing in sensor networks (IPSN), Porto, Portugal 200–211.
Medrano, C., Igual, R., García-Magariño, I., Plaza, I., and Azuara, G. (2017). Combining novelty detectors to improve accelerometer-based fall detection. Med. Biol. Eng. Comput. 55, 1849–1858. doi: 10.1007/s11517-017-1632-z
Medrano, C., Igual, R., Plaza, I., and Castro, M. (2014). Detecting falls as novelties in acceleration patterns acquired with smartphones. PLoS One 9:e94811. doi: 10.1371/journal.pone.0094811
Minh Dang, L., Min, K., Wang, H., Piran, M., and Lee, C. (2020). Sensor-based and vision-based human activity recognition: a comprehensive survey. Pattern Recogn. 108:107561. doi: 10.1016/j.patcog.2020.107561
Mistry, J., and Inden, B. (2018). An approach to sign language translation using the Intel RealSense camera. 2018 10th computer science and electronic engineering (CEEC), Colchester, UK 219–224
Mliki, H., Bouhlel, F., and Hammami, M. (2020). Human activity recognition from UAV-captured video sequences. Pattern Recogn. 100:107140. doi: 10.1016/j.patcog.2019.107140
Müller, M., Röder, T., and Clausen, M. (2005). Efficient content-based retrieval of motion capture data. ACM Trans. Graph. 24, 677–685. doi: 10.1145/1073204.1073247
Narkhede, A. H. (2019). Human activity recognition based on multimodal body sensing San Jose, CA [San Jose State University
Nastac, D. I., Arsene, O., Dragoi, M., Stanciu, I. D., and Mocanu, I. (2019). An AAL scenario involving automatic data collection and robotic manipulation. 3rd IET international conference on technologies for active and assisted living (TechAAL 2019), London, UK 1–6.
Negin, F., Ozdemir, F., Yuksel, K. A., Akgul, C. B., and Ercil, A. (2013). A decision forest based feature selection framework for action recognition from RGB-depth cameras. 2013 21st signal processing and communications applications conference (SIU), Haspolat, Turkey 1–4.
Nouredanesh, M., Gordt, K., Schwenk, M., and Tung, J. (2020). Automated detection of multidirectional compensatory balance reactions: a step towards tracking naturally occurring near falls. IEEE Trans. Neural Syst. Rehabil. Eng. 28, 478–487. doi: 10.1109/TNSRE.2019.2956487
Ofli, F., Chaudhry, R., Kurillo, G., Vidal, R., and Bajcsy, R. (2013). Berkeley MHAD: a comprehensive multimodal human action database. 2013 IEEE workshop on applications of computer vision (WACV), Clearwater Beach, FL, USA 53–60
Oniga, S., and Suto, J. (2014). Human activity recognition using neural networks. Proceedings of the 2014 15th international carpathian control conference (ICCC), Velke Karlovice, Czech Republic 403–406.
Özdemir, A., and Barshan, B. (2014). Detecting falls with wearable sensors using machine learning techniques. Sensors 14, 10691–10708. doi: 10.3390/s140610691
Paolini, G., Masotti, D., Antoniazzi, F., Salmon Cinotti, T., and Costanzo, A. (2019). Fall detection and 3-D indoor localization by a custom RFID reader embedded in a smart e-health platform. IEEE Transact. Microw. Theory Techniq. 67, 5329–5339. doi: 10.1109/TMTT.2019.2939807
Patro, S. G. K., and Sahu, K. K. (2015). Normalization: a preprocessing stage. IARJSET 2, 20–22. doi: 10.17148/IARJSET.2015.2305
Peng, Y., Wu, Z., and Jiang, J. (2010). A novel feature selection approach for biomedical data classification. J. Biomed. Inform. 43, 15–23. doi: 10.1016/j.jbi.2009.07.008
Pierleoni, P., Belli, A., Palma, L., Paoletti, M., Raggiunto, S., and Pinti, F. (2019). Postural stability evaluation using wearable wireless sensor. 2019 IEEE 23rd international symposium on consumer technologies (ISCT), Ancona, Italy 256–260
Pires, I. M., Hussain, F., Garcia, N. M., Lameski, P., and Zdravevski, E. (2020). Homogeneous data normalization and deep learning: a case study in human activity classification. Fut. Internet 12:194. doi: 10.3390/fi12110194
Piwek, L., Ellis, D. A., Andrews, S., and Joinson, A. (2016). The rise of consumer health wearables: promises and barriers. PLoS Med. 13:e1001953. doi: 10.1371/journal.pmed.1001953
Plentz, P. D. M., and De Pieri, E. R. (2018). An overview on real-time constraints for ambient intelligence (AmI). 2018 IEEE/ACS 15th international conference on computer systems and applications (AICCSA), Aqaba, Jordan 1–7.
Poulose, A., Kim, J. H., and Han, D. S. (2022). HIT HAR: human image threshing machine for human activity recognition using deep learning models. Comput. Intell. Neurosci. 2022, 1–21. doi: 10.1155/2022/1808990
Qiu, S., Zhao, H., Jiang, N., Wang, Z., Liu, L., An, Y., et al. (2022). Multi-sensor information fusion based on machine learning for real applications in human activity recognition: state-of-the-art and research challenges. Informat. Fusion 80, 241–265. doi: 10.1016/j.inffus.2021.11.006
Raeis, H., Kazemi, M., and Shirmohammadi, S. (2021). Human activity recognition with device-free sensors for well-being assessment in smart homes. IEEE Instrument. Measur. Magaz. 24, 46–57. doi: 10.1109/MIM.2021.9513637
Rafferty, J., Nugent, C. D., Liu, J., and Chen, L. (2017). From activity Recognition to intention Recognition for assisted living within smart homes. IEEE Trans. Hum. Mach. Syst. 47, 368–379. doi: 10.1109/THMS.2016.2641388
Ramanujam, E., Perumal, T., and Padmavathi, S. (2021). Human activity recognition with smartphone and wearable sensors using deep learning techniques: a review. IEEE Sensors J. 21, 13029–13040. doi: 10.1109/JSEN.2021.3069927
Ranieri, C. M., MacLeod, S., Dragone, M., Vargas, P. A., and Romero, R. A. F. (2021). Activity recognition for ambient assisted living with videos, inertial units and ambient sensors. Sensors 21:768. doi: 10.3390/s21030768
Reiss, A., and Stricker, D. (2012). Introducing a new benchmarked dataset for activity monitoring. 2012 16th international symposium on wearable computers, 108–109.
Rodriguez, M. D., Ahmed, J., and Shah, M. (2008). Action MACH a spatio-temporal maximum average correlation height filter for action recognition. 2008 IEEE conference on computer vision and pattern recognition, Newcastle, UK 1–8
Roggen, D., Calatroni, A., Rossi, M., Holleczek, T., Forster, K., Troster, G., et al. (2010). Collecting complex activity datasets in highly rich networked sensor environments. 2010 seventh international conference on networked sensing systems (INSS), Kassel, Germany 233–240
Roland, L., Lidauer, L., Sattlecker, G., Kickinger, F., Auer, W., Sturm, V., et al. (2018). Monitoring drinking behavior in bucket-fed dairy calves using an ear-attached tri-axial accelerometer: a pilot study. Comput. Electron. Agric. 145, 298–301. doi: 10.1016/j.compag.2018.01.008
Ronald, M., Poulose, A., and Han, D. S. (2021). iSPLInception: An inception-ResNet Deep learning architecture for human activity Recognition. IEEE. Access 9, 68985–69001. doi: 10.1109/ACCESS.2021.3078184
Rupasinghe, I. D. M. S., and Maduranga, M. W. P. (2022). Towards ambient assisted living (AAL): design of an IoT-based elderly activity monitoring system. Int. J. Eng. Manufact. 12, 1–10. doi: 10.5815/ijem.2022.02.01
Ryselis, K., Petkus, T., Blažauskas, T., Maskeliūnas, R., and Damaševičius, R. (2020). Multiple Kinect based system to monitor and analyze key performance indicators of physical training. HCIS 10:51. doi: 10.1186/s13673-020-00256-4
Saleh, M., Abbas, M., and Le Jeannes, R. B. (2021). FallAllD: an open dataset of human falls and activities of daily living for classical and Deep learning applications. IEEE Sensors J. 21, 1849–1858. doi: 10.1109/JSEN.2020.3018335
Saliba, M., Drapeau, N., Skime, M., Hu, X., Accardi, C. J., Athreya, A. P., et al. (2023). PISTACHIo (PreemptIon of diSrupTive behAvior in CHIldren): real-time monitoring of sleep and behavior of children 3–7 years old receiving parent–child interaction therapy augment with artificial intelligence — the study protocol, pilot study. Pilot Feasibility Stud. 9:23. doi: 10.1186/s40814-023-01254-w
Sanchez-Comas, A., Synnes, K., and Hallberg, J. (2020). Hardware for recognition of human activities: a review of smart home and AAL related technologies. Sensors 20:4227. doi: 10.3390/s20154227
Sarabia-Jácome, D., Usach, R., Palau, C. E., and Esteve, M. (2020). Highly-efficient fog-based deep learning AAL fall detection system. Internet Things 11:100185. doi: 10.1016/j.iot.2020.100185
Sathya, R., and Abraham, A. (2013). Comparison of supervised and unsupervised learning algorithms for pattern classification. Int. J. Adv. Res. Artif. Intellig. 2, 34–38. doi: 10.14569/IJARAI.2013.020206
Schuldt, C., Laptev, I., and Caputo, B. (2004). Recognizing human actions: a local SVM approach. Proceedings of the 17th international conference on pattern recognition, 2004. ICPR 2004. Cambridge, UK, 3, 32–36.
Seidenari, L., Varano, V., Berretti, S., Del Bimbo, A., and Pala, P. (2013). Recognizing actions from depth cameras as weakly aligned multi-part bag-of-poses. Proceedings of the IEEE conference on computer vision and pattern recognition (CVPR) workshops Portland, OR, USA. doi: 10.1109/CVPRW.2013.77
Seneviratne, S., Hu, Y., Nguyen, T., Lan, G., Khalifa, S., Thilakarathna, K., et al. (2017). A survey of wearable devices and challenges. IEEE Commun. Surv. Tutor. 19, 2573–2620. doi: 10.1109/COMST.2017.2731979
Senigagliesi, L., Ciattaglia, G., Disha, D., and Gambi, E. (2022). Classification of human activities based on automotive radar spectral images using machine learning techniques: a case study. 2022 IEEE radar conference (RadarConf22), New York City, NY, USA 1–6.
Shahroudy, A., Liu, J., Ng, T.-T., and Wang, G. (2016). NTU RGB+D: a large scale dataset for 3D human activity analysis. 2016 IEEE conference on computer vision and pattern recognition (CVPR), Las Vegas, NV, USA 1010–1019.
Sheikh, S. Y., and Jilani, M. T. (2023). A ubiquitous wheelchair fall detection system using low-cost embedded inertial sensors and unsupervised one-class SVM. J. Ambient. Intell. Humaniz. Comput. 14, 147–162. doi: 10.1007/s12652-021-03279-6
Shewell, C., Nugent, C., Donnelly, M., Wang, H., and Espinilla, M. (2017). Indoor localisation through object detection within multiple environments utilising a single wearable camera. Heal. Technol. 7, 51–60. doi: 10.1007/s12553-016-0159-x
Shi, T., Sun, X., Xia, Z., Chen, L., and Liu, J. (2016). Fall detection algorithm based on Triaxial accelerometer and magnetometer. Eng. Lett. 24, 157–163.
Shumba, A.-T., Montanaro, T., Sergi, I., Fachechi, L., De Vittorio, M., and Patrono, L. (2022). Leveraging IoT-aware technologies and AI techniques for real-time critical healthcare applications. Sensors 22:7675. doi: 10.3390/s22197675
Siddiqi, M. H., and Alsirhani, A. (2022). An efficient feature selection method for video-based activity recognition systems. Math. Probl. Eng. 2022, 1–13. doi: 10.1155/2022/5486004
Sikder, N., and Nahid, A.-A. (2021). KU-HAR: an open dataset for heterogeneous human activity recognition. Pattern Recogn. Lett. 146, 46–54. doi: 10.1016/j.patrec.2021.02.024
Sindhu Meena, K., and Suriya, S. (2020). “A survey on supervised and unsupervised learning techniques” in Proceedings of international conference on artificial intelligence, smart grid and smart city applications (Cham: Springer International Publishing), 627–644.
Singh, A., Rehman, S. U., Yongchareon, S., and Chong, P. H. J. (2020). Sensor technologies for fall detection systems: a review. IEEE Sensors J. 20, 6889–6919. doi: 10.1109/JSEN.2020.2976554
Siriwardhana, C., Madhuranga, D., Madushan, R., and Gunasekera, K. (2019). “Classification of activities of daily living based on depth sequences and audio” in 2019 14th conference on industrial and information systems (ICIIS) (Kandy, Sri Lanka), 278–283.
Sivakumar, S., Gopalai, A. A., Gouwanda, D., and Lim, K. H. (2018). Estimation of joint angle from ground reaction force in human gait. 2018 IEEE-EMBS conference on biomedical engineering and sciences (IECBES), Sarawak, Malaysia, 623–628.
Sokullu, R., Balcı, A., and Demir, E. (2019). “The role of drones in ambient assisted living systems for the elderly” in Enhanced living environments. Lecture notes in computer science. eds. I. Ganchev, N. Garcia, C. Dobre, C. Mavromoustakis, and R. Goleva, vol. 11369 (Cham: Springer), 295–321.
Sonawani, S., and Patil, K. (2023). Air quality measurement, prediction and warning using transfer learning based IOT system for ambient assisted living. Int. J. Pervas. Comput. Communicat. doi: 10.1108/IJPCC-07-2022-0271
Sophia, S., Sridevi, U. K., Boselin, P. S. R., and Thamaraiselvi, P. (2021). “Ambient-assisted living of disabled elderly in an intelligent home using behavior prediction—a reliable Deep learning prediction system” in Computational analysis and deep learning for medical care (Hoboken: Wiley), 329–342.
Sousa Lima, W., Souto, E., El-Khatib, K., Jalali, R., and Gama, J. (2019). Human activity recognition using inertial sensors in a smartphone: an overview. Sensors 19:3213. doi: 10.3390/s19143213
Spasova, V., Iliev, I., and Petrova, G. (2016). Privacy preserving fall detection based on simple human silhouette extraction and a linear support vector machine. Int. J. Bioautomat. 20, 237–252.
Srimath, S., Ye, Y., Sarker, K., Sunderraman, R., and Ji, S. (2021). Human activity Recognition from RGB video streams using 1D-CNNs. 2021 IEEE SmartWorld, ubiquitous intelligence & computing, advanced & trusted computing, scalable computing & communications, internet of people and smart city innovation (SmartWorld/SCALCOM/UIC/ATC/IOP/SCI), 295–302.
Stavropoulos, T. G., Papastergiou, A., Mpaltadoros, L., Nikolopoulos, S., and Kompatsiaris, I. (2020). IoT wearable sensors and devices in elderly care: a literature review. Sensors 20:2826. doi: 10.3390/s20102826
Steven Eyobu, O., and Han, D. (2018). Feature representation and data augmentation for human activity classification based on wearable IMU sensor data using a deep LSTM neural network. Sensors 18:2892. doi: 10.3390/s18092892
Stodczyk, R. (2020). Ambient assisted living an overview of current applications, end-users and acceptance. Biomed. J. Sci. Techn. Res. 30, 23374–23384. doi: 10.26717/BJSTR.2020.30.004949
Su, B., Wu, H., and Sheng, M. (2017). Human action recognition method based on hierarchical framework via Kinect skeleton data. 2017 international conference on machine learning and cybernetics (ICMLC), Ningbo, China 83–90.
Sucerquia, A., López, J., and Vargas-Bonilla, J. (2017). SisFall: a fall and movement dataset. Sensors 17:198. doi: 10.3390/s17010198
Sun, X., Wei, Y., Liang, Shuang, Tang, X., and Sun, J. (2015). Cascaded hand pose regression. 2015 IEEE conference on computer vision and pattern recognition (CVPR), Boston, MA, USA 824–832.
Sung, Jaeyong, Ponce, C., Selman, B., and Saxena, A. (2012). Unstructured human activity detection from RGBD images. 2012 IEEE international conference on robotics and automation, 842–849. Saint Paul, MN
Surendran, D., Janet, J., Prabha, D., and Anisha, E. (2018). A study on devices for assisting Alzheimer patients. 2018 2nd international conference on I-SMAC (IoT in social, mobile, analytics and cloud) (I-SMAC), Palladam, India 620–625.
Suto, J., Oniga, S., and Sitar, Petrica Pop (2016). Comparison of wrapper and filter feature selection algorithms on human activity recognition. Available at: https://ieeexplore.ieee.org/abstract/document/7496749/ (Accessed September 16, 2022).
Tay, N. C., Connie, T., Ong, T. S., Teoh, A. B. J., and Teh, P. S. (2023). A review of abnormal behavior detection in activities of daily living. IEEE Access 11, 5069–5088. doi: 10.1109/ACCESS.2023.3234974
Tazari, M. R., Wichert, R., and Norgall, T. (2011). “Towards a unified ambient assisted living and personal health environment” in Ambient assisted living (Berlin Heidelberg: Springer), 141–155.
Torti, E., Fontanella, A., Musci, M., Blago, N., Pau, D., Leporati, F., et al. (2019). Embedding recurrent neural networks in wearable systems for real-time fall detection. Microprocess. Microsyst. 71:102895. doi: 10.1016/j.micpro.2019.102895
Tu, J., Liu, M., and Liu, H. (2018). Skeleton-based human action recognition using spatial temporal 3D convolutional neural networks. 2018 IEEE international conference on multimedia and expo (ICME), San Diego, CA, USA 1–6.
Um, T. T., Pfister, F. M. J., Pichler, D., Endo, S., Lang, M., Hirche, S., et al. (2017). “Data augmentation of wearable sensor data for parkinson’s disease monitoring using convolutional neural networks” in Proceedings of the 19th ACM international conference on multimodal interaction. eds. E. Lank, A. Vinciarelli, E. Hoggan, S. Subramanian, and S. A. Brewster (New York: Association for Computing Machinery), 216–220.
Vavoulas, G., Chatzaki, C., Malliotakis, T., Pediaditis, M., and Tsiknakis, M. (2016). The MobiAct dataset: recognition of activities of daily living using smartphones. Proceedings of the international conference on information and communication technologies for ageing well and e-health, Rome, Italy 143–151.
Wang, Y., Cang, S., and Yu, H. (2019). A survey on wearable sensor modality centred human activity recognition in health care. Expert Syst. Appl. 137, 167–190. doi: 10.1016/j.eswa.2019.04.057
Wang, J., Chen, Y., Hao, S., Peng, X., and Hu, L. (2019). Deep learning for sensor-based activity recognition: a survey. Pattern Recogn. Lett. 119, 3–11. doi: 10.1016/j.patrec.2018.02.010
Wang, X., Ellul, J., and Azzopardi, G. (2020). Elderly fall detection systems: a literature survey. Front. Robot. AI 7:71. doi: 10.3389/frobt.2020.00071
Wang, L., Huynh, D. Q., and Koniusz, P. (2020). A comparative review of recent Kinect-based action Recognition algorithms. IEEE Transact. Image Proc. 29, 15–28. doi: 10.1109/TIP.2019.2925285
Wang, S., Zhang, X., Liu, X., Zhang, J., Ma, S., and Gao, W. (2017). Utility-driven adaptive preprocessing for screen content video compression. IEEE Trans. Multimed. 19, 660–667. doi: 10.1109/TMM.2016.2625276
Warunsin, K., and Phairoh, T. (2022). Wristband fall detection system using deep learning. 2022 7th international conference on computer and communication systems (ICCCS), Wuhan, China 223–227.
Wu, C., Zhang, J., Savarese, S., and Saxena, A. (2015). Watch-n-patch: unsupervised understanding of actions and relations. 2015 IEEE conference on computer vision and pattern recognition (CVPR), Boston, MA, USA 4362–4370.
Xia, L., Chen, C.-C., and Aggarwal, J. K. (2012). View invariant human action recognition using histograms of 3D joints. 2012 IEEE computer society conference on computer vision and pattern recognition workshops, 20–27. Providence, RI, USA
Xie, J., Guo, K., Zhou, Z., Yan, Y., and Yang, P. (2020). ART: adaptive and real-time fall detection using COTS smart watch. 2020 6th international conference on big data computing and communications (BIGCOM), Deqing, China 33–40.
Yue, S., Yang, Y., Wang, H., Rahul, H., and Katabi, D. (2020). BodyCompass. Proc. ACM Interact. Mobile Wearable Ubiquitous Technol. 4, 1–25. doi: 10.1145/3397311
Zdravevski, E., Lameski, P., Trajkovik, V., Kulakov, A., Chorbev, I., Goleva, R., et al. (2017). Improving activity recognition accuracy in ambient-assisted living systems by automated feature engineering. IEEE Access 5, 5262–5280. doi: 10.1109/ACCESS.2017.2684913
Zebari, R. R., Mohsin Abdulazeez, A., Zeebaree, D. Q., Zebari, D. A., and Saeed, J. N. (2020). A comprehensive review of dimensionality reduction techniques for feature selection and feature extraction. J. Appl. Sci. Technol. Trends 1, 56–70. doi: 10.38094/jastt1224
Zeng, L., Duan, X., Pan, Y., and Deng, M. (2023). Research on the algorithm of helmet-wearing detection based on the optimized yolov4. Vis. Comput. 39, 2165–2175. doi: 10.1007/s00371-022-02471-9
Zerrouki, N., Harrou, F., Sun, Y., and Houacine, A. (2018). Vision-based human action classification using adaptive boosting algorithm. IEEE Sensors J. 18, 5115–5121. doi: 10.1109/JSEN.2018.2830743
Zhang, Y., Chen, Y., Wang, J., and Pan, Z. (2023). Unsupervised deep anomaly detection for multi-sensor time-series signals. IEEE Trans. Knowl. Data Eng. 35, 2118–2132. doi: 10.1109/TKDE.2021.3102110
Keywords: human activity recognition, ambient assisted living, wearable systems, frail people, deep learning
Citation: Guerra BMV, Torti E, Marenzi E, Schmid M, Ramat S, Leporati F and Danese G (2023) Ambient assisted living for frail people through human activity recognition: state-of-the-art, challenges and future directions. Front. Neurosci. 17:1256682. doi: 10.3389/fnins.2023.1256682
Edited by:
Lei Zhang, Nanjing Normal University, ChinaReviewed by:
Alwin Poulose, Indian Institute of Science Education and Research, Thiruvananthapuram, IndiaEnnio Gambi, Marche Polytechnic University, Italy
Irina Mocanu, Polytechnic University of Bucharest, Romania
Copyright © 2023 Guerra, Torti, Marenzi, Schmid, Ramat, Leporati and Danese. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Emanuele Torti, emanuele.torti@unipv.it