- Ubiquitous and Personal Computing Lab, Faculty of Engineering, Kyoto University of Advanced Science (KUAS), Kyoto, Japan
Sleep apnea is one of the most common sleep disorders, which, if left untreated, may have severe health consequences in the long term. Many sleep apnea patients remain non-diagnosed due to lacking access to medical tests. In recent years, portable and wearable sensors that measure blood oxygen saturation (SpO) are becoming common and affordable for daily use, and they open the door for affordable and accessible sleep apnea screening in the context of everyday life. To learn about the advancement in SpO-based sleep apnea screening, we conducted a survey of published studies. We searched databases including Springer, Science Direct, Web of Science, ACM Digital Library, and IEEE Xplore using the keywords “sleep apnea” AND (“SpO2” OR “blood oxygen saturation”) AND (“machine learning” OR “deep learning”). After screening 835 results, we included 31 publications for a full-text review. Analysis shows that SpO-based sleep apnea screening studies consist of three main categories: (1) individual apnea events detection, (2) apnea-hypopnea index prediction, and (3) apnea severity classification. We found two significant research gaps: a lack of sufficient and diverse publicly available datasets, and the absence of standardized protocols for data collection, signal preprocessing, and model bench marking. Future research should focus on addressing these gaps to enhance the effectiveness and reliability of AI-driven sleep apnea screening methods using SpO signals.
1 Introduction
According to a comprehensive review by Benjafield et al. (1), nearly 1 billion people are affected by sleep apnea, with prevalence exceeding 50% in some countries. Undiagnosed sleep apnea has been shown to be associated with high comorbidities and mortality, and reduced quality of life (2). Many people with sleep apnea do not experience noticeable symptoms, leading to a lack of motivation for diagnostic testing (3, 4). This review assesses recent advances in AI algorithms for the screening of sleep apnea, emphasizing the use of SpO due to its non-invasive nature and effectiveness, and aims to highlight future research directions.
The rise of home-based sleep apnea tests (HSAT) has highlighted the potential of wearable devices in supporting sleep health in everyday life. Smartwatches such as the Apple Watch, Samsung Watch, Google Pixel Watch, and Fitbit have become popular due to their convenience and functionality (5, 6). These devices generally include a green light reflective photoplethysmography (PPG) sensor for measuring blood oxygen saturation (SpO). Given that sleep apnea impacts both airway and SpO levels (7–9), many studies have been conducted to explore SpO as a light alternative to PSG for home-based sleep apnea detection, especially when combined with AI-driven computational methods (10–13). However, there remains a gap in understanding current modelling practices and performance. This mini-review aimed to explore such gaps and potential solutions in sleep apnea detection. Section 2 outlines the review methodology. Section 3 examines employed databases, SpO processing techniques, feature extraction methods, sleep apnea screening approaches, model development and performance across each screening task. The final section will evaluate remaining limitations and propose future research directions.
2 Materials and methods
We followed a search and selection process that is consistent with the methodology for mini-reviews, as outlined (14). Keywords “sleep apnea” AND “SpO2” AND “machine learning” were used to search publications in 4 databases: Science Direct (), ACM Digital Library (), IEEE Xplore (), and Springer Link (), yielding a total of 824 entries. All entries retrieved were imported into Rayyan (15) to streamline the review process and eliminate duplicates. Through Rayyan’s duplication removal feature, 31 duplicate articles were identified and removed. The remaining articles were then screened based on predefined inclusion and exclusion criteria. Inclusion criteria required that articles focus on developing an application, model, or algorithm specifically for the screening of sleep apnea. In addition, studies needed to employ blood oxygen saturation (SpO) signals as a primary input for the screening approach and employ machine learning algorithms as part of the methodology. Exclusion criteria were applied to further refine the selection. Articles were excluded if SpO was not the main signal used in the algorithm. Studies on pediatric populations were excluded due to the distinct nature of sleep apnea in children compared to adults. Articles that were not publicly accessible or not written in English were also excluded.
Following the title and abstract screening, 52 articles remained. 21 articles were excluded from the analysis: 2 of these excluded articles were identified as review papers, 3 additional articles employed demographic data as the primary input for regression or classification tasks, 9 excluded articles focused solely on apnea event detection using definitions established by the American Academy of Sleep Medicine (AASM), not incorporating machine learning algorithms as required. Finally, 7 articles were excluded due to a lack of relevance to the overall topic. A final review based on the main content resulted in a final selection of 31 articles deemed relevant for this review, the main findings are provided below.
3 Results
3.1 Sleep datasets
More than half of the studies () used proprietary datasets collected within research facilities, with dataset size ranging from fewer than 50 sleep records (16, 17) to several hundreds (18–21). While proprietary datasets allow tailored data collection and expert labeling, their limited public availability poses challenges for comparison and benchmarking.
In contrast, public datasets provide valuable alternatives. Widely-used ones include the Apnea-ECG Database (AED) (22), the St.Vincent’s University Hospital/University College Dublin Sleep Apnea Database (UCD) (23), and more recently, the OSASUD dataset (24). However, the relatively small sizes of these datasets (often fewer than 100 records) limit their applicability primarily to epoch-wise, rather than subject-wise screening (8, 9). The Sleep Heart Health Study (SHHS) (25), with over 5,000 recordings and high-resolution labels, is ideal for deep learning models. Other publicly available datasets, such as the Wisconsin Sleep Cohort dataset () (26), Cleveland Family Study (CFS) dataset () (27), Osteoporotic Fractures in Men Study (MROS) dataset () (28) and the Multi-Ethnic Study of Atherosclerosis (MESA) dataset () (29), offer valuable data but have been less utilized in sleep apnea research so far.
3.2 Pre-processing SpO signals
Physiological signals such as SpO are prone to movement contamination and thus require pre-processing to remove noises. However, our analysis revealed that many studies () proceed with raw SpO without pre-processing, and there is a lack of standardized protocols for filtering noise or assessing signal quality.
SpO are typically calibrated within a range of 70% to 100% saturation, with an accuracy of 2% to 4% (30). Consequently, readings below 70% may be inaccurate, prompting some studies to use thresholds of 70% or 65% to remove unreliable readings (3, 31). A lower threshold of 50% has also been used in several studies to account for physiological limitations and equipment errors (9, 12, 17, 20, 32). In addition to cut-off thresholds, some studies applied further noise reduction techniques, such as removing data points where consecutive SpO values differ by more than a predefined value (e.g., 4%) (3, 12, 20, 33).
Few studies () explored optimal SpO signal bands. One study identified the apnea-related band as 0.014–0.033 Hz (33), while (21) argued that the shape of SpO signal is similar to a sinusoid with 0.02 Hz frequency and therefore used a 0.02 Hz IIR Butter-worth low-pass filter to suppress and smooth the SpO signal. Another study employed a complex Wavelet filter to eliminate noise from muscle movements (4). Furthermore, Stuban and Niwayama (34) demonstrated that lowering the low-pass filter frequency to a value closer to the fundamental frequency of the PPG signal reduced noise without compromising measurement accuracy. An additional 10 dB of signal-to-noise ratio (SNR) is recommended for accurate SpO measurement (35). However, no study provided a detailed justification for these methods, nor did they analyzed how these techniques influenced signal quality and subsequent classification performance.
The most common pre-processing practice involves using a threshold between 50% and 100% to eliminate hardware errors. However, to determine the optimal filter settings, future studies are needed to conduct a comprehensive benchmarking of all possible filter options.
3.3 Feature construction and selection
Two primary approaches dominate feature extraction methods. The first approach relies on manual feature extraction, leveraging the researchers’ domain expertise. This approach has been applied in 19 studies, where hand-crafted features were derived to ensure model interpretability and applicability across various shallow learning models. While this method can be time-consuming, it remains valuable for its transparency and ease of understanding. A comprehensive list of SpO features is detailed in Xie and Minn (9), Gutiérrez-Tobal et al. (36), Levy et al. (37).
To optimize features for model construction, several techniques have been employed to select features with strong discriminating power. These include forward stepwise logistic regression (18), recursive feature elimination (38), fast correlation-based filter (36), maximum relevance minimum redundancy (11), and heat-map (39). One commonly used feature is the Oxygen Desaturation Index (ODI), due to its strong correlation with the Apnea-Hypopnea Index (AHI) which is a standard measurement of sleep apnea severity (7, 11, 20, 33, 39–42). However, ODI requires a minimum sampling rate of 1 Hz, which limits its applicability in datasets from lower-frequency devices like smartwatches. Other features such as entropies (7, 12, 40), Lempel-Ziv complexity (LZ) (7, 33, 43), and demographic features (e.g., age, gender, neck circumference, body mass index (BMI)) (41, 44, 45) have also gained prominence. These features have been ranked highly in importance compared to others.
Deep learning (DL) has recently emerged as a novel feature extraction paradigm. DL models, such as convolutional neural networks (CNNs) (46, 47) and long short-term memory (LSTM) networks (8), are capable of automatically extracting features, potentially uncovering patterns that manual feature extraction may overlook. For instance, a study by Lyden et al. (48) demonstrated that CNN and LSTM models achieved high performance in epoch-wise classification, with accuracy, sensitivity, and precision exceeding 90%, even when working with reduced signal sampling rates. This highlights the effectiveness of DL in apnea screening. However, the trade-off is that the interpretability and explainability of these extracted features remain an ongoing challenge as visualizing how the variables are interconnected and weighted within the network is virtually impossible (49).
Manually crafting features is a viable approach to enhancing the explainability and interpretability of machine learning methods, particularly when aiming for clinical acceptance. Unlike features derived from deep learning models, which often suffer from the “black box” problem that limits transparency, handcrafted features are well-defined, easy to visualize, and straightforward to interpret. Feature-ranking techniques, such as SHAP or Grad-CAM, can further facilitate a deeper understanding of how machine learning algorithms work. Ensuring that methods are explainable and transparent not only improves their transition to real-world applications but also enables targeted interventions when the model produces incorrect predictions.
3.4 Apnea screening model development
All studies in this review employ supervised learning for sleep apnea screening. As shown in Figure 1, the studies were categorized into three main problem formulations, epoch-wise screening, AHI regression, and subject-wise screening.

Figure 1. Summary of SpO-based sleep apnea screening in terms of target outputs, primary achievements, research gaps and challenges, and future applications.
3.4.1 Epoch-wise model development
Epoch-wise classification, used in 18 studies, involves segmenting a night’s SpO recording into 1-min epochs, with models classifying each epoch as positive or negative for an apnea event. A few studies framed the problem as multiclass classification to distinguish among normal, hypopnea, obstructive and central apnea events (31, 50). While it is true that sleep apnea screening can be done by calculating AHI by effective respiratory event detection, only 5 studies further post-process the epoch-wise classification results to estimate AHI or assess apnea severity (16, 19, 20, 51, 52).
3.4.2 Subject-wise model development
Subject-wise classification aims to assign a whole night’s SpO recording either a binary (positive and negative) or a multiclass label (normal, mild, moderate, or severe). This approach classifies the entire night’s recording, eliminating the need to process large number of 1-min epochs, thereby reducing computational cost. While many studies focused on binary classification (7, 12, 36, 38, 43, 53), only a few have applied multi-class classification (40, 54). Another approach is AHI regression, in which a machine learning model aims to predict AHI as a continuous variable, and later categorize sleep apnea severity into predefined classes (e.g., normal, mild, moderate, severe) (11, 33, 38, 52). This approach allows for more granular predictions, which can be particularly useful in identifying borderline cases or tracking changes in AHI over time.
3.4.3 Class balancing
A common issue in sleep apnea classification is the imbalanced distribution among classes. Epoch-wise classifications often have a skewed distribution towards normal epochs [e.g., 90% normal in (4)]. Even with high AHI, normal epochs can dominate [e.g., 69% normal epochs in (16)]. Subject-wise classifications are more balanced but may still skew towards severe cases (e.g., 56% severe in (36)). Several studies address class imbalance with techniques such as random sampling, SMOTE, and ADASYN (10, 21, 33, 38, 44, 45). These methods create synthetic data to balance classes but may alter natural data distribution and impact model generalizability (55).
3.5 Model performance in sleep apnea screening
The studies reviewed reveal a combination of classification techniques for the screening of sleep apnea. Table 1 shows various shallow and deep learning models have been applied. However, it is difficult to conclude which method is the most optimal due to the lack of standardized evaluation metrics. For classification problems, evaluation metrics include accuracy (ACC), sensitivity (SE), specificity (SPEC), F1 score, Kappa coefficient, Matthews correlation coefficient (MCC), and area under the curve (AUC). On the other hand, regression problems require a distinct set of evaluation metrics, such as correlation coefficient, intraclass correlation coefficient (ICC), Bland-Altman plots, root mean squared error (RMSE), mean absolute error (MAE), and -squared. There has yet to be a consensus as to which metrics are the best for evaluating model performance.

Table 1. A summary of the dataset used, the signal sampling rate, preprocessing methods, applied models, and result reported in included studies.
3.5.1 Epoch-wise model performance
Epoch-wise classification has shown high performance in sleep apnea events detection. For instance, Lyden et al. (48) reported impressive results using shallow models combined with DL-based features, achieving accuracy of 97.04%, specificity of 97.19%, and sensitivity of 96.94%. This study simulated smartwatch data by adding Gaussian noise to down-sampled SpO signals at various signal-to-noise ratios. Most models, except SVM, were minimally affected by reduced sampling frequency, although performance dropped significantly below a 30dB signal-to-noise ratio. Naive Bayes models with LSTM-based features provided the most stable performance.
A major challenge in epoch-wise classification is the delay between sleep apnea events and SpO desaturation, with delays reported from 10 to 40 s (21, 51, 56). In addition, subjects sleep only 75.9% of the time, and hypopnea events are more common than apnea events or periods of normal breathing (e.g., 73% in the SHHS dataset). Notably, 11.5% of apneic events lack SpO desaturation (51), complicating model accuracy. To address these issues, Bernardini et al. (16) focused on detecting clusters of anomalies, providing valuable clinical insights despite lower performance metrics (81.5% ACC, 67.2% SE). The study accurately classified sleep apnea severity in 21 out of 30 cases. Punjabi (57) highlighted that the distribution of apneic events over the night is crucial for understanding their health impact.
In an effort to reduce classification errors, Bark et al. (10) developed a selective model that improves accuracy by rejecting low-confidence predictions, achieving 90.26% ACC, 91.29% SE, and 89.21% SPEC. Despite these promising results, model generalization remains a concern due to small sample sizes. Analysis suggests that RNN and LSTM models generally outperform CNNs in this field (10, 16, 48, 55).
3.5.2 Subject-wise model performance
The OxiNet model by Levy et al. (11) demonstrated high performance and generalization, analyzing 12,923 PSG recordings from multiple databases. Despite a slight decrease in performance on external datasets, the model achieved an F1-score above 0.75 and an ICC greater than 0.92. Subject-wise classification models generally excel at distinguishing severe sleep apnea but struggle with lower severity cut-offs (12, 33). For instance, Levy et al. (11) reported a high misclassification rate for healthy subjects as mild apnea, particularly in the MrOS dataset (44% misclassified). Similarly, Liang (12) achieved better performance with a 30/h cut-off than 5/h, but still faced issues with misclassification. The model by Gutiérrez-Tobal et al. (33) had high sensitivity but lower specificity, indicating a tendency to overestimate severity.
Performance issues may arise from imbalanced data and binary cut-off thresholds. Ganglberger et al. (19) suggested adjusting the AHI threshold could better reflect severity, particularly near the borderline. Studies adjusting the cut-off to 10/h showed more balanced performance (7, 18, 43).
Few studies address post-processing, which is important for identifying and correcting issues not evident during initial model development. Papini et al. (58) proposed a post-processing step to reassess results based on severity discrepancies, considering factors like cardiac comorbidity and medication.
4 Discussion
Our analysis examined key aspects of AI-driven SpO2-based sleep apnea screening, including commonly used datasets, signal preprocessing methods, feature extraction and selection, and model performance. Although the results are promising, the variability in devices, algorithms, and study designs makes it difficult to draw definitive conclusions about which devices and algorithms represent the state-of-the-art. Based on our findings, we discuss the research gaps and opportunities associated with sleep apnea screening at home using SpO2 measurements.
4.1 Research gaps
Our analysis identified two major research gaps in AI-empowered sleep apnea screening using SpO2 measurements. The first gap pertains to the limitations in the quantity and diversity of datasets. Specifically, there is a scarcity of large-scale, open-access datasets. Large volumes of training data are essential for developing robust AI models, and sharing open datasets is increasingly crucial for advancing research. However, we found that most studies used proprietary datasets that are not publicly available. Despite our focus on wearable deceives for apnea detection, we identified only four datasets collected with such devices (19, 36, 53, 56), none of which are openly accessible. In addition, most of the open sleep datasets were collected in Western countries and predominantly included data from Caucasian individuals. The lack of data from diverse populations, especially Black, Hispanic, and Asian groups, poses a challenge to developing generalizable AI models across different demographic groups (11, 39).
Furthermore, in the context of home-based or self-tracking, the user range has expanded beyond patients to include healthy, young individuals in everyday life. The innovation of portable and wearable sensors has highlighted the urgent need for datasets derived from smartwatches, smart rings, and similar devices. The closer these datasets reflect real-world conditions, the more effectively machine learning algorithms can be applied in practice. Another promising trend is the use of longitudinal data, which provides a more detailed and stable representation of health status over time. In the near future, sleep apnea data from healthy young individuals, collected through consumer devices over extended periods, will be essential for advancing research and improving screening models.
The second major gap is the need for standardization in data collection, signal preprocessing, and model benchmarking. This lack of standardization creates challenges for reproducibility, comparability and generalizability. Data collected across different studies vary significantly due to differences in the type of devices used, the protocols followed (e.g., timing, duration, sensor placement), and the environmental conditions under which data are gathered. Signal preprocessing is another area where standardization is absent. Steps such as noise filtering and artifact removal are often performed differently across studies, leading to inconsistencies in the data quality. In addition, variations in datasets used for model training and testing, coupled with differences in evaluation metrics, hinder direct comparisons between models (19). Many studies focus solely on accuracy (e.g., (11, 31, 47)), which can be misleading, especially in the presence of unbalanced data. The generalizability of models is frequently overlooked; only one study extensively investigated model performance on different datasets other than the training set (37). In contrast, several studies reported decreased model performance when applied to new datasets (21, 31, 46, 47). To facilitate meaningful cross-model comparisons, sharing source code is recommended; however, only a few studies have done so (21, 37, 45).
4.2 Opportunities and future directions
To address the identified research gaps in AI-powered sleep apnea screening using SpO2 measurements, future research should focus on enhancing dataset quality and standardization. Developing and sharing large-scale, open-access datasets that include diverse populations is crucial for training robust and generalizable AI models. Collaborative data initiatives involving academia, industry, and healthcare organizations can facilitate the creation of comprehensive datasets. Additionally, establishing standardized protocols for data collection and signal preprocessing will improve reproducibility and comparability across studies. Formulating and disseminating guidelines for data collection and preprocessing, along with creating benchmarking frameworks, can help ensure consistency and facilitate meaningful cross-study comparisons.
Moreover, advancing model evaluation and generalizability is essential for improving AI performance in sleep apnea screening. Future research should emphasize evaluating models on diverse datasets and adopting robust evaluation metrics that account for data imbalance and biases. Encouraging transparency by sharing methodologies, preprocessing scripts, and source code can enhance reproducibility and foster a more reliable research environment. Exploring multi-modal data approaches and innovative sensor technologies could also improve the accuracy and robustness of models. By addressing these areas, the field can make significant strides toward developing more effective and generalizable AI solutions for sleep apnea detection.
Finally, research on AI-based sleep apnea research has been lacking a user-centered perspective, and this needs to be addressed in future studies. In addition to improving model performance, future research must also prioritize clinical applicability and user adoption (59). Keeping experts involved by collecting their feedback to continuously retrain the models can help better align the inner workings of the models with expert decision-making processes. Engaging with both patients and clinicians through pilot studies, usability testing, and observational trials can provide valuable insights into model interpretability, user trust, and practical deployment. Last but not the least, designing user-friendly interfaces that visualize model predictions in an easy-to-understand manner is important for fostering user acceptance and real-world applicability.
Author contributions
NHH: Conceptualization, Methodology, Writing – original draft, Writing – review & editing. ZL: Conceptualization, Funding acquisition, Methodology, Supervision, Validation, Writing – original draft, Writing – review & editing.
Funding
The author(s) declare that financial support was received for the research and/or publication of this article. This study was supported by the JSPS KAKENHI Grant Number 21K17670.
Conflict of interest
The authors declare that the research was carried out in the absence of commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Benjafield AV, Ayas NT, Eastwood PR, Heinzer R, Ip MS, Morrell MJ, et al. Estimation of the global prevalence and burden of obstructive sleep apnoea: a literature-based analysis. Lancet Respir Med. (2019) 7:687–98. doi: 10.1016/S2213-2600(19)30198-5
2. Gahungu N, Shariar A, Playford D, Judkins C, Gabbay E. Transfer learning artificial intelligence for automated detection of atrial fibrillation in patients undergoing evaluation for suspected obstructive sleep apnoea: a feasibility study. Sleep Med. (2021) 85:166–71. doi: 10.1016/j.sleep.2021.07.014
3. Jayawardhana M, de Chazal P. Enhanced detection of sleep apnoea using heart-rate, respiration effort and oxygen saturation derived from a photoplethysmography sensor. In: 2017 39th Annual International Conference of the IEEE Engineering in Medicine and Biology Society (EMBC). IEEE (2017). p. 121–4.
4. Sharma M, Kumbhani D, Yadav A, Acharya UR. Automated sleep apnea detection using optimal duration-frequency concentrated wavelet-based features of pulse oximetry signals. Appl Intell. (2022) 52:1325–37. doi: 10.1007/s10489-021-02422-2
5. Liang Z, Ploderer B. How does fitbit measure brainwaves: a qualitative study into the credibility of sleep-tracking technologies. Proc ACM Interact Mob Wearable Ubiquitous Technol. (2020) 4:1–29. doi: 10.1145/3380994.35846237
6. Liang Z, Ploderer B, Liu W, Nagata Y, Bailey J, Kulik L, et al. SleepExplorer: a visualization tool to make sense of correlations between personal sleep data and contextual factors. Pers Ubiquit Comput. (2016) 20:985–1000. doi: 10.1007/s00779-016-0960-6
7. Alvarez D, Hornero R, Marcos JV, Lopez M. Obstructive sleep apnea detection using clustering classification of nonlinear features from nocturnal oximetry. In: 2007 29th Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE (2007). p. 1937–40.
8. Pathinarupothi RK, Rangan ES, Gopalakrishnan E, Vinaykumar R, Soman K. Single sensor techniques for sleep apnea diagnosis using deep learning. In: 2017 IEEE international conference on healthcare informatics (ICHI). IEEE (2017). p. 524–9.
9. Xie B, Minn H. Real-time sleep apnea detection by classifier combination. IEEE Trans Inf Technol Biomed. (2012) 16:469–77. doi: 10.1109/TITB.2012.2188299
10. Bark B, Nam B, Kim IY. SelANet: decision-assisting selective sleep apnea detection based on confidence score. BMC Med Inform Decis Mak. (2023) 23:190. doi: 10.1186/s12911-023-02292-3
11. Levy J, Álvarez D, Del Campo F, Behar JA. Deep learning for obstructive sleep apnea diagnosis based on single channel oximetry. Nat Commun. (2023) 14:4881. doi: 10.1038/s41467-023-40604-3
12. Liang Z. Multiscale attention entropy (MSAE) of overnight pulse oximetry for assessing sleep apnea. In: Proceedings of the 2023 7th International Conference on Medical and Health Informatics (2023). p. 77–80.
13. Mostafa SS, Baptista D, Ravelo-García AG, Juliá-Serdá G, Morgado-Dias F. Greedy based convolutional neural network optimization for detecting apnea. Comput Methods Programs Biomed. (2020) 197:105640. doi: 10.1016/j.cmpb.2020.105640
14. Sutton A, Clowes M, Preston L, Booth A. Meeting the review family: exploring review types and associated information retrieval requirements. Health Inf Libr J. (2019) 36:202–22. doi: 10.1111/hir.12276.
15. Ouzzani M, Hammady H, Fedorowicz Z, Elmagarmid A. Rayyan—a web and mobile app for systematic reviews. Syst Rev. (2016) 5:1–10. doi: 10.1186/s13643-016-0384-4
16. Bernardini A, Brunello A, Gigli GL, Montanari A, Saccomanno N. AIOSA: an approach to the automatic identification of obstructive sleep apnea events based on deep learning. Artif Intell Med. (2021) 118:102133. doi: 10.1016/j.artmed.2021.102133
17. Zhang J, Zhang Q, Wang Y, Qiu C. A real-time auto-adjustable smart pillow system for sleep apnea detection and treatment. In: Proceedings of the 12th International Conference on Information Processing in Sensor Networks (2013). p. 179–90.
18. Alvarez D, Hornero R, Marcos JV, Penzel T, Wessel N. Prospective evaluation of logistic regression models from overnight oximetry to assist in sleep apnea diagnosis. In: 2011 11th International Conference on Intelligent Systems Design and Applications. IEEE (2011). p. 920–4.
19. Ganglberger W, Bucklin AA, Tesh RA, Da Silva Cardoso M, Sun H, Leone MJ, et al. Sleep apnea and respiratory anomaly detection from a wearable band and oxygen saturation. Sleep Breath. (2021) 26:1–12. doi: 10.1007/s11325-021-02465-2
20. Hang L-W, Wang H-L, Chen J-H, Hsu J-C, Lin H-H, Chung W-S, et al. Validation of overnight oximetry to diagnose patients with moderate to severe obstructive sleep apnea. BMC Pulm Med. (2015) 15:1–13. doi: 10.1186/s12890-015-0017-z
21. Piorecky M, Bartoň M, Koudelka V, Buskova J, Koprivova J, Brunovsky M, et al. Apnea detection in polysomnographic recordings using machine learning techniques. Diagnostics. (2021) 11:2302. doi: 10.3390/diagnostics11122302
22. Penzel T, Moody GB, Mark RG, Goldberger AL, Peter JH. The apnea-ECG database. In: Computers in Cardiology 2000. Vol. 27 (Cat. 00CH37163). IEEE (2000). p. 255–8.
23. Goldberger AL, Amaral LA, Glass L, Hausdorff JM, Ivanov PC, Mark RG, et al. Physiobank, Physiotoolkit, and Physionet: components of a new research resource for complex physiologic signals. Circulation. (2000) 101(23):e215–20.
24. Bernardini A, Brunello A, Gigli GL, Montanari A, Saccomanno N. OSASUD: a dataset of stroke unit recordings for the detection of obstructive sleep apnea syndrome. Sci Data. (2022) 9:177. doi: 10.1038/s41597-022-01272-y
25. Quan SF, Howard BV, Iber C, Kiley JP, Nieto FJ, O’Connor GT, et al. The sleep heart health study: design, rationale, and methods. Sleep. (1997) 20:1077–85. doi: 10.1093/sleep/20.12.1077
26. Young T. Rationale, design, and findings from the wisconsin sleep cohort study: toward understanding the total societal burden of sleep-disordered breathing. Sleep Med Clin. (2009) 4(1):37–46. doi: 10.1016/j.jsmc.2008.11.003
27. Redline S, Tishler PV, Tosteson TD, Williamson J, Kump K, Browner I, et al. The familial aggregation of obstructive sleep apnea. Am J Respir Crit Care Med. (1995) 151:682–7. doi: 10.1164/ajrccm/151.3_Pt_1.682
28. Blackwell T, Yaffe K, Ancoli-Israel S, Redline S, Ensrud KE, Stefanick ML, et al. Associations between sleep architecture and sleep-disordered breathing and cognition in older community-dwelling men: the osteoporotic fractures in men sleep study. J Am Geriatr Soc. (2011) 59:2217–25. doi: 10.1111/j.1532-5415.2011.03731.x
29. Chen X, Wang R, Zee P, Lutsey PL, Javaheri S, Alcántara C, et al. Racial/ethnic differences in sleep disturbances: the multi-ethnic study of atherosclerosis (MESA). Sleep. (2015) 38:877–88. doi: 10.5665/sleep.4732
30. Nitzan M, Romem A, Koppel R. Pulse oximetry: fundamentals and technology update. Med Dev Evid Res. (2014) 7:231–9. doi: 10.2147/MDER.S47319
31. Singtothong C, Siriborvornratanakul T. Deep-learning based sleep apnea detection using sleep sound, SpO2, and pulse rate. Int J Inf Tecnol. (2024) 16:4869–74. doi: 10.1007/s41870-024-01906-x
32. Ma B, Wu Z, Li S, Benton R, Li D, Huang Y, et al. Development of a support vector machine learning and smart phone internet of things-based architecture for real-time sleep apnea diagnosis. BMC Med Inform Decis Mak. (2020) 20:1–13. doi: 10.1186/s12911-020-01329-1
33. Gutiérrez-Tobal GC, Álvarez D, Vaquerizo-Villar F, Crespo A, Kheirandish-Gozal L, Gozal D, et al. Ensemble-learning regression to estimate sleep apnea severity using at-home oximetry in adults. Appl Soft Comput. (2021) 111:107827. doi: 10.1016/j.asoc.2021.107827
34. Stuban N, Niwayama M. Optimal filter bandwidth for pulse oximetry. Rev Sci Instrum. (2012) 83:1–6. doi: 10.1063/1.4759491
35. Brouse CJ, Gatzke R, Freeman D, Chen Y. Increasing the dynamic range of a pulse oximeter using heart rate characteristics. In: Computing in Cardiology 2014. IEEE (2014). p. 793–6.
36. Gutiérrez-Tobal GC, Álvarez D, Crespo A, Del Campo F, Hornero R. Evaluation of machine-learning approaches to estimate sleep apnea severity from at-home oximetry recordings. IEEE J Biomed Health Inform. (2018) 23:882–92. doi: 10.1109/JBHI.2018.2823384
37. Levy J, Álvarez D, Rosenberg AA, Alexandrovich A, Del Campo F, Behar JA. Digital oximetry biomarkers for assessing respiratory function: standards of measurement, physiological interpretation, and clinical use. NPJ Digit Med. (2021) 4:1. doi: 10.1038/s41746-020-00373-5
38. Rahman MJ, Morshed BI. A minimalist method toward severity assessment and progression monitoring of obstructive sleep apnea on the edge. ACM Trans Comput Healthc (HEALTH). (2021) 3:1–16. doi: 10.1145/3479432
39. Li Z, Li Y, Zhao G, Zhang X, Xu W, Han D. A model for obstructive sleep apnea detection using a multi-layer feed-forward neural network based on electrocardiogram, pulse oxygen saturation, and body mass index. Sleep Breath. (2021) 25:1–8. doi: 10.1007/s11325-021-02302-6
40. Sánchez-Morillo D, López-Gordo M, León A. Novel multiclass classification for home-based diagnosis of sleep apnea hypopnea syndrome. Expert Syst Appl. (2014) 41:1654–62. doi: 10.1016/j.eswa.2013.08.062
41. Wu C-H, Lee J-H, Kuo TB, Lai C-T, Li LP, Yang CC. Improving the diagnostic ability of the sleep apnea screening system based on oximetry by using physical activity data. J Med Biol Eng. (2020) 40:858–67. doi: 10.1007/s40846-020-00566-z
42. Yadollahi A, Moussavi Z. Acoustic obstructive sleep apnea detection. In: 2009 Annual International Conference of the IEEE Engineering in Medicine and Biology Society. IEEE (2009). p. 7110–3.
43. Mostafa SS, Carvalho JP, Morgado-Dias F, Ravelo-García A. Optimization of sleep apnea detection using SpO and ANN. In: 2017 XXVI International Conference on Information, Communication and Automation Technologies (ICAT). IEEE (2017). p. 1–6.
44. Mencar C, Gallo C, Mantero M, Tarsia P, Carpagnano GE, Foschino Barbaro MP, et al. Application of machine learning to predict obstructive sleep apnea syndrome severity. Health Inform J. (2020) 26:298–317. doi: 10.1177/1460458218824725
45. Rodrigues Jr JF, Pepin J-L, Goeuriot L, Amer-Yahia S. An extensive investigation of machine learning techniques for sleep apnea screening. In: Proceedings of the 29th ACM International Conference on Information & Knowledge Management (2020). p. 2709–16.
46. Albuhayri MA. CNN model for sleep apnea detection based on SpO signal. Comput Inf Sci. (2023) 16:1–39. doi: 10.5539/cis.v16n1p39
47. Mostafa SS, Mendonca F, Ravelo-Garcia AG, Juliá-Serdá GG, Morgado-Dias F. Multi-objective hyperparameter optimization of convolutional neural network for obstructive sleep apnea detection. IEEE Access. (2020) 8:129586–99. doi: 10.1109/ACCESS.2020.3009149
48. Lyden B, Dair Z, O’Reilly R. Classification of sleep apnea via SpO in a simulated smartwatch environment. In: 2022 IEEE-EMBS International Conference on Biomedical and Health Informatics (BHI). IEEE (2022). p. 1–5.
49. Brennan HL, Kirby SD. Barriers of artificial intelligence implementation in the diagnosis of obstructive sleep apnea. J Otolaryngol Head Neck Surg. (2022) 51:16. doi: 10.1186/s40463-022-00566-w
50. Rolón RE, Gareis IE, Larrateguy LD, Di Persia LE, Spies RD, Rufiner HL. Automatic scoring of apnea and hypopnea events using blood oxygen saturation signals. Biomed Signal Process Control. (2020) 62:102062. doi: 10.1016/j.bspc.2020.102062
51. Deviaene M, Testelmans D, Buyse B, Borzée P, Van Huffel S, Varon C. Automatic screening of sleep apnea patients based on the SpO signal. IEEE J Biomed Health Inform. (2018) 23:607–17. doi: 10.1109/JBHI.2018.2817368
52. Hwang SH, Cho JG, Choi BH, Baek HJ, Lee YJ, Jeong D-U, et al. Real-time automatic apneic event detection using nocturnal pulse oximetry. IEEE Trans Biomed Eng. (2017) 65:706–12. doi: 10.1109/TBME.2017.2715405
53. Bilge PC, Fatma EK, Cansu S, Haydar S, Deniz K, Alisher K, et al. Scanning of obstructive sleep apnea syndrome using smartwatch: a comparison of smartwatch and polysomnography. J Clin Neurosci. (2024) 119:212–9. doi: 10.1016/j.jocn.2023.12.009
54. Chen J-W, Liu C-M, Wang C-Y, Lin C-C, Qiu K-Y, Yeh C-Y, et al. A deep neural network-based model for OSA severity classification using unsegmented peripheral oxygen saturation signals. Eng Appl Artif Intell. (2023) 122:106161. doi: 10.1016/j.engappai.2023.106161
55. Mostafa SS, Mendonça F, Morgado-Dias F. A systematic review of detecting sleep apnea using deep learning. Sensors. (2019) 19:4934. doi: 10.3390/s19224934
56. Jeon Y, Heo K, Kang SJ. Real-time sleep apnea diagnosis method using wearable device without external sensors. In: 2020 IEEE International Conference on Pervasive Computing and Communications Workshops (PerCom Workshops). IEEE (2020). p. 1–5.
57. Punjabi NM. Counterpoint: is the apnea-hypopnea index the best way to quantify the severity of sleep-disordered breathing? No. Chest. (2016) 149:16–9. doi: 10.1378/chest.14-2261
58. Papini GB, Fonseca P, van Gilst MM, Bergmans JW, Vullings R, Overeem S. Wearable monitoring of sleep-disordered breathing: estimation of the apnea–hypopnea index using wrist-worn reflective photoplethysmography. Sci Rep. (2020) 10:13512. doi: 10.1038/s41598-020-69935-7
Keywords: sleep apnea, SpO2, oximeter, mobile health (mHealth), digital health, machine learning, deep learning, apnea-hypopnea index
Citation: Hoang NH and Liang Z (2025) AI-driven sleep apnea screening with overnight blood oxygen saturation: current practices and future directions. Front. Digit. Health 7:1510166. doi: 10.3389/fdgth.2025.1510166
Received: 12 October 2024; Accepted: 24 March 2025;
Published: 9 April 2025.
Edited by:
Feng Liu, Stevens Institute of Technology, United StatesReviewed by:
Sayonara de Fatima Faria Barbosa, University of Cincinnati, United StatesCopyright: © 2025 Hoang and Liang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Nhung H. Hoang, bmh1bmc5NzEzQGdtYWlsLmNvbQ==