- 1Geriatric Research Education and Clinical Center and Sleep Medicine Section, Louis Stokes Cleveland VA Medical Center, Cleveland, OH, United States
- 2School of Medicine, Case Western Reserve University, Cleveland, OH, United States
- 3Department of Quantitative Sciences, Lerner Research Institute Cleveland Clinic, Cleveland, OH, United States
Background: Adherence to positive airway pressure (PAP) therapy for sleep apnea is suboptimal, particularly in the veteran population. Accurately identifying those best suited for other therapy or additional interventions may improve adherence. We evaluated various machine learning algorithms to predict 90-day adherence.
Methods: The cohort of VA Northeast Ohio Health Care system patients who were issued a PAP machine (January 1, 2010–June 30, 2015) had demographics, comorbidities, and medications at the time of polysomnography obtained from the electronic health record. The data were split 60:20:20 into training, calibration, and validation data sets, with no use of validation data for model development. We constructed models for the first 90-day adherence period (% nights ≥4 h use) using the following algorithms: linear regression, least absolute shrinkage and selection operator, elastic net, ridge regression, gradient boosted machines, support vector machine regression, Bayes-based models, and neural nets. Prediction performance was evaluated in the validation data set using root mean square error (RMSE).
Results: The 5,047 participants were 38.3 ± 11.9 years old, and 96.1% male, with 36.8% having coronary artery disease and 52.6% with depression. The median adherence was 36.7% (interquartile range: 0%, 86.7%). The gradient boosted machine was superior to other machine learning techniques (RMSE 37.2). However, the performance was similar and not clinically useful for all models without 30-day data. The 30-day PAP data and using raw diagnoses and medications (vs. grouping by type) improved the RMSE to 24.27.
Conclusion: Comparing multiple prediction algorithms using electronic medical record information, we found that none has clinically meaningful performance. Better adherence predictive measures may offer opportunities for personalized tailoring of interventions.
1 Introduction
Obstructive sleep apnea (OSA) affects an estimated 26% of the U.S. population and 47% of veterans with sleep disorders (Peppard et al., 2013; Alexander et al., 2016). OSA causes daytime dysfunction, decreased quality of life, and increased rates of morbidity and mortality, particularly in the veteran population (Guggisberg et al., 2007; Alexander et al., 2016).
Consistent use of positive airway pressure (PAP) therapy, the mainstay of therapy, is effective in reversing OSA pathophysiology. PAP therapy improves blood pressure control, cardiac remodeling, measures of cardiac electrophysiological derangement, and atrial fibrillation recurrence after ablation and cardioversion (Bonsignore et al., 2002; Baranchuk, 2012; Colish et al., 2012; Baranchuk et al., 2013; Gottlieb et al., 2014; Li et al., 2014; Campos-Rodriguez et al., 2017). Treatment impacts patient-centered outcomes, including improved sleep quality, decreased daytime sleepiness, and improved quality of life. Perhaps most impressive, OSA therapy decreased overall health care resource utilization and costs in Canada, a country with nationalized health care, making the lack of health care access unlikely to account for the effect (Albarrak et al., 2005). Taken in toto, therapy for OSA not only decreases morbidity but may also decrease health care costs.
OSA therapy is effective. However, adherence to PAP therapy is low-−24% stopped therapy within 3 months, with another 23% being non-adherent with therapy after 1 year (Aloia et al., 2008). Long known to be a problem, adherence has not improved substantially over the past 20 years, despite attempts of intensive interventions (Rotenberg et al., 2016). Patients, providers, and the health care system would reap enhanced efficiency and cost savings by identifying and focusing on those most likely to become non-adherent to therapy.
Although many studies describe associations between patient characteristics or sleep study data and adherence, to our knowledge, no predictive model has been developed. Improving prediction could guide resource allocation and improve outcomes. We developed, validated, and compared several machine learning algorithms to predict adherence within the first 90 days (percentage nights used at least 4 h a night). We hypothesize that screening for non-adherence can be developed using a subset of commonly collected electronic medical record (EMR) variables. We additionally hypothesize that comorbidities, particularly cardiac and psychiatric comorbidities, contribute most to adherence prediction model performance.
2 Methods
2.1 Participants and study design
This retrospective cohort study evaluated adult veteran sleep medicine clinic patients of the VA (Veterans Administration) Northeast Ohio Health Care system. There were 5,548 PAP adherence monitoring records started between January 1, 2010, and June 30, 2015 (Figure 1). Of the 5,174 unique records, 5,047 were able to be matched with VA electronic medical records. The VA Northeast Ohio Health Care system institutional review board found this retrospective study exempt from review.
2.2 PAP adherence
All individuals with records of PAP disbursement had ResMed© PAP machines with electronic modems. Adherence data were automatically downloaded every morning to EncoreAnywhere ©. Adherence was defined as the percentage use of at least 4 h a day in the first 90-day period after their PAP disbursement. In addition, we collected 30-day adherence and efficacy metrics, including the percentage of days used more than 4 h/night, mean and 90th percentile pressure, residual apnea–hypopnea index, residual central apnea index, Hunter–Cheyne–Stokes respiration, and leak.
2.3 Other measures
All other variables were obtained from the electronic medical record from the time of PAP disbursement or, if not available at that time, the time closest to PAP disbursement. Height and weight were used to compute the body mass index (kg/m2). Race was categorized as white, black, or other. Participants' past medical history (obesity, cardiac disease, psychiatric diagnoses, other sleep disorders, diabetes, chronic kidney disease, dementia, stroke, liver disease, pain syndromes, alcohol abuse, and tobacco use) and medications (benzodiazepines, benzodiazepine receptor agonists, other sleep medications, tricyclic antidepressants, antidepressants, antipsychotics, mood stabilizers, α2δ ligands, and stimulants) were obtained. Cardiac disease was defined as a diagnosis of arrhythmia, heart failure, and/or coronary artery disease. Psychiatric disease was defined as schizophrenia, psychosis disorders, depression, posttraumatic stress disorder (PTSD), bipolar disorder, anxiety, and obsessive-compulsive disorder. Other sleep disorders included insomnia, hypersomnia, hypoventilation, circadian rhythm disorders, parasomnia, movement disorders, and narcolepsy. Other sleep medications included ramelteon, diphenhydramine, melatonin, doxepin, mirtazapine, and trazodone. Antidepressants included serotonin reuptake inhibitors, serotonin norepinephrine reuptake inhibitors, mirtazapine, trazodone, and nefazodone. Stimulants included modafinil, armodafinil, methylphenidate, and dextroamphetamine.
3 Statistical analyses
Participant characteristics were summarized as mean ± standard deviation (SD), median (interquartile range, IQR), or n (%). Standardized mean differences were calculated between adherent and non-adherent participants.
3.1 Developing and optimizing adherence prediction models
The cohort was randomly partitioned 60:20:20 into training (n = 4,039), calibration (n = 1,008), and validation data sets (n = 1,008). Only the training and calibration data sets were used in developing the models, including cross-validation to select hyperparameters. Multiple multivariate imputation via chained equations using a random forest algorithm in the mice package was used to estimate missing data (Buuren and Groothuis-Oudshoorn, 2011; Li et al., 2015). Adherence as a continuous variable of the percentage use ≥4 h a night (%use) was the outcome. All other variables (features) were included in the model based on clinical and physiologic plausibility rather than variable selection via machine learning approaches for feature selection. Models tested included linear regression, least absolute shrinkage and selection operator (LASSO), Bayesian LASSO, elastic net, random forest, conditional inference random forest, gradient boosted machine (GBM), support vector machine (SVM) regression (both linear kernel and radial basis function kernel), spike and slab regression, and several neural network models—flat and multilayer neural networks (variable number and size of hidden layers, initialization weights, initial learning rates, and optimizer algorithms), model-averaged neural networks, Bayesian regularized neural networks, and neural networks with feature extractions. The caret package was used to cross-validate parameter estimates and optimize hyperparameter values using 10-fold cross-validation performed 4 times (Kuhn, 2008, 2019). A grid search was used to identify the model with a minimum cross-validated root mean square error (RMSE) (Kuhn, 2008, 2019). The model parameters tested are found in Table 1. These models were then recalibrated via the Dalton method (Dalton, 2013), using the calibration data set. Briefly, adherence was calculated for the calibration data set using the model developed in the training data set. The difference between the actual and predicted adherence was then calculated. A natural spline regression of the difference between actual and calculated adherence (dependent variable) vs. the predicted adherence (independent variable) to find offset parameters to calibrate the model. Model performance was evaluated in the validation data set with RMSE as the primary measure and mean absolute error (MAE) as the secondary measure. A calibration slope and intercept were evaluated from models fit on the validation data set to evaluate calibration (Alba et al., 2017).
3.2 Secondary analyses
We evaluated which group of variables had the highest impact on prediction performance. The best model was built on the training data set by removing from the main analysis groups of variables as follows: psychiatric medications, psychiatric comorbidity, cardiovascular comorbidity, pain comorbidity, sleep comorbidity, and tobacco and alcohol use/abuse. Analyses were evaluated in the validation data set with RMSE and MAE. In addition, we examined the effect of adding 30-day adherence data on model performance as this is a common time point to check in on therapy adherence. We also examined the models' performance on non-grouped (raw) data.
All analyses were conducted using R software version 3.4.3 (R Core Development Team, Vienna, Austria) (Core Team, 2014). This analysis used the mice, caret, glmnet, randomForest, and tidyverse packages extensively (Liaw and Wiener, 2002; Buuren and Groothuis-Oudshoorn, 2011; Simon et al., 2011; Fritsch et al., 2019; Kuhn, 2019; Wickham et al., 2019).
4 Results
4.1 Study population
Table 2 reports the baseline characteristics of the cohort. The 5,047 participants in the overall cohort were 38.3 ± 11.9 years old, and 3.9% female. The 90-day adherence distribution was bimodal with modes at 0% and 100% use >4 h/night; participant adherence at 90 days was 44.0% ± 40.4% (mean ± SD), with 36.9% of participants adherent by the Centers of Medicare & Medicaid Services standard of at least 70% of nights used at least 4 h a night. Among those who were adherent based on this standard, mean use >4 h/night was 91.2% ± 10.6%, whereas non-adherent individuals used PAP at least 4 h/night on 16.5% ± 21.6% of nights. The cohort had a high number of people with cardiometabolic comorbidity: 62.5% with obesity, 49.3% with diabetes, and 36.8% with coronary artery disease. In addition, there was notable psychiatric comorbidity as evidenced by 52.6% with depression, 29.6% with anxiety, and 25.0% with PTSD history. There was also substantial overlap in the psychiatric diagnoses, with 37.2% of the cohort with two or more diagnoses. The non-adherent participants were significantly younger (aged 37.9 ± 12.2 years vs. 40.4 ± 11.2 years) with more history of psychiatric comorbidity, alcohol abuse (24.5% vs. 13.8%), and tobacco use (40.1% vs. 29.6%). Non-adherent participants had a substantially increased use of benzodiazepines, antipsychotics, and mood stabilizers. Cardiac comorbidities were evenly balanced across groups.
4.2 Adherence model performance
The best model after hyperparameter tuning was GBM, with an RMSE of 37.2 and an MAE of 33.1 (Table 3, Figure 2); this indicates that predictions by this model were off by 37% on average from the actual value. However, other models including the least sophisticated—cross-validated linear regression—were only slightly worse on this data set. The GBM performed the best in models that included 30-day adherence and PAP effectiveness information (Table 4). Models with 30-day PAP data performed substantially better than all other models. However, removing features (e.g., psychiatric comorbidity) did not substantially affect model performance (Table 5). Models that did not group variables (e.g., cardiac comorbidity vs. individual diagnoses of heart failure, coronary artery disease, etc.) but instead used each diagnosis performed better than those with grouped variables (RMSE 37.2 in GBM with grouped variables vs. RMSE 36 for the raw data model, including 102 features) (Table 6). Models that were restricted by eliminating variables with near-zero variance were not substantially worse than the full feature set (42 variables, GBM RMSE 36.1).
Table 3. Performance characteristics of machine learning algorithms for 90-day adherence in the validation cohort*.
Table 4. Performance characteristics of machine learning algorithms for 90-day adherence with additional 30-day adherence information in the validation cohort*.
Table 6. Performance characteristics of GBM algorithm in models with individual diagnoses and medications (raw model), individual diagnoses and indications with near-zero variance features removed, and grouped diagnoses and medication classes*.
5 Discussion
This investigation compared several machine learning algorithms for predicting adherence to PAP therapy in a veteran population using robust methodologies. We found the GBM algorithm to be incrementally the most accurate model after hyperparameter tuning. However, other evaluated models performed similarly, and none was deemed clinically distinct at this point. Even though model performance overall was subpar, these models could be used to identify those people who are highly likely to be extremely adherent (and therefore need limited intervention) or not use their PAP at all (and, therefore, be candidates for alternative therapy). Adding 30-day adherence data substantially improved the accuracy of the predictions for all models; however, there is a high probability of data leakage because 30-day adherence is also included in the outcome, 90-day adherence. Additional model improvements would be expected if sleep study, symptom, and contextual factors were added to models. This is the first study to our knowledge to (1) try to predict adherence as opposed to finding associations with variables, (2) compare multiple machine learning algorithms using reproducible methods to evaluate a sleep condition, and (3) use data that were readily available from the EMR to make such predictions.
The next logical step to consider is whether incorporating other data improves predictive performance. Work in medication adherence has consistently shown that prior adherence to medications is a good predictor of future medication adherence (Muntner et al., 2014; Kumamaru et al., 2018; Zullig et al., 2019). Therefore, information on adherence to medication and no-show rates to scheduled medical visits may improve prediction accuracy. However, other variables, such as age, although strongly associated with adherence, had no ability to discriminate between who would or would not be adherent. Attitudes about therapy are associated with adherence and may aid in determining who will or will not be adherent (Balachandran et al., 2013). Perhaps the addition of a point-of-care questionnaire on attitudes before the start of PAP therapy could improve adherence prediction accuracy. Furthermore, integrating social determinants of health may improve prediction accuracy as shown in models of cardiovascular risk (Dalton et al., 2017).
Any adherence prediction model that is developed for VA data will need to be externally validated because there are inherent and dramatic differences between the VA population and the non-VA general sleep population. In the VA, there is a very high percentage of men, a non-representative racial mix, and much more psychiatric comorbidity. Hence, these findings may not be representative of EMR prediction algorithm performance in other populations. In particular, model discrimination can be lower when the risk profile of the analyzed population is relatively homogeneous, as may be the case with the population in this analysis.
Given the recent popularity of neural networks, it is perhaps counterintuitive that they were not able to accurately predict veterans' future adherence and that GBM—an ensemble of decision trees—had superior performance. However, neural networks are not a panacea. Not only is algorithm performance very dependent on the data and field for which it is being developed, but there is also evidence that tree-based models may be an improvement on neural networks in several situations (Fernández-Delgado et al., 2014). SVM outperformed neural networks in both binary classification for image identification and corporate bankruptcy prediction (Moghaddam and Yang, 2001; Shin et al., 2005) and multiclass prediction problems such as financial time-series forecasting and protein folding (Ding and Dubchak, 2001; Tay and Cao, 2001). Neural networks work best with large training sets in cases in which there are complex hierarchical relationships. In addition, there are numerous hyperparameters for neural nets, and hyperparameter value selection is complicated. Finding the ideal architecture by tuning the numerous hyperparameters inherent to neural networks is not intuitive, and even a grid search is not guaranteed to guide one in the optimal direction for hyperparameter optimization. Furthermore, neural networks are prone to finding local, rather than global, minima, and overfitting. This may be why when testing 179 algorithms in 121 data sets, tree-based methods were found to be the best family of models (Fernández-Delgado et al., 2014). In addition, the interpretability of models is crucial in medicine to detect bias and build trust in the model. Neural networks suffer from poor interpretability. Even small changes in parameters can lead to substantial shifts in model output. Because neural networks are often non-intuitive and opaque—with no clear explanation for why one set of variables slightly differing from another yielded substantially different results—using these models would be hard to justify in medicine in their current form. However, there is fervor in the community to develop tools to allow a better understanding of model decision-making (Teng et al., 2022). Despite the recent popularity of neural networks, they are not the best tool for every data analytic job.
Several study strengths and limitations are worth noting. This study used a separate training and validation data set (no part of the validation data set was used in any part of the training algorithm), which safeguards validity. The present study enrolled a broad cohort of every eligible patient who received a PAP machine, regardless of whether they used it or not. This large data set allowed advanced machine learning models to be used, such as neural nets, which smaller data sets would have difficulty converging. This expansive approach of gathering the entire clinical population given PAP therapy may improve performance in other clinical populations, at least within the VA. Furthermore, data included a limited subset of features that are readily gleaned from any EMR, giving the ability to implement such a method at scale in health systems. The evaluation of multiple machine learning algorithms allows for comparison in the validation data set. Several limitations are also worth noting. The cohort is from a VA population, which is largely male with a higher psychiatric and cardiac comorbidity; therefore, this veteran cohort may not be generalizable to the broader OSA population. This EMR data set misses potentially important predictors by not assessing patient symptoms or sleep study data. Medical treatment has evolved over the 10-year span that was sampled in this study and may affect adherence patterns. Additional bias in the data may incorrectly classify people because it lacks contextual data such as beliefs, socioeconomic status, and implicit bias in health care; this concern has higher prominence in data sets with limited diversity (e.g., low number of women in the present data set). If used for decision-making, further validation of such algorithms is paramount. It is notoriously difficult to understand the decision process of neural net models, and this is a detractor in medicine because the ability to explain the process is often important in understanding and improving the error rate.
Study findings could be broadened by examining populations enriched with women and minorities. Reevaluation and redevelopment of an adherence algorithm in larger, diverse cohorts may yet show the utility of these variables in determining future adherence. Future work may focus on identifying subgroups within the non-adherent population to guide the provision of enhanced interventions to promote adherence or encourage the use of a different treatment modality. Further refinement of adherence models with additional EMR data (e.g., anthropometry, physiologic variables, and past adherence to medications) may yield a more potent ability to predict future adherence. Adding contextual and social factors, such as marital status, presence of a bed partner, socioeconomic variables, and attitudes, could also improve prediction performance.
We were able to develop and calibrate a model for PAP adherence using EMR data. After evaluating several machine learning algorithms, including neural nets, we found GBM to be superior in predicting 90-day PAP therapy adherence; however, no model had sufficient performance for clinical application. Further work on refining the feature set to predict adherence is warranted. Improving the prediction of future adherence may open avenues for personalized therapy at OSA diagnosis.
Data availability statement
The data analyzed in this study is subject to the following licenses/restrictions: adherence data is provided by Philips Respironics. VA Privacy Officer would need to review and clear all datasets before release. Requests to access these datasets should be directed to ZHJhbm5hbWF5JiN4MDAwNDA7Z21haWwuY29t.
Ethics statement
The requirement of ethical approval was waived by Louis Stokes Cleveland VA Medical Center Institutional Review Board for the studies involving humans because study is exempt per 45 CFR 46.104(d) (4). The studies were conducted in accordance with the local legislation and institutional requirements. The Ethics Committee/institutional review board also waived the requirement of written informed consent for participation from the participants or the participants' legal guardians/next of kin because retrospective chart review with minimal risk could not reasonably be done with informed consent.
Author contributions
AM: Conceptualization, Formal analysis, Funding acquisition, Project administration, Resources, Writing—original draft, Writing—review & editing. JD: Conceptualization, Methodology, Supervision, Writing—review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This study was supported by a Clinical Science Research and Development Career Development Award IK2CX001882 from the United States (U.S.) Department of Veterans Affairs Clinical Sciences Research and Development Service. JD was also supported by the National Institutes of Aging R01AG055480. The funding sources had no role in the design and conduct of the study; collection, management, analysis, and interpretation of the data; or the decision to submit the manuscript for publication. There were no off-label or investigational use of drugs or devices used in this study. The contents of this work do not represent the views of the Department of Veterans Affairs or the United States government.
Acknowledgments
We wish to thank Christy Stitt from Philips Respironics for obtaining and organizing the adherence data for this project.
Conflict of interest
AM receives consulting fees from Liva Nova PLC for DSMB board participation.
The remaining author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Alba, A. C., Agoritsas, T., Walsh, M., Hanna, S., Iorio, A., Devereaux, P. J., et al. (2017). Discrimination and calibration of clinical prediction models: users' guides to the medical literature. JAMA. 318, 1377–1384. doi: 10.1001/jama.2017.12126
Albarrak, M., Banno, K., Sabbagh, A. A., Delaive, K., Walld, R., Manfreda, J., et al. (2005). Utilization of healthcare resources in obstructive sleep apnea syndrome: a 5-year follow-up study in men using CPAP. Sleep. 28, 1306–1311. doi: 10.1093/sleep/28.10.1306
Alexander, M., Ray, M. A., Hébert, J. R., Youngstedt, S. D., Zhang, H., Steck, S. E., et al. (2016). The national veteran sleep disorder study: descriptive epidemiology and secular trends, 2000-2010. Sleep. 39, 1399–1410. doi: 10.5665/sleep.5972
Aloia, M. S., Goodwin, M. S., Velicer, W. F., Arnedt, J. T., Zimmerman, M., Skrekas, J., et al. (2008). Time series analysis of treatment adherence patterns in individuals with obstructive sleep apnea. Ann. Behav. Med. Publ. Soc. Behav. Med. 36, 44–53. doi: 10.1007/s12160-008-9052-9
Balachandran, J. S., Yu, X., Wroblewski, K., and Mokhlesi, B. (2013). A brief survey of patients' first impression after CPAP titration predicts future CPAP adherence: a pilot study. J. Clin. Sleep Med. 9, 199–205. doi: 10.5664/jcsm.2476
Baranchuk, A. (2012). Sleep apnea, cardiac arrhythmias, and conduction disorders. J. Electrocardiol. 45, 508–512. doi: 10.1016/j.jelectrocard.2012.03.003
Baranchuk, A., Pang, H., Seaborn, G. E. J., Yazdan-Ashoori, P., Redfearn, D. P., Simpson, C. S., et al. (2013). Reverse atrial electrical remodelling induced by continuous positive airway pressure in patients with severe obstructive sleep apnoea. J. Interv. Card Electrophysiol. 36, 247–253. doi: 10.1007/s10840-012-9749-3
Bonsignore, M. R., Parati, G., Insalaco, G., Marrone, O., Castiglioni, P., Romano, S., et al. (2002). Continuous positive airway pressure treatment improves baroreflex control of heart rate during sleep in severe obstructive sleep apnea syndrome. Am. J. Respir. Crit. Care Med. 166, 279–286. doi: 10.1164/rccm.2107117
Buuren, S., and Groothuis-Oudshoorn, K. (2011). Mice: multivariate imputation by chained equations in R. J. Stat. Softw. 45, 1–67. doi: 10.18637/jss.v045.i03
Campos-Rodriguez, F., Gonzalez-Martinez, M., Sanchez-Armengol, A., Jurado-Gamez, B., Cordero-Guevara, J., Reyes-Nuñez, N., et al. (2017). Effect of continuous positive airway pressure on blood pressure and metabolic profile in women with sleep apnoea. Eur. Respir. J. 50:1700257. doi: 10.1183/13993003.00257-2017
Colish, J., Walker, J. R., Elmayergi, N., Almutairi, S., Alharbi, F., Lytwyn, M., et al. (2012). Obstructive sleep apnea: effects of continuous positive airway pressure on cardiac remodeling as assessed by cardiac biomarkers, echocardiography, and cardiac MRI. Chest. 141, 674–681. doi: 10.1378/chest.11-0615
Core Team R. (2014). R: A language and environment for statistical computing. R Foundation for Statistical Computing, Vienna, Austria. 2013. ISBN 3-900051-07-0. Available online at: http://www.R-project.org/ (accessed September 24, 2022).
Dalton, J. E. (2013). Flexible recalibration of binary clinical prediction models. Stat. Med. 32, 282–289. doi: 10.1002/sim.5544
Dalton, J. E., Perzynski, A. T., Zidar, D. A., Rothberg, M. B., Coulton, C. J., Milinovich, A. T., et al. (2017). Accuracy of cardiovascular risk prediction varies by neighborhood socioeconomic position: a retrospective cohort study. Ann. Intern. Med. 167, 456–464. doi: 10.7326/M16-2543
Ding, C. H. Q., and Dubchak, I. (2001). Multi-class protein fold recognition using support vector machines and neural networks. Bioinformatics. 17, 349–358. doi: 10.1093/bioinformatics/17.4.349
Fernández-Delgado, M., Cernadas, E., Barro, S., and Amorim, D. (2014). Do we need hundreds of classifiers to solve real world classification problems? J. Mach. Learn. Res. 15, 3133–3181.
Fritsch, S., Guenther, F., and Wright, M. N. (2019). Neuralnet: Training of Neural Networks. Available online at: https://CRAN.R-project.org/package=neuralnet (accessed September 24, 2022).
Gottlieb, D. J., Punjabi, N. M., Mehra, R., Patel, S. R., Quan, S. F., Babineau, D. C., et al. (2014). CPAP versus oxygen in obstructive sleep apnea. N. Engl. J. Med. 370, 2276–2285. doi: 10.1056/NEJMoa1306766
Guggisberg, A. G., Hess, C. W., and Mathis, J. (2007). The significance of the sympathetic nervous system in the pathophysiology of periodic leg movements in sleep. Sleep. 30, 755–766. doi: 10.1093/sleep/30.6.755
Kuhn, M. (2008). Building predictive models in r using the caret package. J. Stat. Softw. 28, 1–26. doi: 10.18637/jss.v028.i05
Kuhn, M. (2019). Caret: Classification and Regression Training. Available online at: https://CRAN.R-project.org/package=caret (accessed September 24, 2022).
Kumamaru, H., Lee, M. P., Choudhry, N. K., Dong, Y. H., Krumme, A. A., Khan, N., et al. (2018). Using previous medication adherence to predict future adherence. J. Manag. Care Spec. Pharm. 24, 1146–1155. doi: 10.18553/jmcp.2018.24.11.1146
Li, L., Wang, Z. W., Li, J., Ge, X., Guo, L. Z., Wang, Y., et al. (2014). Efficacy of catheter ablation of atrial fibrillation in patients with obstructive sleep apnoea with and without continuous positive airway pressure treatment: a meta-analysis of observational studies. Europace. 16, 1309–1314. doi: 10.1093/europace/euu066
Li, P., Stuart, E. A., and Allison, D. B. (2015). Multiple imputation: a flexible tool for handling missing data. JAMA. 314, 1966–1967. doi: 10.1001/jama.2015.15281
Moghaddam, B., and Yang, M. (2001). Sex with Support Vector Machines. In: Advances in Neural Information Processing 13 (MIT Press) p. 960–6.
Muntner, P., Yun, H., Sharma, P., Delzell, E., Kent, S. T., Kilgore, M. L., et al. (2014). Ability of low antihypertensive medication adherence to predict statin discontinuation and low statin adherence in patients initiating treatment after a coronary event. Am. J. Cardiol. 114, 826–831. doi: 10.1016/j.amjcard.2014.06.009
Peppard, P. E., Young, T., Barnet, J. H., Palta, M., Hagen, E. W., Hla, K. M., et al. (2013). Increased prevalence of sleep-disordered breathing in adults. Am. J. Epidemiol. 177, 1006–1014. doi: 10.1093/aje/kws342
Rotenberg, B. W., Murariu, D., and Pang, K. P. (2016). Trends in CPAP adherence over twenty years of data collection: a flattened curve. J. Otolaryngol. 45:43. doi: 10.1186/s40463-016-0156-0
Shin, K. S., Lee, T. S., and Kim, H. (2005). Jung: an application of support vector machines in bankruptcy prediction model. Expert. Syst. Appl. 28, 127–135. doi: 10.1016/j.eswa.2004.08.009
Simon, N., Friedman, J., Hastie, T., and Tibshirani, R. (2011). Regularization paths for cox's proportional hazards model via coordinate descent. J. Stat. Softw. 39, 1–13. doi: 10.18637/jss.v039.i05
Tay, F. E. H., and Cao, L. (2001). Application of support vector machines in financial time series forecasting. Omega. 29, 309–317. doi: 10.1016/S0305-0483(01)00026-3
Teng, Q., Liu, Z., Song, Y., Han, K., and Lu, Y. (2022). A survey on the interpretability of deep learning in medical diagnosis. Multimed. Syst. 28, 2335–2355. doi: 10.1007/s00530-022-00960-4
Wickham, H., Averick, M., Bryan, J., Chang, W., McGowan, L. D., François, R., et al. (2019). Welcome to the tidyverse. J. Open Sour. Softw. 4:1686. doi: 10.21105/joss.01686
Keywords: sleep apnea, machine learning, adherence, positive airway pressure, compliance
Citation: May AM and Dalton JE (2024) Comparison of machine learning approaches for positive airway pressure adherence prediction in a veteran cohort. Front. Sleep 3:1278086. doi: 10.3389/frsle.2024.1278086
Received: 15 August 2023; Accepted: 29 January 2024;
Published: 16 February 2024.
Edited by:
Stuart F. Quan, Harvard Medical School, United StatesReviewed by:
Samuel Huang, Virginia Commonwealth University, United StatesAlyssa Huff, Seattle Children's Research Institute, United States
Copyright © 2024 May and Dalton. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Anna M. May, ZHJhbm5hbWF5JiN4MDAwNDA7Z21haWwuY29t