Skip to main content

PERSPECTIVE article

Front. Psychiatry, 18 January 2024
Sec. Perinatal Psychiatry
This article is part of the Research Topic Women in Psychiatry 2023: Perinatal Psychiatry View all 11 articles

Recentering responsible and explainable artificial intelligence research on patients: implications in perinatal psychiatry

  • 1School of Nursing, Columbia University School of Nursing, New York, NY, United States
  • 2Department of Psychiatry, Weill Cornell Medicine, New York, NY, United States

In the setting of underdiagnosed and undertreated perinatal depression (PD), Artificial intelligence (AI) solutions are poised to help predict and treat PD. In the near future, perinatal patients may interact with AI during clinical decision-making, in their patient portals, or through AI-powered chatbots delivering psychotherapy. The increase in potential AI applications has led to discussions regarding responsible AI and explainable AI (XAI). Current discussions of RAI, however, are limited in their consideration of the patient as an active participant with AI. Therefore, we propose a patient-centered, rather than a patient-adjacent, approach to RAI and XAI, that identifies autonomy, beneficence, justice, trust, privacy, and transparency as core concepts to uphold for health professionals and patients. We present empirical evidence that these principles are strongly valued by patients. We further suggest possible design solutions that uphold these principles and acknowledge the pressing need for further research about practical applications to uphold these principles.

1 Introduction

An estimated one in five women in the United States experience depression between conception and 6 months postpartum (1). For this reason, the United States Preventive Services Task Force recommends routine perinatal depression (PD) screening (2). Other studies have estimated perinatal anxiety may be more prevalent, but there are not yet widely used, perinatal-specific screening tools such as the Edinburgh Postnatal Depression Scale (3). Clinician time constraints, resource availability, validity of screening tools across the perinatal period, and patient willingness to report mental health symptoms present challenges to identification (4, 5). Many also face barriers to treatment including limited access to perinatal psychiatrists and financial constraints (6, 7). As a result, therapeutic interventions are initiated late after symptoms develop (e.g., secondary prevention) (5), and as many as 75% of women who experience symptoms of perinatal mood and anxiety disorders (PMADs) go untreated altogether (8). Left untreated, PMADs can lead to poor maternal and infant outcomes, including pre-eclampsia (9), prematurity (10), low birth weight (11, 12), and behavioral dysregulation (1315), lower vagal tone, decreased immunity, and adverse neurodevelopmental outcomes in the child (16). Suicide and overdoses are leading causes death in the first year postpartum (1719).

In the setting of underdiagnosed and undertreated PMADs, novel strategies are needed. Artificial intelligence (AI) is poised to help address these barriers to PMADs diagnosis and treatment. AI involves the development of computer systems that are able to perform tasks that normally require human intelligence such as classifying and predicting disease, recommending treatments, and producing interactive chatbots that leverage large language models to generate humanlike responses (20). For example, AI models can accurately identify women with an elevated risk for PD, enabling primary prevention, before the onset of severe symptoms (46). Primary prevention is significantly more effective than secondary prevention for at-risk women (21). Numerous studies have reported on predictive models that can predict PD with high accuracy using a range of datasets (2224). Specifically, researchers have used electronic health records (EHRs) (25), administrative claims data (26), actively and passively collected data from smartphones (27), social media data (27), and patient-reported outcomes (27, 28), to predict PD, and, in one study, postpartum psychiatric admission. AI-powered chatbots have also become increasingly researched in mental health generally for their potential to deliver psychotherapy (29, 30), and may soon be tailored to the unique needs of perinatal individuals (31, 32).

Artificial intelligence for perinatal psychiatry is now in the early, proof-of-concept stages, but researchers, health systems, and companies are planning for implementation (32). The American College of Obstetrics and Gynecology (ACOG) has also issued guidelines highlighting the potential for AI and other technologies to enhance pregnancy-related care (33). Now is the time to ensure that patient perspectives are understood and integrated into care delivery that uses AI; failure to do so may undermine the very trust in clinicians that is so fundamental in mental healthcare. Research on the bioethical implications of AI in healthcare has raised many important patient safety and quality considerations (34). These conversations have generated popular terms such as responsible AI (RAI), which calls for AI that is ethical, transparent, and accountable while also aligning with stakeholder expectations (35, 36). Similarly, there has been significant interest in explainable AI (XAI), which is a set of techniques ensuring that a person can understand how or why an AI system came up with a certain output (37). XAI can be considered a subdomain of RAI because transparency and explainability help identify biases and support autonomous decision-making.

However, RAI and XAI work has primarily assumed that AI was patient-adjacent: that patients are recipients of AI-propagated benefits or harms, but that they do not interact directly with it. Rather, we argue that there is a need to equally consider patient-centered AI, in which the patient interacts directly with AI during the course of their care. There are many envisioned use cases of patient-centered AI in perinatal psychiatry. Patient engagement and shared decision-making continue to rise in importance in healthcare, making it likely that AI-generated risk predictions will be discussed or presented to patients in clinical care as part of the decision-making process (38). Additionally, new United States policies prohibiting “information blocking” have also led to a deluge of information being returned to patients through patient portals (39), and this information may soon include AI-generated risk predictions that are integrated into EHRs. Finally, AI-powered chatbots delivering psychotherapy represent another example of patients directly interacting with AI (40). Considering patient-centered AI calls for understanding how to responsibly implement AI in a manner that further fosters, rather than undermines, trust. Therefore, in this perspective, we will describe RAI and XAI research, relate it to patient-centered AI, articulate important bioethical concepts in patient-centered RAI and XAI, and describe why these concepts are of particular importance for perinatal psychiatry. We argue that patient-centered AI must consider bioethical constructs while also factoring in important elements of context including type of data being used to develop AI, persons with whom AI or data are shared, what tasks AI supports, and intended time period of use (i.e., preconception, pregnancy, or postpartum).

2 The current landscape of RAI and XAI

Many ethical concerns have been raised about the uses of AI in healthcare. From a legal and regulatory standpoint, the European Union (EU) has issued seven key requirements for trustworthy AI, which relate to human agency and oversight, technical safety and rigor, privacy, transparency, fairness, societal and environmental well-being, and algorithm accountability (21). The EU also has strict regulations, the General Data Protection Regulation (GDPR), regarding data sharing and secondary uses of personal data, which relate to AI algorithm development (41). Such guidance and regulations are lacking, however, in the United States and many other nations worldwide, though the White House has proposed the AI Bill of Rights, containing many elements similar to the GDPR (42).

Responsible AI encompasses the conceptual and empirical work investigating responsible uses of AI (36). Four key ethical principles of beneficence, non-maleficence, autonomy, and justice have been adapted to the AI use case, with new acknowledgments of the interplay between these principles and the external agencies, patients, and the complex technical and clinical environments with which AI is used (43). Research has shown that deliberately communicating information about plans to enact these principles when implementing AI in healthcare is associated with favorable attitudes, satisfaction, and high usage intentions among clinicians (36). Some have proposed practical guidelines to operationalize these principles when AI is implemented in clinical practice, including checklists, governance processes, and training for clinicians (44). We have identified that communicating the purpose (why the AI was developed and what it does), process (how it operates), and performance (how well it functions) of an AI system helps users form greater trust (45). In general, however, the majority of RAI work has been highly theoretical, and more practical guidance is needed (46). This research has also primarily considered clinicians to be the end users, while patients are passive participants in the AI–they are considered recipients of the benefits and harms of AI, but not agents interacting directly with it (4749).

A fifth bioethical principle, explainability, has been proposed as relating to, and upholding, the other four principles previously mentioned (36, 50). Explainability, or XAI, is the idea that humans should have insight into how an AI system works. In the literature, it is an often-used yet poorly defined term. Some have defined it as broadly as technological transparency in general, while others argue it has specific requirements including the ability of a human operator to reconstruct how an AI system uses data, moves through a decision-making process, and arrives at a conclusion (37, 51). XAI has been described as a necessary component of RAI because without explainability, it is difficult for clinicians to exercise their independent clinical judgment (i.e., act autonomously), or for them to ensure that AI is helping, not harming, all of their patients–not just a select few (i.e., upholding beneficence, non-maleficence, and justice) (36). XAI is necessary because of the “black box” problem; for many types of models, particularly those that use deep learning (a type of machine learning), the internal workings of the algorithm are unknown to the developers or end users. Explainability measures attempt to “shine a light” into the black box by using additional algorithms to understand how and why an AI model is producing certain output (52, 53). For example, XAI may tell a clinician which variables included in a model are most contributing to a specific patient’s risk of developing PMADs. To further enhance clinical utility, researchers have also explored visualizations of these explainability measures (54). However, as with RAI research in general, nearly all of the research into XAI techniques considers clinicians to be the primary end users (38). Therefore, the unique information needs and bioethical concerns of patients who may be directly interacting with AI in the course of their care have not been fully considered.

Concepts of RAI and XAI are also complex as perspectives may vary based on context. For example, a recent survey of 610 United States-based adults (sampled to reflect racial, sex, and age demographics of the United States) found that willingness to share health information varied by the type of data (sexual health information, imaging data, genetic data, or mental health information) and persons with whom the information would be shared (e.g., health technology companies, doctors and nurses, and chosen friends and family members). Of note for perinatal psychiatry, participants were least comfortable sharing sexual and mental health information, respectively (55).

3 RAI and XAI matter to perinatal psychiatry patients

Trust in AI is one of the biggest challenges in clinical practice and one of the goals of AI implementation has been to foster appropriate trust. Much of the AI implementation research to date has focused on presenting model output and fostering trust in AI among clinicians (45). Maintaining the trust of patients who may be directly interacting with AI is of equal importance, as trust is a fundamental requirement of patient-centered perinatal psychiatry.

Perinatal psychiatry presents a complex bioethical case because of the sensitivity of both pregnancy and mental health data, and the multiple layers of autonomy. In pregnancy, the autonomy, harms, and benefits afforded to the pregnant person, newborn or fetus, and partner must be weighed simultaneously (56, 57). Mental health issues in the perinatal person can negatively impact the newborn or fetus, and partners play an important supportive role in caring for both the perinatal patient experiencing PMADs and the newborn. There is also a layered network of care providers that may at different points in time include, obstetrics, midwifery, doulas, mental health professionals, primary care providers, and pediatricians. Although, the perinatal person may have different levels of comfort disclosing PMADS symptoms to different roles, for example their own obstetrician vs. their child’s pediatrician. Furthermore, perinatal persons may be reticent to accept AI solutions that can predict or treat PMADs. Some concerns relate to the way their personal health data is being used. Numerous research studies have noted that integrating multiple datasets including EHRs, claims data, patient-reported outcomes, and data collected from smartphones and social media improves prediction accuracy (2628). Perinatal people may be alarmed to discover their EHRs are being repurposed and used for AI-based risk prediction, and may first learn about it when encountering an AI-based risk prediction with their clinicians. Similarly, perinatal persons may have unique concerns about wearable data being hacked or inappropriately shared (e.g., disclosing their location and movements in ways that may be dangerous). For example, many have shed light on the privacy violations and monetization of data collected from period tracking applications (58, 59). These issues are often compounded by the broader concerns about revealing a mental health diagnosis during the perinatal period, such as fear of stigma or loss of custody of one’s baby (22, 60). Moreover, ethical considerations and factors needed to build trust differ by demographic variables, such as race and/or ethnicity, due to well-documented issues of algorithmic bias (61). Therefore, it is imperative that approaches to patient-centered AI for perinatal applications inclusively represent different patient groups, taking special care to ensure minoritized groups may equitably reap benefits and not be disproportionately exposed to harms of AI.

Although in its nascent stages, research on perinatal patients’ attitudes toward AI confirms their strong desire to be involved and informed about how AI is being used in their care. For example, one survey examined the perspectives of 258 English-speaking United States persons registered with an online survey-sampling platform regarding their attitudes toward AI use in mental healthcare. The sample was reflective of the United States population based on race. While the study involved persons with inadequate health literacy (24%), those with less than a Bachelor’s degree (47%), and persons reporting Hispanic/Latino ethnicity (6.5%), these numbers are generally lower than national averages. The authors found most participants reported it was “very important” for them to understand AI output, that participants with a history of pregnancy were significantly more likely to feel this transparency was important, many participants were concerned about medical harm resulting from AI, inappropriate data sharing, and that AI may lead to their mental health provider not knowing them as well (62). Similarly, another study surveyed 150 pregnant persons in Spain from a single hospital. The respondents had somewhat higher rates of tertiary education than the Spanish public (41% in the study vs. 33% nationally). They found that participants strongly endorsed the need for AI to be responsible, trustworthy, useful, and safe, many have privacy concerns, and, importantly, XAI would increase the trust and confidence of participants who were averse to AI being used in their care (63). Importantly, in both studies, participants reported generally high levels of openness to AI being used in one’s healthcare, but all wanted their clinical care team to play a role in ensuring the safe, responsible use of AI (62, 63). These issues should be further evaluated in a sample that is better representative of those that have lower literacy, less education, and a greater proportion of those with Hispanic/Latino ethnicity.

The United States-based study described also, found differences in participants’ comfort with AI based on the task the AI performed. Participants were least comfortable with diagnosis delivery tasks (i.e., AI telling someone they have a mental health condition) and recommending medication. Participants were relatively more comfortable with non-pharmacological treatment recommendations and AI performing mental health assessments (62).

4 How can we uphold patient-centered RAI and XAI in perinatal psychiatry?

A practical understanding of patient-centered RAI in perinatal psychiatry is critically needed to support patients in seeking care, receiving care, and maintaining a positive therapeutic alliance with mental health professionals when AI may be involved. As a first step, it is important to identify the bioethical issues of importance to patients that may be operationalized in practice in the future. Previously, our team has synthesized research on RAI generally (64), as well as the ethics of using consumer-generated data (65), AI in psychiatry (20), and maternal health (66, 67) to identify core ethical concepts to be considered in the perinatal psychiatry space.

We propose six bioethical constructs for upholding patient-centered RAI, including autonomy, privacy, beneficence, justice, transparency, and trust (also shown in Figure 1). Autonomy calls for patient self-determination regarding how their data are used by those developing AI, and provide informed consent in ways that are specific to the data type, recipient, and use case. We assert that this autonomy lies with the birthing person involving informal support (e.g., partners, family members) as desired, or in rare circumstance where medical proxies may be required. Relatedly, privacy calls for the appropriate and confidential uses of personal health data to train AI, and for data use to be deferential to patient wishes as described in a consent process. Communication of privacy-related practices is even more crucial from a reproductive health perspective given the sensitivity of the data that may be involved. When data involve the infant, privacy from the perspective of the non-birthing partner may also become a consideration. Beneficence calls for AI to demonstrably improve patient outcomes and, as an implied partner to this, that risks and harms from AI are minimized (i.e., non-maleficence). In perinatal settings, beneficence must not only consider the birthing person but also ensuring potential harms to the child are minimized while prioritizing the autonomy and wellbeing of the birthing person. Justice calls for the equal and fair access of the benefits of AI to be evenly distributed among all patients. As part of this, it is important to consider the diversity of patients affected by AI, which includes considering potential sources of algorithmic bias as well as differing information needs and preferences based on different levels of health literacy, numeracy, and experiences of implicit bias in healthcare interactions. Given United States-based disparities in perinatal health outcomes, the disproportionate share of adverse perinatal outcomes occurring in low-and-middle-income countries, and the lack of mental health support for minoritized groups, justice holds particular importance in perinatal psychiatry. Transparency calls for XAI that considers the patient as the end user, and ensures that patients can understand how AI is used in their care, how their personal health data was used for that AI, and, crucially, how AI arrived at a specific conclusion relating to their health or healthcare. This principle may be the most different from the general discussions of XAI and RAI as communicating AI output to patients has not previously been explored. Finally, underlying each of these principles is Trust—the need to ensure that trust between patients, clinicians, and AI is not undermined. Because AI research both within and beyond perinatal psychiatry has demonstrated that clinicians’ perceptions of AI play a significant role in patients’ attitudes toward it (62, 63, 68, 69), we must consider the interplay of trust between patient, clinician, and AI.

Figure 1
www.frontiersin.org

Figure 1. Temporal, person, data, and task-based elements that must be considered in the context of bioethical principles to support RAI and XAI for perinatal psychiatry. * Patient reported outcomes.

Our empirical research with 258 pregnancy-capable (female sex at birth) United States-based survey respondents (also described above) showed that these concepts resonated with participant groups relevant to perinatal psychiatry. The majority of participants surveyed about the potential use case of AI for mental health endorsed that it was “very important” that they could understand which of their individual risk factors for depression are used by AI (85%), AI will decrease the chance of negative outcomes (77%), they were aware of how their personal data was being used for AI (75%), they were able to make up their own mind about their risk for depression based on AI output (71%), they can understand how likely it is that they develop depression within the next year according to AI (71%), and AI will improve depression and/or depressive symptoms (62%) (62).

The endorsed importance of the six bioethical constructs among survey participants suggests they may serve as a future research agenda for those studying the ethical implications of patients interacting with AI in mental healthcare. However, given the sensitivities and complexities related to perinatal psychiatry, there are also data, person, task, and temporal factors that must be considered. Figure 1 outlines important elements to consider as it pertains to each of these factors in the context of perinatal psychiatry. First, patients may have different needs or perspectives, or there may be different persons to involve based on temporal factors, specifically whether they are preconception, currently pregnant, or postpartum. This also considers that patient perspectives and needs are fluid within and outside of these time frames. Second, there are numerous persons who may influence data used for AI or may have access to AI-related information, including both informal support persons and health professionals. Which persons are involved and patient perspectives regarding their involvement will likely be individually dependent. Third, AI may be informed using various different kinds of patient and health-system generated data elements, some of which have been described. While these elements may not differ much for perinatal persons, the information may have increased sensitivity which should be considered, for example, app-related data that track ovulation or pregnancy, or mental health reported outcomes. Fourth, AI may be used to support diagnosis/classification, treatment recommendation, or treatment support related tasks. Needs and opinions on diagnosis and classification tasks may vary based on perceived severity of disease (e.g., psychosis vs. moderate depression) or other factors. Based on our previous work, patient comfort with treatment recommendation tasks may vary based on whether or not the treatment involves pharmacotherapy, and these perceptions may also differ throughout the preconceptions, pregnancy, and postpartum life-cycle. Fourth, treatment support may come in many different forms with varied levels of autonomy, such as using patient-reported outcomes (PROs) to support health-professional led care to having a nearly-autonomous chatbot leading therapy. Last, these factors must continue to be weighed with the foundation of bioethical principles described above. There are likely more questions than answers related to AI use for perinatal psychiatry in these nascent stages, but we advocate that a research agenda should consider these bioethical constructs and contextual factors in order to ensure AI in perinatal psychiatry may be safe, inclusive, and patient-centered.

It may be possible to design patient education materials, or features in the AI-integrated technology itself, to uphold bioethical principles in practice while still considering important contextual factors. For example, computational interpretability methods that explain important predictors in AI-generated risk prediction [e.g., SHapley Additive exPlanations, or SHAP (70)] may improve patient and clinician trust in the AI and foster its appropriate, safe use. Because visualizations objectively improve comprehension (71, 72), these methods should include visualizations that are comprehended by patients, not just clinicians. Additionally, patient educational materials or “InfoButtons” embedded directly in AI-enhanced technologies can explain an AI’s purpose, its use of personal health data, and its performance, which may support autonomy and transparency. Communication of this information should also follow inclusive design principles to uphold distributive justice (73). Based on described shortcomings in the samples included in previous relevant research (62, 63), it is also important that further empirical studies be conducted with samples that reflect the ethnicity, literacy, and educational backgrounds of diverse perinatal populations. This may be accomplished scientifically by instituting sampling quotas or over-sampling typically under-represented groups. Practically, it will be important to work with community-based organizations to support: recruitment, use of appropriate language, and fostering trust with groups that have a history of facing discrimination in a healthcare setting.

In the near future, perinatal patients may interact with AI during clinical decision-making (38), in their patient portals (39), or through AI-powered chatbots delivering psychotherapy (40). Current discussions of RAI are limited in their consideration of the patient as an active participant with AI. Therefore, we propose a patient-centered, rather than a patient-adjacent, approach to RAI and XAI, that identifies autonomy, beneficence, justice, trust, privacy, and transparency as core concepts to uphold. Although we suggest possible design solutions, research about practical applications to uphold these principles is needed as a next step.

Data availability statement

The original contributions presented in the study are included in the article/supplementary material, further inquiries can be directed to the corresponding author.

Author contributions

MT: Conceptualization, Writing – original draft. AH: Writing – review & editing. NB: Conceptualization, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported by the National Institute of Mental Health (R41MH124581-S1; PI—Pathak). MT is supported by the National Institute of Nursing Research (R00NR019124; PI—MT), and NB is supported by the National Institute on Minority Health and Health Disparities (R00MD015781; PI—NB).

Conflict of interest

MT and AH are co-founders and have equity in Iris OB Health, New York.

The remaining author declares that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Gavin, NI, Gaynes, BN, Lohr, KN, Meltzer-Brody, S, Gartlehner, G, and Swinson, T. Perinatal depression: a systematic review of prevalence and incidence. Obstet Gynecol. (2005) 106:1071–83. doi: 10.1097/01.AOG.0000183597.31630.db

Crossref Full Text | Google Scholar

2. Curry, SJ, Krist, AH, Owens, DK, Barry, MJ, Caughey, AB, Davidson, KW, et al. Interventions to prevent perinatal depression: US preventive services task force recommendation statement. JAMA. (2019) 321:580–7. doi: 10.1001/jama.2019.0007

Crossref Full Text | Google Scholar

3. Leach, LS, Poyser, C, and Fairweather-schmidt, K. Maternal perinatal anxiety: a review of prevalence and correlates. Clin Psychol. (2017) 21:4–19. doi: 10.1111/cp.12058

Crossref Full Text | Google Scholar

4. O’Connor, E, Senger, CA, Henninger, ML, Coppola, E, and Gaynes, BN. Interventions to prevent perinatal depression: evidence report and systematic review for the US preventive services task force. JAMA. (2019) 321:588–601. doi: 10.1001/jama.2018.20865

Crossref Full Text | Google Scholar

5. Yeaton-Massey, A, and Herrero, T. Recognizing maternal mental health disorders: beyond postpartum depression. Curr Opin Obstet Gynecol. (2019) 31:116–9. doi: 10.1097/GCO.0000000000000524

PubMed Abstract | Crossref Full Text | Google Scholar

6. Dennis, CL, and Chung-Lee, L. Postpartum depression help-seeking barriers and maternal treatment preferences: a qualitative systematic review. Birth. (2006) 33:323–31. doi: 10.1111/j.1523-536X.2006.00130.x

PubMed Abstract | Crossref Full Text | Google Scholar

7. Sobey, WS. Barriers to postpartum depression prevention and treatment: a policy analysis. J Midwifery Womens Health. (2002) 47:331–6. doi: 10.1016/S1526-9523(02)00277-5

Crossref Full Text | Google Scholar

8. Byatt, N, Levin, LL, Ziedonis, D, Moore Simas, TA, and Allison, J. Enhancing participation in depression Care in Outpatient Perinatal Care Settings: a systematic review. Obstet Gynecol. (2015) 126:1048–58. doi: 10.1097/AOG.0000000000001067

PubMed Abstract | Crossref Full Text | Google Scholar

9. Kurki, T, Hiilesmaa, V, Raitasalo, R, Mattila, H, and Ylikorkala, O. Depression and anxiety in early pregnancy and risk for preeclampsia. Obstet Gynecol. (2000) 95:487–90.

PubMed Abstract | Google Scholar

10. Grote, NK, Bridge, JA, Gavin, AR, Melville, JL, Iyengar, S, and Katon, WJ. A meta-analysis of depression during pregnancy and the risk of preterm birth, low birth weight, and intrauterine growth restriction. Arch Gen Psychiatry. (2010) 67:1012–24. doi: 10.1001/archgenpsychiatry.2010.111

PubMed Abstract | Crossref Full Text | Google Scholar

11. Dole, N, Savitz, DA, Hertz-Picciotto, I, Siega-Riz, AM, McMahon, MJ, and Buekens, P. Maternal stress and preterm birth. Am J Epidemiol. (2003) 157:14–24. doi: 10.1093/aje/kwf176

Crossref Full Text | Google Scholar

12. Rondó, PHC, Ferreira, RF, Nogueira, F, Ribeiro, MCN, Lobert, H, and Artes, R. Maternal psychological stress and distress as predictors of low birth weight, prematurity and intrauterine growth retardation. Eur J Clin Nutr. (2003) 57:266–72. doi: 10.1038/sj.ejcn.1601526

PubMed Abstract | Crossref Full Text | Google Scholar

13. Groome, LJ, Swiber, MJ, Bentz, LS, Holland, SB, and Atterbury, JL. Maternal anxiety during pregnancy: effect on fetal behavior at 38 to 40 weeks of gestation. J Dev Behav Pediatr. (1995) 16:391–6. doi: 10.1097/00004703-199512000-00001

PubMed Abstract | Crossref Full Text | Google Scholar

14. Susman, EJ, Schmeelk, KH, Ponirakis, A, and Gariepy, JL. Maternal prenatal, postpartum, and concurrent stressors and temperament in 3-year-olds: a person and variable analysis. Dev Psychopathol. (2001) 13:629–52. doi: 10.1017/S0954579401003121

PubMed Abstract | Crossref Full Text | Google Scholar

15. O’Connor, TG, Heron, J, Golding, J, Beveridge, M, and Glover, V. Maternal antenatal anxiety and children’s behavioural/emotional problems at 4 years. Report from the Avon longitudinal study of parents and children. Br J Psychiatry. (2002) 180:502–8. doi: 10.1192/bjp.180.6.502

Crossref Full Text | Google Scholar

16. Field, T. Prenatal anxiety effects: a review. Infant Behav Dev. (2017) 49:120–8. doi: 10.1016/j.infbeh.2017.08.008

Crossref Full Text | Google Scholar

17. Metz, TD, Rovner, P, Hoffman, MC, Allshouse, AA, Beckwith, KM, and Binswanger, IA. Maternal deaths from suicide and overdose in Colorado, 2004-2012. Obstet Gynecol. (2016) 128:1233–40. doi: 10.1097/AOG.0000000000001695

PubMed Abstract | Crossref Full Text | Google Scholar

18. Goldman-Mellor, S, and Margerison, CE. Maternal drug-related death and suicide are leading causes of postpartum death in California. Am J Obstet Gynecol. (2019) 221:e489.e1–9. doi: 10.1016/j.ajog.2019.05.045

PubMed Abstract | Crossref Full Text | Google Scholar

19. Centers for Disease Control and Prevention (2023). Pregnancy-related deaths: Data from 14 US maternal mortality review committees. Available at: https://www.cdc.gov/reproductivehealth/maternal-mortality/docs/pdf/MMR-Data-Brief_2019-h.pdf (Accessed October 11, 2023).

Google Scholar

20. Doraiswamy, PM, Blease, C, and Bodner, K. Artificial intelligence and the future of psychiatry: insights from a global physician survey. Artif Intell Med. (2020) 102:101753. doi: 10.1016/j.artmed.2019.101753

PubMed Abstract | Crossref Full Text | Google Scholar

21. Cannarsa, M. “Ethics guidelines for trustworthy AI,” in The Cambridge Handbook of Lawyering in the Digital Age. Eds. LA DiMatteo, A Janssen, P Ortolani, F de Elizalde, M Cannarsa, and M Durovic Cambridge University Press (2021). 283–97.

Google Scholar

22. Saqib, K, Khan, AF, and Butt, ZA. Machine learning methods for predicting postpartum depression: scoping review. JMIR Ment Health. (2021) 8:e29838. doi: 10.2196/29838

PubMed Abstract | Crossref Full Text | Google Scholar

23. Cellini, P, Pigoni, A, Delvecchio, G, Moltrasio, C, and Brambilla, P. Machine learning in the prediction of postpartum depression: a review. J Affect Disord. (2022) 309:350–7. doi: 10.1016/j.jad.2022.04.093

Crossref Full Text | Google Scholar

24. Davidson, L, and Boland, MR. Towards deep phenotyping pregnancy: a systematic review on artificial intelligence and machine learning methods to improve pregnancy outcomes. Brief Bioinform. (2021) 22:bbaa369. doi: 10.1093/bib/bbaa369

Crossref Full Text | Google Scholar

25. Zhang, Y, Wang, S, Hermann, A, Joly, R, and Pathak, J. Development and validation of a machine learning algorithm for predicting the risk of postpartum depression among pregnant women. J Affect Disord. (2021) 279:1–8. doi: 10.1016/j.jad.2020.09.113

PubMed Abstract | Crossref Full Text | Google Scholar

26. Betts, KS, Kisely, S, and Alati, R. Predicting postpartum psychiatric admission using a machine learning approach. J Psychiatr Res. (2020) 130:35–40. doi: 10.1016/j.jpsychires.2020.07.002

PubMed Abstract | Crossref Full Text | Google Scholar

27. Bilal, AM, Fransson, E, Bränn, E, Eriksson, A, Zhong, M, Gidén, K, et al. Predicting perinatal health outcomes using smartphone-based digital phenotyping and machine learning in a prospective Swedish cohort (Mom2B): study protocol. BMJ Open. (2022) 12:e059033. doi: 10.1136/bmjopen-2021-059033

PubMed Abstract | Crossref Full Text | Google Scholar

28. Garbazza, C, Mangili, F, D'Onofrio, TA, Malpetti, D, D’Agostino, A, Cicolin, A, et al. 0621 a machine learning model to predict the risk of perinatal depression from sleep data in healthy pregnant women. Sleep. (2023) 46:A273. doi: 10.1093/sleep/zsad077.0621

Crossref Full Text | Google Scholar

29. Fitzpatrick, KK, Darcy, A, and Vierhile, M. Delivering cognitive behavior therapy to young adults with symptoms of depression and anxiety using a fully automated conversational agent (Woebot): a randomized controlled trial. JMIR Mental Health. (2017) 4:e7785. doi: 10.2196/mental.7785

Crossref Full Text | Google Scholar

30. Williams, AD, and Andrews, G. The effectiveness of internet cognitive behavioural therapy (iCBT) for depression in primary care: a quality assurance study. PLoS One. (2013) 8:e57447. doi: 10.1371/journal.pone.0057447

PubMed Abstract | Crossref Full Text | Google Scholar

31. Chua, JYX, Choolani, M, Chee, CYI, Chan, YH, Lalor, JG, Chong, YS, et al. Insights of parents and parents-to-be in using chatbots to improve their preconception, pregnancy, and postpartum health: a mixed studies review. J Midwifery Womens Health. (2023) 68:480–9. doi: 10.1111/jmwh.13472

PubMed Abstract | Crossref Full Text | Google Scholar

32. Novick, AM, Kwitowski, M, Dempsey, J, Cooke, DL, and Dempsey, AG. Technology-based approaches for supporting perinatal mental health. Curr Psychiatry Rep. (2022) 24:419–29. doi: 10.1007/s11920-022-01349-w

PubMed Abstract | Crossref Full Text | Google Scholar

33. Committee on Obstetric Practice A. Presidential task force on redefining the postpartum visit committee on obstetric practice. Obstet Gynecol. (2021).

Google Scholar

34. Anderson, M, and Anderson, SL. How should AI be developed, validated, and implemented in patient care? AMA J Ethics. (2019) 21:E125–30. doi: 10.1001/amajethics.2019.125

Crossref Full Text | Google Scholar

35. Sivarajah, U, Wang, Y, Olya, H, and Mathew, S. Responsible artificial intelligence (AI) for digital health and medical analytics. Inf Syst Front. (2023) 25:2117–22. doi: 10.1007/s10796-023-10412-7

PubMed Abstract | Crossref Full Text | Google Scholar

36. Wang, W, Chen, L, Xiong, M, and Wang, Y. Accelerating AI adoption with responsible AI signals and employee engagement mechanisms in health care. Inf Syst Front. (2021) 25:2239–56. doi: 10.1007/s10796-021-10154-4

Crossref Full Text | Google Scholar

37. Amann, J, Blasimme, A, Vayena, E, Frey, D, and Madai, VIPrecise4Q consortium. Explainability for artificial intelligence in healthcare: a multidisciplinary perspective. BMC Med Inform Decis Mak. (2020) 20:310. doi: 10.1186/s12911-020-01332-6

PubMed Abstract | Crossref Full Text | Google Scholar

38. Walsh, CG, McKillop, MM, Lee, P, Harris, JW, Simpson, C, and Novak, LL. Risky business: a scoping review for communicating results of predictive models between providers and patients. Jamia Open. (2021) 4:ooab092. doi: 10.1093/jamiaopen/ooab092

PubMed Abstract | Crossref Full Text | Google Scholar

39. Bonamici 21st century cures act. (2016). Available at: http://www.congress.gov/bill/114th-congress/house-bill/34 (Accessed August 9, 2023).

Google Scholar

40. Abd-Alrazaq, AA, Rababeh, A, Alajlani, M, Bewick, BM, and Househ, M. Effectiveness and safety of using Chatbots to improve mental health: systematic review and meta-analysis. J Med Internet Res. (2020) 22:e16021. doi: 10.2196/16021

PubMed Abstract | Crossref Full Text | Google Scholar

41. General Data Protection Regulation (GDPR) (2023). Official Legal Text. General data protection regulation (GDPR). Available at: https://gdpr-info.eu/ (Accessed September 25, 2023).

Google Scholar

42. Blumenthal-Barby, J, and An, AI. Bill of rights: implications for health care AI and machine learning—a bioethics Lens. Am J Bioeth. (2023) 23:4–6. doi: 10.1080/15265161.2022.2135875

PubMed Abstract | Crossref Full Text | Google Scholar

43. Floridi, L, Cowls, J, Beltrametti, M, and Chatila, R (2018). AI4People—An ethical framework for a good AI society: Opportunities, risks, principles, and recommendations, Atomium. European Institute.

Google Scholar

44. Benjamins, R, Barbado, A, and Sierra, D (2019). Responsible AI by Design in Practice. arXiv [Preprint]. Available at: http://arxiv.org/abs/1909.12838 (Accessed September 27, 2019).

Google Scholar

45. Benda, NC, Novak, LL, Reale, C, and Ancker, JS. Trust in AI: why we should be designing for APPROPRIATE reliance. J Am Med Inform Assoc. (2021) 29:207–12. doi: 10.1093/jamia/ocab238

PubMed Abstract | Crossref Full Text | Google Scholar

46. Morley, J, Floridi, L, Kinsey, L, and Elhalal, A. From what to how: An initial review of publicly available AI ethics tools, methods and research to translate principles into practices. Sci Eng Ethics. (2020) 26:2141–68. doi: 10.1007/s11948-019-00165-5

PubMed Abstract | Crossref Full Text | Google Scholar

47. Vandemeulebroucke, T, Denier, Y, Mertens, E, and Gastmans, C. Which framework to use? A systematic review of ethical frameworks for the screening or evaluation of health technology innovations. Sci Eng Ethics. (2022) 28:26. doi: 10.1007/s11948-022-00377-2

Crossref Full Text | Google Scholar

48. Tang, L, Li, J, and Fantus, S. Medical artificial intelligence ethics: a systematic review of empirical studies. Digit Health. (2023) 9:20552076231186064. doi: 10.1177/20552076231186064

PubMed Abstract | Crossref Full Text | Google Scholar

49. Siala, H, and Wang, Y. SHIFTing artificial intelligence to be responsible in healthcare: a systematic review. Soc Sci Med. (2022) 296:114782. doi: 10.1016/j.socscimed.2022.114782

PubMed Abstract | Crossref Full Text | Google Scholar

50. Floridi, L, and Cowls, J. A unified framework of five principles forAIin society. Mach Learn City. (2022) 21:535–45. doi: 10.1002/9781119815075.ch45

Crossref Full Text | Google Scholar

51. Markus, AF, Kors, JA, and Rijnbeek, PR. The role of explainability in creating trustworthy artificial intelligence for health care: a comprehensive survey of the terminology, design choices, and evaluation strategies. J Biomed Inform. (2021) 113:103655. doi: 10.1016/j.jbi.2020.103655

PubMed Abstract | Crossref Full Text | Google Scholar

52. Chaddad, A, Peng, J, Xu, J, and Bouridane, A. Survey of explainable AI techniques in healthcare. Sensors. (2023) 23:634. doi: 10.3390/s23020634

PubMed Abstract | Crossref Full Text | Google Scholar

53. Allgaier, J, Mulansky, L, Draelos, RL, and Pryss, R. How does the model make predictions? A systematic literature review on the explainability power of machine learning in healthcare. Artif Intell Med. (2023) 143:102616. doi: 10.1016/j.artmed.2023.102616

PubMed Abstract | Crossref Full Text | Google Scholar

54. Diprose, WK, Buist, N, Hua, N, Thurier, Q, Shand, G, and Robinson, R. Physician understanding, explainability, and trust in a hypothetical machine learning risk calculator. J Am Med Inform Assoc. (2020) 27:592–600. doi: 10.1093/jamia/ocz229

PubMed Abstract | Crossref Full Text | Google Scholar

55. Niño de Rivera, S, Masterson Creber, R, Zhao, Y, Mangal, S, Dugdale, LS, and Reading Turchioe, M (n.d.). Research participants’ perspectives on the 2023 National Institutes of Health data management and sharing policy and implications for willingness to participate in research. (2024)

Google Scholar

56. Gehtland, LM, Paquin, RS, Andrews, SM, Lee, AM, Gwaltney, A, Duparc, M, et al. Using a patient portal to increase enrollment in a newborn screening research study: observational study. JMIR Pediatr Parent. (2022) 5:e30941. doi: 10.2196/30941

PubMed Abstract | Crossref Full Text | Google Scholar

57. Tran, VT, Riveros, C, and Ravaud, P. Patients’ views of wearable devices and AI in healthcare: findings from the ComPaRe e-cohort. NPJ Digit Med. (2019) 2:1–8. doi: 10.1038/s41746-019-0132-y

Crossref Full Text | Google Scholar

58. Gross, MS, Hood, A, and Corbin, B. Pay no attention to that man behind the curtain: An ethical analysis of the monetization of menstruation app data. Int J Fem Approaches Bioeth. (2021) 14:144–56. doi: 10.3138/ijfab-2021-03-22

Crossref Full Text | Google Scholar

59. Shipp, L, and Blasco, J. How private is your period?: a systematic analysis of menstrual app privacy policies. Proc Priv Enhanc Technol. (2020) 2020:491–510. doi: 10.2478/popets-2020-0083

Crossref Full Text | Google Scholar

60. Javed, F, Gilani, SO, Latif, S, Waris, A, Jamil, M, and Waqas, A. Predicting risk of antenatal depression and anxiety using multi-layer perceptrons and support vector machines. J Pers Med. (2021) 11:199–216. doi: 10.3390/jpm11030199

PubMed Abstract | Crossref Full Text | Google Scholar

61. Ledford, H. Millions of black people affected by racial bias in health-care algorithms. Nat Publish Group. (2019) 574:608–9. doi: 10.1038/d41586-019-03228-6

PubMed Abstract | Crossref Full Text | Google Scholar

62. Reading Turchioe, M, Harkins, S, Desai, P, Kumar, S, Kim, J, Hermann, A, et al. Women’s perspectives on the use of artificial intelligence (AI)-based technologies in mental healthcare. JAMIA Open. (2023) 6:ooad048. doi: 10.1093/jamiaopen/ooad048

Crossref Full Text | Google Scholar

63. Oprescu, AM, Miró-Amarante, G, García-Díaz, L, Rey, VE, Chimenea-Toscano, A, Martínez-Martínez, R, et al. Towards a data collection methodology for responsible artificial intelligence in health: a prospective and qualitative study in pregnancy. Inf Fusion. (2022) 83-84:53–78. doi: 10.1016/j.inffus.2022.03.011

Crossref Full Text | Google Scholar

64. The IEEE (2017). Global initiative on ethics of autonomous and intelligent systems. Ethically aligned design: A vision for prioritizing human well-being with autonomous and intelligent systems, version 2. IEEE. Available at: http://standards.ieee.org/develop/indconn/ec/autonomous_systems.html.]

Google Scholar

65. Skopac, EL, Mbawuike, JS, Levin, SU, Dwyer, AT, Rosenthal, SJ, Humphreys, AS, et al. An ethical framework for the use of consumer-generated data in health care. (2019). Available at: https://www.mitre.org/sites/default/files/publications/A%20CGD%20Ethical%20Framework%20in%20Health%20Care%20final.pdf

Google Scholar

66. Brandon, AR, Shivakumar, G, Lee, SC, Inrig, SJ, and Sadler, JZ. Ethical issues in perinatal mental health research. Curr Opin Psychiatry. (2009) 22:601–6. doi: 10.1097/YCO.0b013e3283318e6f

PubMed Abstract | Crossref Full Text | Google Scholar

67. Scott, KA, Britton, L, and McLemore, MR. The ethics of perinatal Care for Black Women: dismantling the structural racism in “mother blame” narratives. J Perinat Neonatal Nurs. (2019) 33:108–15. doi: 10.1097/JPN.0000000000000394

Crossref Full Text | Google Scholar

68. Richardson, JP, Smith, C, Curtis, S, Watson, S, Zhu, X, Barry, B, et al. Patient apprehensions about the use of artificial intelligence in healthcare. NPJ Digit Med. (2021) 4:1–6. doi: 10.1038/s41746-021-00509-1

Crossref Full Text | Google Scholar

69. Khullar, D, Casalino, LP, Qian, Y, Lu, Y, Krumholz, HM, and Aneja, S. Perspectives of patients about artificial intelligence in health care. JAMA Netw Open. (2022) 5:e2210309. doi: 10.1001/jamanetworkopen.2022.10309

PubMed Abstract | Crossref Full Text | Google Scholar

70. Lundberg, SM, and Lee, SI. A unified approach to interpreting model predictions In: I Guyon, UV Luxburg, and S Bengio, et al., editors. Advances in Neural Information Processing Systems 30. Cambridge, MA, USA: Curran Associates, Inc. (2017). 4765–74.

Google Scholar

71. Zikmund-Fisher, BJ, Scherer, AM, Witteman, HO, Solomon, JB, Exe, NL, Tarini, BA, et al. Graphics help patients distinguish between urgent and non-urgent deviations in laboratory test results. J Am Med Inform Assoc. (2017) 24:520–8. doi: 10.1093/jamia/ocw169

PubMed Abstract | Crossref Full Text | Google Scholar

72. Reading Turchioe, M, Grossman, LV, Myers, AC, Baik, D, Goyal, P, and Masterson Creber, RM. Visual analogies, not graphs, increase patients’ comprehension of changes in their health status. J Am Med Inform Assoc. (2020) 27:677–89. doi: 10.1093/jamia/ocz217

PubMed Abstract | Crossref Full Text | Google Scholar

73. Benda, NC, Montague, E, and Valdez, RS. Design for inclusivity In: Design for Health. Eds. A Sethumadhavan and F Sasangohar Elsevier (2020). 305–22.

Google Scholar

Keywords: artificial intelligence (AI), explainable AI, bioethics, perinatal psychiatry, patient centered care

Citation: Turchioe MR, Hermann A and Benda NC (2024) Recentering responsible and explainable artificial intelligence research on patients: implications in perinatal psychiatry. Front. Psychiatry. 14:1321265. doi: 10.3389/fpsyt.2023.1321265

Received: 13 October 2023; Accepted: 27 December 2023;
Published: 18 January 2024.

Edited by:

Maria Muzik, University of Michigan, United States

Reviewed by:

Ellen McGinnis, Wake Forest Baptist Medical Center, United States
Nonyé Kanu, Independent researcher, Dallas, TX, United States

Copyright © 2024 Turchioe, Hermann and Benda. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Natalie C. Benda, nb3115@cumc.columbia.edu

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.