Skip to main content

REVIEW article

Front. Robot. AI
Sec. Biomedical Robotics
Volume 11 - 2024 | doi: 10.3389/frobt.2024.1444763

Enhancing Interpretability and Accuracy of AI Models in Healthcare: A Comprehensive Review on Challenges and Future Directions

Provisionally accepted
Mohammad Ennab Mohammad Ennab *Hamid Mcheick Hamid Mcheick
  • Université du Québec à Chicoutimi, Chicoutimi, Canada

The final, formatted version of the article will be published soon.

    Artificial Intelligence (AI) has demonstrated exceptional performance in automating critical healthcare tasks, such as diagnostic imaging analysis and predictive modeling, often surpassing human capabilities. The integration of AI in healthcare promises substantial improvements in patient outcomes, including faster diagnosis and personalized treatment plans. However, AI models frequently lack interpretability, leading to significant challenges concerning their performance and generalizability across diverse patient populations. These opaque AI technologies raise serious patient safety concerns, as non-interpretable models can result in improper treatment decisions due to misinterpretations by healthcare providers. Our systematic review explores various AI applications in healthcare, focusing on the critical assessment of model interpretability and accuracy. We identify and elucidate the most significant limitations of current AI systems, such as the black-box nature of deep learning models and the variability in performance across different clinical settings. By addressing these challenges, our objective is to provide healthcare providers with well-informed strategies to develop innovative and safe AI solutions. This review aims to ensure that future AI implementations in healthcare not only enhance performance but also maintain transparency and patient safety.

    Keywords: Artificial intelligence (AI), Machine Learning (ML), deep learning (DL), healthcare, Interpretability, Explainability, model accuracy

    Received: 06 Jun 2024; Accepted: 27 Sep 2024.

    Copyright: © 2024 Ennab and Mcheick. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) or licensor are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

    * Correspondence: Mohammad Ennab, Université du Québec à Chicoutimi, Chicoutimi, Canada

    Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.