Skip to main content

ORIGINAL RESEARCH article

Front. Cardiovasc. Med., 23 August 2024
Sec. Clinical and Translational Cardiovascular Medicine
This article is part of the Research Topic Novel Translational Advances in Hemodynamics for the Diagnosis and Treatment of Cardiovascular Diseases View all 12 articles

Feasibility of tongue image detection for coronary artery disease: based on deep learning

\r\nMengyao Duan,&#x;Mengyao Duan1,†Boyan Mao,&#x;Boyan Mao2,†Zijian LiZijian Li1Chuhao WangChuhao Wang2Zhixi HuZhixi Hu3Jing Guan
Jing Guan1*Feng Li
\r\nFeng Li1*
  • 1School of Traditional Chinese Medicine, Beijing University of Chinese Medicine, Beijing, China
  • 2School of Life Science, Beijing University of Chinese Medicine, Beijing, China
  • 3School of Traditional Chinese Medicine, Hunan University of Chinese Medicine, Changsha, China

Aim: Clarify the potential diagnostic value of tongue images for coronary artery disease (CAD), develop a CAD diagnostic model that enhances performance by incorporating tongue image inputs, and provide more reliable evidence for the clinical diagnosis of CAD, offering new biological characterization evidence.

Methods: We recruited 684 patients from four hospitals in China for a cross-sectional study, collecting their baseline information and standardized tongue images to train and validate our CAD diagnostic algorithm. We used DeepLabV3 + for segmentation of the tongue body and employed Resnet-18, pretrained on ImageNet, to extract features from the tongue images. We applied DT (Decision Trees), RF (Random Forest), LR (Logistic Regression), SVM (Support Vector Machine), and XGBoost models, developing CAD diagnostic models with inputs of risk factors alone and then with the additional inclusion of tongue image features. We compared the diagnostic performance of different algorithms using accuracy, precision, recall, F1-score, AUPR, and AUC.

Results: We classified patients with CAD using tongue images and found that this classification criterion was effective (ACC = 0.670, AUC = 0.690, Recall = 0.666). After comparing algorithms such as Decision Tree (DT), Random Forest (RF), Logistic Regression (LR), Support Vector Machine (SVM), and XGBoost, we ultimately chose XGBoost to develop the CAD diagnosis algorithm. The performance of the CAD diagnosis algorithm developed solely based on risk factors was ACC = 0.730, Precision = 0.811, AUC = 0.763. When tongue features were integrated, the performance of the CAD diagnosis algorithm improved to ACC = 0.760, Precision = 0.773, AUC = 0.786, Recall = 0.850, indicating an enhancement in performance.

Conclusion: The use of tongue images in the diagnosis of CAD is feasible, and the inclusion of these features can enhance the performance of existing CAD diagnosis algorithms. We have customized this novel CAD diagnosis algorithm, which offers the advantages of being noninvasive, simple, and cost-effective. It is suitable for large-scale screening of CAD among hypertensive populations. Tongue image features may emerge as potential biomarkers and new risk indicators for CAD.

1 Introduction

The World Health Organization (WHO) has declared that cardiovascular disease (CVD), particularly coronary artery disease (CAD), is the leading cause of death due to illness globally, accounting for 17.9 million fatalities per year, representing 32% of all deaths caused by disease (1). This significant public health challenge places a significant financial burden on national health budgets (2). Hypertension is an important independent risk factor for the development of CAD (3). Relevant studies have shown that patients with hypertension have a higher risk of developing CAD, and when the two diseases coexist, there is a significant increase in the risk of cardiovascular death (4, 5). Early diagnosis and prompt treatment of patients with CAD have been shown to significantly improve outcomes and reduce treatment costs (3). However, the gold standard for diagnosing CAD is invasive coronary angiography, which is expensive and can cause complications (6). It is not suitable for early diagnosis and disease risk assessment. Finding non-invasive, cost-effective and efficient methods for early CAD diagnosis is crucial in global public health.

The rapid development of artificial intelligence in recent years has provided new insights into the exploration of non-invasive diagnostic methods for CAD. Clinical data exhibit complex and multidimensional characteristics, in which machine learning (ML) demonstrates advantages over traditional statistical methods (7). ML involves the selection and integration of multiple models. When confronted with the complex nonlinear relationships in clinical data, traditional statistical methods often struggle to accomplish modeling tasks. However, ML algorithms can automatically learn to handle nonlinear relationships and select useful predictive features. These algorithms can more effectively reveal hidden relationships within data and has been increasingly utilized for the diagnosis and risk prediction of clinical diseases.

Some scholars have already utilized ML to develop diagnostic models for CAD, with the predictive variables primarily being clinical risk factors (810). These studies demonstrate the promising application prospects of ML in clinical diagnostic tasks. Recent research has found that, in addition to risk factors, other biological information may also hold significant importance for the diagnosis of CAD, such as facial images and pulse waves (11, 12). In clinical diagnosis, the primary focus is on the patient's symptoms and signs.Traditional Chinese Medicine (TCM) employs unique and effective diagnostic strategies, particularly in observing the external conditions of patients. TCM theory posits that “internal diseases manifest externally,” thereby allowing practitioners to gauge the severity of illnesses through observation. Tongue diagnosis is a critical component of the TCM observation process. The appearance of the tongue, including its color, shape, and coating, has long been utilized in TCM to diagnose various health conditions. From a biomedical perspective, the tongue is a highly vascular organ closely related to the cardiovascular system. Changes in blood circulation and overall systemic health often manifest as observable alterations in the tongue's appearance. Many studies have proven the effectiveness of diagnosing diseases through tongue observation (references 14–27), and we have compiled these studies into Table 1 and provided commentary on each. From a biomedical standpoint, the tongue contains rich physiological and pathological information. It is an important terminal organ with abundant blood supply, closely linked to the cardiovascular system. When there are issues with blood circulation, the tongue's appearance often changes (27, 28). Biomedical research suggests that hypoxemia can lead to changes in tongue color and is associated with various cardiovascular diseases (29). In the case of CHD, the narrowing of the coronary arteries restricts blood flow to the heart, potentially causing systemic changes in overall circulation and oxygenation levels, which manifest on the tongue. Despite this, the tongue has not been effectively utilized in the actual diagnosis process of CAD. Recent advancements in artificial intelligence and deep learning have made it possible to extract and analyze subtle features from medical images, including tongue images, which were previously difficult to quantify. By leveraging these technologies, it is possible to detect patterns and features in tongue images associated with CAD, providing a non-invasive, cost-effective, and accessible diagnostic tool. Therefore, what exactly is the diagnostic value of the tongue for CAD? Can tongue images become a crucial basis for optimizing non-invasive diagnosis of CAD? This is precisely the question this study aims to explore.

Table 1
www.frontiersin.org

Table 1. Research on tongue image-assisted disease diagnosis in the past 3 years.

To this end, we conducted a multi-center cross-sectional clinical study, utilizing deep learning methods to explore the potential connection between tongue image features and CAD. Additionally, we aimed to investigate the feasibility of optimizing CAD diagnostic models by incorporating tongue image features as risk factors. Meanwhile, we also hope to present a new and effective biomarker for the clinical diagnosis of CAD.

2 Materials and methods

2.1 Study population and ethical statement

From March 2019 to November 2022, hypertensive patients aged 18–85 were recruited from the cardiology departments of Dongzhimen Hospital, Dongfang Hospital, the Third Affiliated Hospital of Beijing University of Chinese Medicine, and the First Affiliated Hospital of Hunan University of Chinese Medicine. All participants signed an informed consent form, and the study was conducted in accordance with the Declaration of Helsinki. The ethical review of this study was carried out and approved by the Institutional Review Board (IRB) of Shuguang Hospital affiliated with Shanghai University of Traditional Chinese Medicine (IRB number: 2018-626-55-01), with the clinical trial registration number ChiCTR1900026008. All source codes and data analyzed in this study can be obtained from the corresponding author upon reasonable request.

The diagnostic criteria for hypertension refer to the “Chinese Guidelines for the Prevention and Treatment of Hypertension,” which define hypertension as a systolic blood pressure ≥140 mmHg or diastolic blood pressure ≥90 mmHg, or currently undergoing treatment with antihypertensive medication (30). The diagnosis of CAD is based on the patient's CAG results, that is, a narrowing of the inner diameter of at least one of the coronary arteries (left anterior descending, left circumflex, right coronary artery, or left main) by ≥50%. Initially, 684 patients were recruited, with the following exclusion criteria: (1) patients whose tongue appearance was altered by medications or food, (2) no definitive diagnosis related to CAD, prior percutaneous coronary intervention (PCI) or coronary artery bypass grafting (CABG); (3) those with a severe lack of relevant clinical pathophysiological information; (4) those with poor quality tongue images.

2.2 Data collection

Trained research physicians conducted interviews and took tongue photographs of participants following a standardized collection process. The interviews gathered baseline data on general conditions, socioeconomic status, lifestyle (alcohol consumption, smoking, insomnia), and clinical manifestations. Tongue photographs were collected using a TFDA-1 tongue diagnosis instrument (Figure 1), two hours after breakfast or lunch. The specific steps for image collection are as follows: (1) Power on the Instrument after inspection and adjust the camera parameters. (2) Disinfect the areas of the instrument that may come into direct contact with the participant using 75% alcohol. (3) Instruct the patient to place their face on the chin rest, relax, and stick out their tongue flatly. (4) Turn on the built-in ring light source and complete the image capture. (5) Check the photo; if it is satisfactory, the collection is complete; if not, retake the photo until the image quality meets the standard. Qualification criteria for photo quality: no problems such as occlusion, blurring, fogging, overexposure, or underexposure; the tongue should be relaxed and flattened with no twisting or tension; there should be no foreign objects, staining, or other conditions affecting the appearance of the tongue surface.

Figure 1
www.frontiersin.org

Figure 1. The tongue diagnosis instrument and collection process. 1: lens hood, 2: LED light resource, 3: high-definition camera, 4: chin support plate. Note. Use fixed standard camera parameters when shooting: the color temperature is 5,000 k, the color rendering index is 97, the frame rate is 1/125 s, the aperture is F/6.3, the exposure indicator scale is 0 or ±1. (A) Front view of the device; (B) Side view of the device; (C) Rear view of the device; (D) Schematic of the image acquisition process; (E) Perspective of the operator.

2.3 Data preprocessing

Patients were divided into two groups based on whether they were diagnosed with CAD: the hypertension group and the hypertension combined with CAD group, with the labels recorded as 0 and 1, respectively.

Considering that tongue images also contain other facial information, which is superfluous for this study, we built a deep learning model for semantic segmentation of the tongue body using the DeepLabV3 + framework (Figure 2A). We used 500 images from a national key research and development program tongue image database for model training, implementing a phased training strategy. In the first 50 epochs of training (Figure 2B), the backbone of the model was frozen to focus on fine-tuning the tail end of the network, with a batch size set to 8. Subsequently, in the unfreezing phase, all network layers were involved in training, with the batch size adjusted to 4 and the learning rate set to 0.01. After completing the segmentation of the tongue body, the image size was uniformly cropped and adjusted to 256 × 256 pixels (Figure 2C). Additionally, due to our overall small sample size, data augmentation was performed on the images through rotation, flipping, and translation.

Figure 2
www.frontiersin.org

Figure 2. Data preprocessing for tongue images. (A) DeepLabV3 + framework diagram, (B) model training loss function graph, (C) preprocessing effect on tongue image.

For the baseline data of patients obtained through interviews, there was only a minimal amount of missing data (less than 5% missing). Various interpolation methods were used to fill in the missing data. For discrete variables in the baseline data, such as gender and ethnicity, one-hot encoding was employed for data preprocessing.

2.4 Development of a CAD diagnostic algorithm

The customization of a CAD diagnostic algorithm primarily encompasses two core steps: the classification of images using deep learning frameworks, and the construction of diagnostic models utilizing common ML techniques. In this study, we utilized the ResNet-18 network (Figure 3A), pretrained on the ImageNet dataset, as the foundation for our deep learning architecture. As a deep residual network, ResNet-18 effectively mitigates the vanishing gradient problem encountered in training deep networks through residual learning, making it widely applicable in image recognition, especially in the field of medical image processing. It has demonstrated excellent performance in various tongue image processing tasks.

Figure 3
www.frontiersin.org

Figure 3. Tongue image-based CAD diagnostic algorithm. (A) ResNet-18 framework used in this study, (B) 5-fold cross-validation mean ROC on training set, (C) 5-fold cross-validation mean ROC Comparison on training set, (D) performance comparison of different feature inputs on the validation set.

During the training process, we chose to freeze the first and second layers (layer1 and layer2) of the model, training only the parameters of the third and fourth layers (layer3 and layer4) along with the fully connected layer. This strategy not only helps prevent overfitting, which might arise due to the small size of the dataset, but also further enhances the model's ability to learn representations. Stochastic Gradient Descent (SGD) was employed as the optimizer, with cross-entropy loss function used as the loss function. Upon convergence, the output of the penultimate layer of the model forms a 512-dimensional deep feature vector, which is then linked to a binary classification output layer to produce probabilities.

After extracting the deep feature vectors, we constructed a CAD diagnostic model that integrates tongue image feature vectors with risk factors. To optimize the model's performance, we explored a variety of common ML algorithms, including Random Forest (RF), Support Vector Machine (SVM), Decision Trees (DT), Logistic Regression (LR), and XGBoost, all of which are widely used in disease classification and risk prediction. By comparing the performance of these algorithms, we will select the one with the best performance as the core algorithm for our final CAD diagnostic model.

2.5 Statistical analysis

Using SPSS 27 and Python 3.8 as statistical tools, we processed and analyzed the data. For data that followed a normal distribution, we employed descriptive statistics using X¯±SD, along with one-way ANOVA to explore differences between groups. Before conducting the one-way ANOVA, we performed a homogeneity of variance test, such as Levene's test, to ensure that the variances across groups were roughly equal. For data not meeting the normal distribution criteria, we opted for quartile descriptions and employed the Kruskal–Wallis H test to compare differences between two groups. Additionally, we conducted an in-depth analysis of potential risk factors for CAD using binary logistic regression, presenting the results with adjusted odds ratios (adjusted OR) and their 95% confidence intervals (CI). Throughout the analysis, differences were considered statistically significant when the p-value < 0.05.

2.6 Performance evaluation standard

To evaluate the algorithm's performance, we calculated the Accuracy, Precision, Recall, F-1 Score, AUC (Area Under the Curve), and AUPR (Area Under the Precision-Recall Curve) (Equations 14) (31).

Accuracy=TP+TNTP+FN+FP+TN(1)
Precision=TPTP+FP(2)
Recall=TPTP+FN(3)
F1=2×Precision×RecallPrecision+Recall(4)

For binary classification problems, examples can be divided into four categories based on the combination of their true labels and the predictions made by the classifier: True Positives (TP), False Positives (FP), True Negatives (TN), and False Negatives (FN). In addition, we performed a 5-fold cross-validation on the training set to evaluate the model's effectiveness.

We evaluated the segmentation results of the DeepLabV3 + model using Pixel Accuracy (PA) and Mean Intersection over Union (MIoU) (Equations 5, 6). The formulas for calculating PA and MIoU are as follows, where k represents the number of categories excluding the background, and pij denotes the number of pixels of class i predicted to be class j (32):

PA=i=0kpiii=0kj=0kpij(5)
MIoU=1k+1i=1kpiij=0kpij+j=0kpjipii(6)

3 Results

3.1 Patient recruitment and data analysis

As shown in Figure 4, we recruited a total of 511 hypertension patients from the cardiology departments of Dongzhimen Hospital, Dongfang Hospital, and the Third Affiliated Hospital of Beijing University of Chinese Medicine, and an additional 173 patients from the First Affiliated Hospital of Hunan University of Chinese Medicine. These patients underwent interviews and had standard tongue photographs taken. After screening, we excluded 60 patients (8.77%) with disqualifying tongue images, 164 patients (23.98%) with more than 5% missing data, and 88 patients (12.87%) without a coronary CAD record and who could not be definitively ruled out for CAD. Ultimately, we included a total of 244 hypertension patients and 166 patients with hypertension combined with CAD in our study.

Figure 4
www.frontiersin.org

Figure 4. Study flowchart. (A) Workflow for algorithm development; (B) Workflow for algorithm validation.

Table 2 provides detailed baseline information for both patient groups. Upon analysis, we found significant differences between groups in terms of age, BMI, and the duration of hypertension across the training, validation, and test sets. To further explore potential risk factors, we analyzed the data from the training and validation sets using logistic regression (Table 3). The results indicated that older age, male gender, and having hyperlipidemia are risk factors for CAD, while having a higher education, Insomnia, and using antihypertensive medication were considered protective factors.

Table 2
www.frontiersin.org

Table 2. Basic information of three groups.

Table 3
www.frontiersin.org

Table 3. Binary logistic regression results of the main coronary artery disease risk factors.

3.2 Performance of tongue feature extraction algorithm

In the task of semantic segmentation of the tongue body, we employed a customized algorithm based on the Deeplab V3 + framework, which demonstrated outstanding performance. The overall accuracy exceeded 99%, with a Mean Intersection over Union (mIoU) for the tongue segmentation task reaching 98.77%, and the Mean Pixel Accuracy (mPA) was as high as 99.45%. Such results fully attest to the effectiveness and accuracy of our algorithm in the task of tongue body semantic segmentation.

We developed a CAD diagnostic algorithm based on the ResNet-18 framework, which uses tongue images as input. On the training set, the algorithm showed a mean accuracy of 0.690, a mean AUC value of 0.749, and a mean recall rate of 0.842 (Table 4); on the test set, it achieved an accuracy of 0.670, an AUC value of 0.690, and a recall rate of 0.666 (Table 5). As shown in Figures 3A,C, the algorithm demonstrates certain classification capabilities on both the training and test sets, indicating that tongue images indeed possess classification value for the diagnosis of CAD in patients with hypertension.3.3 Performance of CAD Diagnostic Algorithm.

Table 4
www.frontiersin.org

Table 4. Performance of 5-fold cross-validation with different feature inputs.

Table 5
www.frontiersin.org

Table 5. Performance on the test Set with different feature inputs.

We developed two types of CAD diagnostic models: one based solely on CAD risk factors and the other combining risk factors with deep features of tongue images. To determine the optimal ML approach, we compared the performance of various algorithms, all utilizing risk factors and deep features of tongue images as inputs. Results on the training set showed that different ML algorithms could effectively complete the classification task, with XGBoost exhibiting the best performance (Table 6). Therefore, we ultimately selected XGBoost as the method for algorithm customization. Moreover, we compared algorithm customized only with risk factors to those incorporating tongue image features, it was found that the inclusion of tongue image features significantly enhanced algorithm performance, indicating that adding tongue features as input variables positively contributes to algorithm optimization (Figure 3B; Table 4).

Table 6
www.frontiersin.org

Table 6. Comparison of 5-fold cross-validation results across different algorithms.

We also evaluated the performance of different ML algorithms on the test set (Table 7), and the results were broadly consistent with those on the validation set. Although there was a slight decrease in performance, the algorithms still demonstrated good classification capabilities, with XGBoost continuing to show the best performance. Additionally, we compared algorithms developed solely based on risk factors with those that integrate both risk factors and tongue image features, using the test set for evaluation (Figure 3C; Table 5). This finding confirms the practical diagnostic value of tongue images for CAD and also indicates the potential of tongue images to enhance the efficacy of current diagnostic models for the condition.

Table 7
www.frontiersin.org

Table 7. The performance comparison of different algorithm on the test set.

To more comprehensively evaluate the algorithm's applicability and performance across different populations, we subdivided the test set according to age, gender, and the number of risk factors, and presented the algorithm's performance across various subgroups. As shown in Figure 5, in terms of age distribution, the algorithm demonstrated superior diagnostic ability in the elderly population aged 65 and above. Regarding gender, our algorithm demonstrated relatively stable performance between men and women, with no significant differences. In terms of risk factors, the algorithm's judgment ability significantly improved when the number of risk factors reached or exceeded three; however, with fewer risk factors, the algorithm's performance was comparatively weaker.

Figure 5
www.frontiersin.org

Figure 5. Algorithm performance in subgroups of test group. (A) AUC for the model in individuals under 65 years old; (B) AUC for the model in individuals 65 years and older; (C) AUC for the model in males; (D) AUC for the model in females; (E) AUC for the model in individuals with fewer than 3 risk factors; (F) AUC for the model in individuals with 3 or more risk factors.

4 Discussion

Hypertensive patients constitute a large patient population and are an important risk factor for coronary heart disease. When the two conditions occur simultaneously, they can result in a higher burden of disease and accidental risks (4). Therefore, this study selected hypertensive patients as the target population. Tongue diagnosis is an important diagnostic method in TCM, closely related to the cardiovascular blood flow status. However, it has been largely overlooked in the actual diagnosis of CAD. In recent years, there have been many studies on constructing artificial intelligence diagnostic models for CAD, but none have attempted to utilize the tongue, a biological marker of the human body. We conducted a multi-center cross-sectional study, customizing a CAD diagnostic algorithm that integrates risk factors and tongue features based on clinically accessible data, achieving moderate performance. We innovated a method using deep learning to optimize CAD diagnosis through tongue images. This method has the advantages of being non-invasive, low-cost, and easy to operate compared to coronary angiography, and it exhibits better diagnostic performance than traditional CAD diagnostic algorithms that use risk factors. Our results affirm the practical diagnostic value of the tongue for CAD and demonstrate the feasibility of enhancing CAD diagnostic algorithm performance with tongue images.

For the development of a CAD diagnostic model, the rational use of clinical risk factors is of great importance. Various CAD diagnostic and prediction models, such as the Framingham (33) and the Systematic Coronary Risk Evaluation (SCORE) (34) heavily rely on clinical risk factors as a crucial basis. Inspired by this approach, our study first analyzed potential clinical risk factors for CAD. In our dataset, not all known CAD risk factors played a risk role, and insomnia, previously considered a risk factor (35) acted as a protective factor. We developed a CAD diagnostic algorithm that uses clinical risk factors as inputs, which exhibited moderate diagnostic performance (ACC = 0.770). Diagnostic models developed in other studies based on clinical risk factors showed similar performance (3638). Although the performance of our non-invasive diagnostic model still needs further improvement, which may be related to the grouping method, sample size, and model construction approach, our results still demonstrate the potential of using tongue images for non-invasive diagnosis of CAD. Based on this, we attempted to incorporate tongue image features as inputs to develop a new CAD diagnostic algorithm. To accurately and objectively explore the value of tongue images for CAD diagnosis, we used the TFDA-1 tongue diagnosis instrument for image capture and established a standardized data collection process. We eliminated potential interference from other facial information through tongue body semantic segmentation and image cropping. In customizing the algorithm for diagnosing CAD with tongue images, we employed the Resnet-18 as the deep learning framework. Resnet-18 is a type of deep residual network (39) that has consistently shown good performance in past studies on tongue images (4042). In the analysis of tongue image features using traditional feature engineering, medical prior knowledge is often borrowed to define features based on the color, shape, and texture of the tongue, ensuring that the features possess good interpretability and medical significance.

That is precisely why our research opted for deep learning instead of traditional feature engineering. Despite the complexity and lack of explainability in the decision-making process of deep learning, this method simplifies the feature extraction process compared to traditional tongue image feature extraction engineering (43). It eliminates the need for extensive manual labeling of image features and enables fast and efficient automatic learning of complex feature representations in images, uncovering hidden information. The results indicate that, although there was a slight decline in the algorithm's performance on the test set compared to the validation set, the algorithm still retains certain classification capabilities overall. This suggests that tongue images have a definite diagnostic value for CAD, making the tongue an effective biological marker for CAD diagnosis.

Ultimately, we incorporated both risk factors and tongue image features as inputs and developed a new CAD diagnostic algorithm using XGBoost. This algorithm demonstrates superior performance compared to those utilizing single-type features as inputs. This part of the work validates the practical effectiveness of tongue images in enhancing the performance of CAD diagnostic algorithms, proving the feasibility of supplementing clinical diagnosis with TCM diagnostic theories. It offers a new perspective on integrating traditional medical knowledge with modern technology. Additionally, we also focused on the algorithm's performance across different demographic subgroups. The results indicate that the algorithm has better diagnostic capability in elderly populations aged 65 years and above, which may be related to the higher prevalence of CAD in older individuals. The algorithm performs similarly in both men and women, indicating that our developed algorithm has commendable generalization capability across genders. The model exhibits higher accuracy in judgments when the number of risk factors is three or more, highlighting the importance of considering multiple risk factors in the diagnosis of CAD. This approach enhances our understanding of the model's generalizability, revealing its applicability to patient groups with varying demographic and clinical profiles. It holds significant value for clinical practice, offering a reference for tailoring diagnostic methods based on demographic characteristics to improve diagnostic accuracy. By evaluating the model's performance in different subgroups, we can identify potential biases that might affect accuracy, thereby rendering it more reliable in clinical settings.

Although this study identified the potential value of the tongue in diagnosing CAD, it also has some limitations. Firstly, despite being a multi-center study, there are only four hospitals from two regions involved in the sub-centers, lacking subjects from different ethnicities and countries. Secondly, this study is focused solely on hypertensive populations, and the overall sample size is relatively small, which may limit the possibility of applying the CAD diagnostic algorithm to a wider population. Thirdly, while this study employed standardized tongue image collection equipment to minimize interference from other factors during image capture, it also restricts the application of the model in different scenarios and with different collection devices. Although this study explores the potential value of tongue diagnosis for CAD, future research needs to further validate and optimize our diagnostic model in a wider and larger population, carrying out prospective studies. We also experimented with using different types of cameras, various light sources, and even mobile portable devices for image collection, to further expand the model's applicability and enhance its generalization capability. In future research, we can expect more optimized and interpretable deep learning models to enhance the study results, capturing finer changes in tongue images more accurately, thus further optimizing the findings of this study. Additionally, tongue diagnosis is only an essential component of TCM diagnosis, and we can further focus on the integration of multimodal data, considering the fusion of other biomarkers with tongue images to build a more comprehensive and integrated CAD diagnostic model.

5 Conclusion

Exploring an inexpensive, non-invasive diagnostic tool that can be used for early-stage and large-scale screening of CAD is essential. In this study, we analyzed potential risk factors for CAD, extracted potential diagnostic features from tongue images, and developed a new, well-performing CAD diagnostic algorithm based on these findings. Our work introduces a novel perspective, suggesting that tongue images have applicable diagnostic value for CAD diagnosis. Tongue image features could become new risk indicators for CAD, demonstrating the feasibility of integrating TCM theories with modern technology.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by Shuguang Hospital affiliated with Shanghai University of Traditional Chinese Medicine (IRB number: 2018-626-55-01). The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study.

Author contributions

MD: Data curation, Funding acquisition, Investigation, Methodology, Project administration, Software, Validation, Writing – original draft, Writing – review & editing. BM: Funding acquisition, Methodology, Software, Validation, Writing – original draft, Writing – review & editing. ZL: Data curation, Formal Analysis, Writing – original draft, Writing – review & editing. CW: Data curation, Methodology, Writing – original draft, Writing – review & editing. ZH: Data curation, Methodology, Writing – original draft, Writing – review & editing. JG: Data curation, Investigation, Supervision, Validation, Writing – original draft, Writing – review & editing. FL: Conceptualization, Data curation, Funding acquisition, Investigation, Supervision, Validation, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article.

Our Study was supported by Beijing Municipal Education Science Planning Project [grant numbers CDAA21037]; the China Postdoctoral Science Foundation [grant numbers 2024M750261]; the Postdoctoral Fellowship Program of CPSF [grant numbers GZC20230323]; the Ministry of Science and Technology of the People’s Republic of China [grant numbers 2017FYC1703301]; National Natural Science Foundation of China [grant numbers 12102064].

Acknowledgments

We would like to thank all participants and their family members for being willing to participate in this study.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. World Health Organization. Cardiovascular Diseases (CVDs). Geneva, Switzerland: World Health Organization (2019). Available online at: https://www.who.int/news-room/fact-sheets/detail/cardiovascular-diseases-(cvds)

Google Scholar

2. Roth GA, Mensah GA, Johnson CO, Addolorato G, Ammirati E, Baddour LM, et al. Global burden of cardiovascular diseases and risk factors, 1990–2019: update from the GBD 2019 study. J Am Coll Cardiol. (2020) 76(25):2982–3021. doi: 10.1016/j.jacc.2020.11.010

PubMed Abstract | Crossref Full Text | Google Scholar

3. Meng H, Ruan J, Yan Z, Chen Y, Liu J, Li X, et al. New progress in early diagnosis of atherosclerosis. Int J Mol Sci. (2022) 23(16):8939. doi: 10.3390/ijms23168939

PubMed Abstract | Crossref Full Text | Google Scholar

4. Charchar FJ, Prestes PR, Mills C, Ching SM, Neupane D, Marques FZ, et al. Lifestyle management of hypertension: international society of hypertension position paper endorsed by the world hypertension league and European society of hypertension. J Hypertens. (2024) 42(1):23–49. doi: 10.1097/HJH.0000000000003563

PubMed Abstract | Crossref Full Text | Google Scholar

5. Berge CA, Eskerud I, Almeland EB, Larsen TH, Pedersen ER, Rotevatn S, et al. Relationship between hypertension and non-obstructive coronary artery disease in chronic coronary syndrome (the NORIC registry). PLoS One. (2022) 17(1):e0262290. doi: 10.1371/journal.pone.0262290

PubMed Abstract | Crossref Full Text | Google Scholar

6. Truesdell AG, Alasnag MA, Kaul P, Rab ST, Riley RF, Young MN, et al. Intravascular imaging during percutaneous coronary intervention: JACC state-of-the-art review. J Am Coll Cardiol. (2023) 81(6):590–605. doi: 10.1016/j.jacc.2022.11.045

PubMed Abstract | Crossref Full Text | Google Scholar

7. Suri JS, Bhagawati M, Paul S, Protogeron A, Sfikakis PP, Kitas GD, et al. Understanding the bias in machine learning systems for cardiovascular disease risk assessment: the first of its kind review. Comput Biol Med. (2022) 142:105204. doi: 10.1016/j.compbiomed.2021.105204

PubMed Abstract | Crossref Full Text | Google Scholar

8. Tao S, Yu L, Yang D, Yao R, Zhang L, Huang L, et al. Development and validation of a clinical prediction model for detecting coronary heart disease in middle-aged and elderly people: a diagnostic study. Eur J Med Res. (2023) 28(1):375. doi: 10.1186/s40001-023-01233-0

PubMed Abstract | Crossref Full Text | Google Scholar

9. Mirjalili SR, Soltani S, Heidari Meybodi Z, Marques-Vidal P, Kraemer A, Sarebanhassanabadi M. An innovative model for predicting coronary heart disease using triglyceride-glucose index: a machine learning-based cohort study. Cardiovasc Diabetol. (2023) 22(1):200. doi: 10.1186/s12933-023-01939-9

PubMed Abstract | Crossref Full Text | Google Scholar

10. Wang Z, Sun Z, Yu L, Wang Z, Li L, Lu X. Machine learning-based prediction of composite risk of cardiovascular events in patients with stable angina pectoris combined with coronary heart disease: development and validation of a clinical prediction model for Chinese patients. Front Pharmacol. (2024) 14:1334439. doi: 10.3389/fphar.2023.1334439

PubMed Abstract | Crossref Full Text | Google Scholar

11. Lin S, Li Z, Fu B, Chen S, Li X, Wang Y, et al. Feasibility of using deep learning to detect coronary artery disease based on facial photo. Eur Heart J. (2020) 41(46):4400–11. doi: 10.1093/eurheartj/ehaa640

PubMed Abstract | Crossref Full Text | Google Scholar

12. Wang S, Wu D, Li G, Song X, Qiao A, Li R, et al. A machine learning strategy for fast prediction of cardiac function based on peripheral pulse wave. Comput Methods Programs Biomed. (2022) 216:106664. doi: 10.1016/j.cmpb.2022.106664

PubMed Abstract | Crossref Full Text | Google Scholar

13. Rajakumaran S, Sasikala J. Improvement in tongue color image analysis for disease identification using deep learning based depthwise separable convolution model. Indian J Comput Sci Eng. (2021) 12(1):21–32. doi: 10.21817/indjcse/2021/v12i1/211201082

Crossref Full Text | Google Scholar

14. Jiang T, Guo X, Tu L, Lu Z, Cui J, Ma X, et al. Application of computer tongue image analysis technology in the diagnosis of NAFLD. Comput Biol Med. (2021) 135:104466. doi: 10.1016/j.compbiomed.2021.104466

Crossref Full Text | Google Scholar

15. Mansour RF, Althobaiti MM, Ashour AA. Internet of things and synergic deep learning based biomedical tongue color image analysis for disease diagnosis and classification. IEEE Access. (2021) 9:1. doi: 10.1109/ACCESS.2021.3094226

Crossref Full Text | Google Scholar

16. Li J, Chen Q, Hu X, Yuan P, Cui L, Tu L, et al. Establishment of noninvasive diabetes risk prediction model based on tongue features and machine learning techniques. Int J Med Inform. (2021) 149:104429. doi: 10.1016/j.ijmedinf.2021.104429

PubMed Abstract | Crossref Full Text | Google Scholar

17. Thanikachalam V, Shanthi S, Kalirajan K, Abdel-Khalek S, Omri M, Ladhar LM. Intelligent deep learning based disease diagnosis using biomedical tongue images. CMC Computers Mater Continua. (2022) 3:15. doi: 10.32604/cmc.2022.020965

Crossref Full Text | Google Scholar

18. Jiang T, Lu Z, Hu X, Zeng L, Ma X, Lu Z, et al. Deep learning multi-label tongue image analysis and its application in a population undergoing routine medical checkup. Evid Based Complement Alternat Med. (2022) 2022:3384209. doi: 10.1155/2022/3384209

PubMed Abstract | Crossref Full Text | Google Scholar

19. Li J, Huang J, Jiang T, Tu L, Cui L, Cui J, et al. A multi-step approach for tongue image classification in patients with diabetes. Comput Biol Med. (2022) 149:105935. doi: 10.1016/j.compbiomed.2022.105935

PubMed Abstract | Crossref Full Text | Google Scholar

20. Sivasubramaniam S, Balamurugan SP. Simulated annealing with deep learning based tongue image analysis for heart disease diagnosis. Intell Autom Soft Comput. (2023) 37(1):111–26. doi: 10.32604/iasc.2023.035199

Crossref Full Text | Google Scholar

21. Zhang N, Jiang Z, Li JX, Zhang D. Multiple color representation and fusion for diabetes mellitus diagnosis based on back tongue images. Comput Biol Med. (2023) 155:106652. doi: 10.1016/j.compbiomed.2023.106652

PubMed Abstract | Crossref Full Text | Google Scholar

22. Yuan L, Yang L, Zhang S, Xu Z, Qin J, Liu X, et al. Development of a tongue image-based machine learning tool for the diagnosis of gastric cancer: a prospective multicentre clinical cohort study. EClinicalMedicine. (2023) 57:101834. doi: 10.1016/j.eclinm.2023.101834

PubMed Abstract | Crossref Full Text | Google Scholar

23. Noguchi K, Saito I, Namiki T, Yoshimura Y, Nakaguchi T. Reliability of non-contact tongue diagnosis for Sjögren’s syndrome using machine learning method. Sci Rep. (2023) 13(1):1334. doi: 10.1038/s41598-023-27764-4

PubMed Abstract | Crossref Full Text | Google Scholar

24. Shi Y, Guo D, Chun Y, Liu J, Liu L, Tu L, et al. A lung cancer risk warning model based on tongue images. Front Physiol. (2023) 14:1154294. doi: 10.3389/fphys.2023.1154294

PubMed Abstract | Crossref Full Text | Google Scholar

25. Dai S, Guo X, Liu S, Tu L, Hu X, Cui J, et al. Application of intelligent tongue image analysis in conjunction with microbiomes in the diagnosis of MAFLD. Heliyon. (2024) 10(7):e29269. doi: 10.1016/j.heliyon.2024.e29269

PubMed Abstract | Crossref Full Text | Google Scholar

26. Duan M, Zhang Y, Liu Y, Mao B, Li G, Han D, et al. Machine learning aided non-invasive diagnosis of coronary heart disease based on tongue features fusion. Technol Health Care. (2024) 32(1):441–57. doi: 10.3233/THC-230590

PubMed Abstract | Crossref Full Text | Google Scholar

27. Skalidis E, Zacharis E, Hamilos M, Skalidis I, Anastasiou I, Parthenakis F. Transient lingual ischemia complicating coronary angiography. J Invasive Cardiol. (2019) 31(3):E51. https://www.researchgate.net/publication/331730352_Transient_Lingual_Ischemia_Complicating_Coronary_Angiography30819980

PubMed Abstract | Google Scholar

28. Chang FY, Natesan S, Goh WY, Anerdis JLD. An elderly woman with tongue ischemia. J Am Geriatr Soc. (2016) 64(10):e111–2. doi: 10.1111/jgs.14395

PubMed Abstract | Crossref Full Text | Google Scholar

29. Hamaoka T, McCully KK. Review of early development of near-infrared spectroscopy and recent advancement of studies on muscle oxygenation and oxidative metabolism. J Physiol Sci. (2019) 69:799–811. doi: 10.1007/s12576-019-00697-2

PubMed Abstract | Crossref Full Text | Google Scholar

30. Writing Group of 2018 Chinese Guidelines for the Management of Hypertension, Chinese Hypertension League, Chinese Society of Cardiology. 2018 Chinese guidelines for the management of hypertension. Chin J Cardiovasc Med. (2019) 24:24–56.

Google Scholar

31. Olaniran OR, Alzahrani ARR, Alzahrani MR. Eigenvalue distributions in random confusion matrices: applications to machine learning evaluation. Mathematics. (2024) 12(10):1425. doi: 10.3390/math12101425

Crossref Full Text | Google Scholar

32. He J, Cheng Y, Wang W, Ren Z, Zhang C, Zhang W. A lightweight building extraction approach for contour recovery in complex urban environments. Remote Sens. (2024) 16(5):740. doi: 10.3390/rs16050740

Crossref Full Text | Google Scholar

33. Mahmood SS, Levy D, Vasan RS, Wang TJ. The framingham heart study and the epidemiology of cardiovascular disease: a historical perspective. Lancet. (2014) 383(9921):999–1008. doi: 10.1016/S0140-6736(13)61752-3

PubMed Abstract | Crossref Full Text | Google Scholar

34. Conroy RM, Pyörälä K, Fitzgerald AP, Sans S, Menotti A, De Backer G, et al. Estimation of ten-year risk of fatal cardiovascular disease in Europe: the SCORE project. Eur Heart J. (2003) 24(11):987–1003. doi: 10.1016/s0195-668x(03)00114-3

PubMed Abstract | Crossref Full Text | Google Scholar

35. Javaheri S, Redline S. Insomnia and risk of cardiovascular disease. Chest. (2017) 152(2):435–44. doi: 10.1016/j.chest.2017.01.026

PubMed Abstract | Crossref Full Text | Google Scholar

36. Huang Y, Ren Y, Yang H, Ding Y, Liu Y, Yang Y, et al. Using a machine learning-based risk prediction model to analyze the coronary artery calcification score and predict coronary heart disease and risk assessment. Comput Biol Med. (2022) 151(Pt B):106297. doi: 10.1016/j.compbiomed.2022.106297

PubMed Abstract | Crossref Full Text | Google Scholar

37. Zhang L, Niu M, Zhang H, Wang Y, Zhang H, Mao Z, et al. Nonlaboratory-based risk assessment model for coronary heart disease screening: model development and validation. Int J Med Inform. (2022) 162:104746. doi: 10.1016/j.ijmedinf.2022.104746

PubMed Abstract | Crossref Full Text | Google Scholar

38. Li D, Xiong G, Zeng H, Zhou Q, Jiang J, Guo X. Machine learning-aided risk stratification system for the prediction of coronary artery disease. Int J Cardiol. (2021) 326:30–4. doi: 10.1016/j.ijcard.2020.09.070

PubMed Abstract | Crossref Full Text | Google Scholar

39. He KM, Zhang X, Ren SQ, Sun J. Deep residual learning for image recognition. 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (2016). p. 770–8. doi: 10.1109/cvpr.2016.90

Crossref Full Text | Google Scholar

40. Ma C, Zhang P, Du S, Li Y, Li S. Construction of tongue image-based machine learning model for screening patients with gastric precancerous lesions. J Pers Med. (2023) 13(2):271. doi: 10.3390/jpm13020271

PubMed Abstract | Crossref Full Text | Google Scholar

41. Li J, Yuan P, Hu X, Huang J, Cui L, Cui J, et al. A tongue features fusion approach to predicting prediabetes and diabetes with machine learning. J Biomed Inform. (2021) 115:103693. doi: 10.1016/j.jbi.2021.103693

PubMed Abstract | Crossref Full Text | Google Scholar

42. Okawa J, Hori K, Izuno H, Fukuda M, Ujihashi T, Kodama S, et al. Developing tongue coating status assessment using image recognition with deep learning. J Prosthodont Res. (2023) 68(3):425–31. doi: 10.2186/jpr.JPR_D_23_00117

PubMed Abstract | Crossref Full Text | Google Scholar

43. Bhatnagar V, Bansod PP. Challenges and solutions in automated tongue diagnosis techniques: a review. Crit Rev Biomed Eng. (2022) 50(1):47–63. doi: 10.1615/CritRevBiomedEng.2022044392

PubMed Abstract | Crossref Full Text | Google Scholar

Keywords: coronary artery disease, deep learning, hypertension, tongue image, early diagnosis

Citation: Duan M, Mao B, Li Z, Wang C, Hu Z, Guan J and Li F (2024) Feasibility of tongue image detection for coronary artery disease: based on deep learning. Front. Cardiovasc. Med. 11:1384977. doi: 10.3389/fcvm.2024.1384977

Received: 11 February 2024; Accepted: 7 August 2024;
Published: 23 August 2024.

Edited by:

Hitomi Anzai, Tohoku University, Japan

Reviewed by:

Eugenio Vocaturo, National Research Council (CNR), Italy
Chang Won Jeong, Wonkwang University, Republic of Korea
Nikolaos Papandrianos, University of Thessaly, Greece

Copyright: © 2024 Duan, Mao, Li, Wang, Hu, Guan and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Feng Li, 1575054688@qq.com; Jing Guan, guanjingishere@126.com

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.