Skip to main content

ORIGINAL RESEARCH article

Front. Oncol., 08 November 2022
Sec. Head and Neck Cancer
This article is part of the Research Topic Methods in Head and Neck Cancer View all 20 articles

Deep learning for the diagnosis of suspicious thyroid nodules based on multimodal ultrasound images

Yi Tao&#x;Yi Tao1†Yanyan Yu*&#x;Yanyan Yu2*†Tong WuTong Wu1Xiangli XuXiangli Xu3Quan DaiQuan Dai1Hanqing KongHanqing Kong1Lei ZhangLei Zhang1Weidong YuWeidong Yu1Xiaoping LengXiaoping Leng1Weibao QiuWeibao Qiu4Jiawei Tian*Jiawei Tian1*
  • 1Department of Ultrasound, The Second Affiliated Hospital of Harbin Medical University, Harbin, China
  • 2The National-Regional Key Technology Engineering Laboratory for Medical Ultrasound, Guangdong Key Laboratory for Biomedical Measurements and Ultrasound Imaging, School of Biomedical Engineering, Health Science Center, Shenzhen University, Shenzhen, China
  • 3Department of Ultrasound, The Second Hospital of Harbin, Harbin, China
  • 4Shenzhen Key Laboratory of Ultrasound Imaging and Therapy, Paul C. Lauterbur Research Center for Biomedical Imaging, Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China

Objectives: This study aimed to differentially diagnose thyroid nodules (TNs) of Thyroid Imaging Reporting and Data System (TI-RADS) 3–5 categories using a deep learning (DL) model based on multimodal ultrasound (US) images and explore its auxiliary role for radiologists with varying degrees of experience.

Methods: Preoperative multimodal US images of 1,138 TNs of TI-RADS 3–5 categories were randomly divided into a training set (n = 728), a validation set (n = 182), and a test set (n = 228) in a 4:1:1.25 ratio. Grayscale US (GSU), color Doppler flow imaging (CDFI), strain elastography (SE), and region of interest mask (Mask) images were acquired in both transverse and longitudinal sections, all of which were confirmed by pathology. In this study, fivefold cross-validation was used to evaluate the performance of the proposed DL model. The diagnostic performance of the mature DL model and radiologists in the test set was compared, and whether DL could assist radiologists in improving diagnostic performance was verified. Specificity, sensitivity, accuracy, positive predictive value, negative predictive value, and area under the receiver operating characteristics curves (AUC) were obtained.

Results: The AUCs of DL in the differentiation of TNs were 0.858 based on (GSU + SE), 0.909 based on (GSU + CDFI), 0.906 based on (GSU + CDFI + SE), and 0.881 based (GSU + Mask), which were superior to that of 0.825-based single GSU (p = 0.014, p< 0.001, p< 0.001, and p = 0.002, respectively). The highest AUC of 0.928 was achieved by DL based on (G + C + E + M)US, the highest specificity of 89.5% was achieved by (G + C + E)US, and the highest accuracy of 86.2% and sensitivity of 86.9% were achieved by DL based on (G + C + M)US. With DL assistance, the AUC of junior radiologists increased from 0.720 to 0.796 (p< 0.001), which was slightly higher than that of senior radiologists without DL assistance (0.796 vs. 0.794, p > 0.05). Senior radiologists with DL assistance exhibited higher accuracy and comparable AUC than that of DL based on GSU (83.4% vs. 78.9%, p = 0.041; 0.822 vs. 0.825, p = 0.512). However, the AUC of DL based on multimodal US images was significantly higher than that based on visual diagnosis by radiologists (p< 0.05).

Conclusion: The DL models based on multimodal US images showed exceptional performance in the differential diagnosis of suspicious TNs, effectively increased the diagnostic efficacy of TN evaluations by junior radiologists, and provided an objective assessment for the clinical and surgical management phases that follow.

1 Introduction

Thyroid cancer has become the most common endocrine malignancy, with an increasing incidence of approximately 7%–15% annually (1, 2). Ultrasound (US) is widely used as a first-line screening tool for the clinical examination of thyroid lesions, with the advantages of no exposure to radiation, real-time dynamic imaging, and simplicity of procedure (1, 3). Multiple versions of the Thyroid Imaging Reporting and Data System (TI-RADS) have been proposed for US imaging to standardize and improve the diagnostic consistency and accuracy of thyroid lesions, and each risk stratification system has its advantages (1, 36).

Nevertheless, US diagnosis of thyroid nodules (TNs) is subjective to a certain extent. Various diagnostic results of US evaluation of TNs were obtained from different observers, especially less-experienced radiologists, who showed relatively lower accuracy. In previous studies, moderate variability in the interobserver agreement was found among different TI-RADS scores (7). There was fair agreement in margin, echotexture, and echogenicity (k = 0.34, 0.26, and 0.34, respectively) for interobserver variability (810). Clinically, there is a wide range of malignant risks (approximately 2%–90%) and some overlapping US features for the TNs of TI-RADS 3–5 categories; therefore, it was difficult for radiologists to accurately differentiate between benign and malignant TNs (1113), resulting in overdiagnosis or misdiagnosis.

Fine-needle aspiration (FNA) is a relatively effective method for the preoperative diagnosis of TNs (14). The radiologists assess the malignant probability of TNs and then recommend patients for FNA or US follow-up according to TI-RADS. However, FNA is an invasive procedure with some possible complications, such as bleeding, and FNA results are also dependent on the size, composition of TNs, and skills of radiologists. Moreover, approximately 20% of the FNA results were rendered inconclusive, which led to uncertainty in the next course of clinical treatment (1517). The development of artificial intelligence (AI) technology has shown great potential in reducing the influence of subjectivity and improving the consistency of diagnosis.

In the past two decades, machine-learning methods have been used in TN characterization, which is usually known as “radiomics” (18, 19). Radiomics can automatically extract features in the region of interest (ROI), which tends to be difficult to discern with the naked eye. It should be noted that high-throughput features extracted by radiomics from the ROI are easily affected by the segmentation strategy and imaging parameters. Deep learning (DL) is a machine-learning concept that has shown strong capability in medical image characterization and outperforms traditional machine-learning methods. With the help of artificial neural networks, DL has been widely applied to differentiate breast, thyroid, and liver lesions with good performance (2022). However, radiologists cannot be completely replaced with AI technology. It is crucial to integrate DL methods into clinical practice; therefore, they can aid radiologists in diagnosis, evaluation, and decision-making (23). In this study, the diagnostic performances of junior and senior radiologists with and without a DL assistant were compared.

Most previous studies using DL for the diagnosis of TNs have concentrated on grayscale US (GSU) imaging. However, beyond conventional GSU, some new US technologies such as color Doppler flow imaging (CDFI), elastography, and contrast-enhanced ultrasonography are commonly used to assist in the diagnosis of GSU for TNs, which have been proven to improve the diagnostic accuracy in the clinical evaluation (13, 24, 25). This indicated that the features of blood flow and hardness also played an important role in thyroid US diagnosis. Therefore, in our study, new DL models based on multimodal US imaging were proposed to explore their application value in improving the diagnostic accuracy of suspicious thyroid lesions and the role of auxiliary diagnosis for radiologists.

2 Materials and methods

2.1 Patients

This retrospective study was approved by the Ethics Committee of The Second Affiliated Hospital of Harbin Medical University, and the requirement for informed consent was waived (approval number KY2021-152). Consecutive patients who had undergone thyroid surgery at The Second Affiliated Hospital of Harbin Medical University between September 9, 2020, and June 6, 2021, were enrolled. The inclusion criteria of the enrolled patients were as follows: lesions with (1) complete or high-quality transverse and longitudinal section images (2), complete surgical records and pathological results (3), no preoperative operation such as FNA and ablation or surgical treatment of TNs, and (4) US examination in our hospital within 1 week before surgery. Finally, 1,138 TNs of TI-RADS 3–5 categories from 781 patients were included in the study. The postoperative pathological results were used as the gold standard. The mean diagnostic age of patients was 47.74 ± 10.60 years (range, 21–79 years). According to the pathological results, there were 550 (48.33%) malignant and 588 (51.67%) benign TNs. The workflow of the selection is shown in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1 Flowchart of enrolling patients with thyroid nodules.

2.2 Ultrasound image acquisition and analysis

Preoperative thyroid US examinations were performed by two radiologists with 10 years of experience (Q.D. and H.K.) using a US device (Hitachi HI VISION Avius, Hitachi Medical Corporation, Tokyo, Japan) equipped with a 5- to 13-MHz linear probe. According to the Chinese TI-RADS (C-TIRADS) issued by the Chinese Society of Ultrasound in Medicine in 2020, thyroid scanning and imaging parameter adjustments were guided and completed (6). The GSU, CDFI, and strain elastography (SE) images of the TNs were acquired in transverse and longitudinal sections, which showed obvious characteristics and were saved in BMP format.

The ultrasonographic features were evaluated for all 1,138 TNs in our study. To maintain consistency, the images were independently analyzed by two experienced radiologists (L.Z. and W.Y.) in a double-blind manner, and results were obtained through consultation by consensus when discrepancies arose. GSU features, including the maximum diameter, position, echotexture, echogenicity, composition, orientation, margin, punctate echogenic foci, halo, and posterior features, were evaluated visually according to the C-TIRADS. CDFI could indicate tumor blood flow characteristics using the vascular distribution pattern and Adler grade (0–3) standards (26). Tumor tissue hardness was evaluated on a scale of 1–4 according to the Asteria standard by SE (27).

2.3 Construction of deep learning

2.3.1 Pretreatment of multimodal and double-view ultrasound images

Four modalities of TN images were included in our research: GSU, CDFI, SE, and ROI mask (Mask) images. Each modal image was captured from both horizontal and vertical perspectives. The multimodal and double-view US images of a TN in the right lobe of a 65-year-old female patient with pathologically proven papillary carcinoma are illustrated in Figure 2. The Masks of the TNs were manually segmented using ImageJ (version 1.48, National Institutes of Health, USA) by two radiologists (Q.D. and Y.T.). The total data set was separated into training, validation, and test data sets, with a ratio of 4:1:1.25.

FIGURE 2
www.frontiersin.org

Figure 2 Multimodal ultrasound images of a thyroid nodule in the right lobe of a 65-year-old female patient with a pathologically proven papillary carcinoma. (A) GSU, (B) CDFI, (C) SE, and (D) Mask ultrasound images in transverse section. (E) GSU, (F) CDFI, (G) SE, and (H) Mask ultrasound images in longitudinal section. GSU, grayscale ultrasound; CDFI, color Doppler flow imaging; SE, strain elastography; Mask, region of interest mask.

2.3.2 Deep residual learning with attention block

Deep networks can extract more abstract information from low-level feature maps, which enables them to perform better than shallow networks. The residue strategy provides a skip connection to solve the degradation problem, making it possible to train a very deep network. To make full use of the multimodal image features, ResNet-50 (28) was used as the backbone for feature extraction in our method. In the ResNet-50, there is one convolutional layer and 16 residual blocks. For the essential composition of ResNet-50, a residual block is defined as follows:

y=x+F(x,W)(1)

where x and y denote the input and output feature maps of the residual block, respectively. F refers to the residual function, which is learned by stacked convolutional layers with different kernel sizes in the residual block. The right side of the equation is obtained by feedforward neural networks with skip connections, which allow gradients to propagate through the networks.

All available multimodal images were preprocessed to a size of 224 × 224 × 3 pixels, where 224 denotes the width and height and 3 denotes the channels of images. The training and validation data sets were randomly divided into five parts for fivefold cross-validation. Multimodal US images of the same patient were sent to the training, validation, or testing data set as one sample. During the training process, the parameters of the modal were optimized by forward and backward propagation computing until the prediction reached a high accuracy related to the ground truth. The feedforward process can be mathematically expressed as follows:

hl=Rl(Wlhl1+bl)(2)

where l denotes the number of layers. hl represents the output feature map of the l layer with hl−1 as the input. W and b denote the weights and biases of the convolutional filter bank, respectively. R is a rectified linear activation (ReLU) function. In back propagation, the parameters of the network are updated by optimizing the following binary cross-entropy loss.

Because of the low contrast and small area of TNs in thyroid US images, it is necessary to obtain effective feature information. However, the key channels and spatial position of the lesion cannot be identified because the information obtained by the convolution operation with the kernel in ResNet is local and may fail to capture effective features from the global image. To solve this problem, we combined the convolutional bottleneck attention module (29) and ResNet-50 to learn the weights for our feature maps (Figure 3). Two attention units were inserted before the first and after the last residual block to obtain abstract features from both the higher and lower layers, as shown in Figure 3A. There are two types of attention mechanisms in the attention unit: spatial attention and channel attention, as shown in Figure 3B. Channel-wise attention was used to select features that could calculate the strongest channel-wise activation values. Spatial attention performs average pooling and max pooling along the channel axis on the feature map to obtain the activated feature map with a local receptive field in the spatial dimension. To complement channel attention, spatial attention was applied to find the informative region for the input feature map in the spatial dimension.

FIGURE 3
www.frontiersin.org

Figure 3 The overall network architecture. (A) Architecture of the backbone network ResNet-50 with two attention units. (B) The attention unit.

2.3.3 Implementation

To establish the DL model, we used 588 benign and 550 malignant TNs with multimodal and double-view images as the data set. Furthermore, fivefold cross-validation was applied to the data sets.

To evaluate the performance of the four types of sonography in thyroid cancer diagnosis, we performed experiments with multimodal inputs (i.e., GSU, CDFI, SE, and Mask). The four streams in Figure 4 correspond to the four modalities. All four modalities (Figure 4), as well as one or multiple modalities of the same patient, were taken as the inputs. Popular ResNet-50 was used as the feature extraction backbone (Figure 3A). The features obtained by multiple network streams from the different modalities and views were averaged and then applied to fully connected layers to predict the classification result. In our experiments, each network stream had its own independent parameters.

FIGURE 4
www.frontiersin.org

Figure 4 The illustration of multimodality inputs and feature fusion.

The framework was implemented on a Dell-T7920 workstation equipped with an NVIDIA GeForce RTX3090 GPU and 64 GB of memory. The Adam optimization algorithm for minibatch gradient descent was used for training with a batch size of 32. The learning rate was initially set to 0.00001 and reduced by 0.1 every 30 epochs. A pretrained model was used for parameter initialization. The models with the smallest loss values within 100 training epochs were selected as the final models to generate classification results. We set the same epochs for training every modal, including the double- and single-view modes.

2.4 Comparing the diagnosis of the deep learning model and radiologists

In this section, we investigate the diagnostic performance of the DL models and radiologists using 228 cases from the test set. According to a survey, the diagnostic accuracy of radiologists increased when they classified the final category into either dichotomous prediction or malignant risk (9). In our study, radiologists diagnosed TNs of the test set based on multimodal US images, and the results were compared with those of the DL method. Five senior radiologists with 5–10 years of experience and five junior radiologists with 1–3 years of experience independently evaluated the TNs and were blinded the diagnosis to the postoperative pathological results. The radiologist then performed a second diagnosis based on the results of the DL and arrived at the final diagnosis. The diagnostic performance of the radiologist alone and in combination with DL assistance was compared.

2.5 Statistical analysis

R software (version 1.8) and MedCalc (version 11.2, Ostend, Belgium) were used to analyze the data. The data set was randomly divided into five non-overlapping groups, whereas there was no data intersection for the same subject for each group. After fivefold cross-validation, the accuracy, sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and area under the receiver operating characteristics curves (AUC) were obtained to evaluate the performance of the presented DL model in the test set. The Delong test results in terms of the AUC for the test data set were introduced to evaluate the statistical difference between DL based on different combined US images and radiologists with variable levels. A 95% confidence interval was used to estimate the range of these evaluation values; p-values of less than 0.05 (two-tailed) were considered statistically significant.

3 Results

3.1 General and ultrasonic characteristic analysis

Among the 781 patients, 135 (17.29%) were men and 646 (82.71%) were women. The mean diagnostic age of the patients was 45.92 ± 10.18 years (range, 22–67 years) for men and 48.11 ± 10.66 years (range, 21–79 years) for women. The average size of malignant TNs (12.5 ± 7.40 mm) was significantly larger than that of benign TNs (9.70 ± 6.40 mm) (p< 0.001) (Table 1).

TABLE 1
www.frontiersin.org

Table 1 Comparing the characteristics of benign and malignant thyroid nodules.

The US characteristics of the 1,138 TNs were statistically analyzed, and the results are listed in Table 1. Except for echotexture (p = 0.649), the risk features of GSU were significantly different between malignant and benign nodules (p< 0.05). We found some significantly different US features in the vascular distribution pattern, Adler grading, and Asteria standard in this study (p< 0.05).

3.2 Diagnostic performance of deep learning models

The performances of the various DL models for differentiating TNs are summarized in Table 2 and Figure 5. In our study, a total of eight DL models were established based on multimodal US imaging. We found that the feature fusion of images from both transverse and longitudinal sections could achieve better performance than that from a single section (Supplementary Text S1).

TABLE 2
www.frontiersin.org

Table 2 Comparing the deep learning diagnostic performance based on multimodal ultrasound images.

FIGURE 5
www.frontiersin.org

Figure 5 The ROC curves of DL-based single GSU and multimodality. ROC, receiver operating characteristics; DL, deep learning; GSU, gray-scale ultrasound.

The AUCs of DL using multimodal US imaging (0.909 based on [G + C]US, 0.858 based on [G + E]US, and 0.906 based on [G + C + E]US) outperformed those of GSU imaging alone (0.825) (p = 0.014, p< 0.001, and p< 0.001, respectively). There was a statistically significant difference in the diagnosis between the (G + C)US and (G + E)US images (0.909 vs. 0.858, p = 0.001). However, the AUC of the DL model based on (G + C + E)US exhibited an excellent performance similar to that based on (G + C)US (0.906 vs. 0.909, p = 0.294), which were both markedly better than DL based on (G + E)US (0.906 vs. 0.858, p = 0.002; 0.909 vs. 0.858, p = 0.001, respectively). In addition, the accuracy, specificity, PPV, and NPV of DL based on (G + C + E)US were better than those of (G + C)US; however, only the PPV was statistically significant (87.8% vs. 85.2%, p = 0.045; Supplementary Text S2).

Furthermore, after adding the Mask feature, the diagnostic performance was obviously better than that of GSU alone (0.881 vs. 0.825, p = 0.002), and the AUCs of (G + C)US, (G + E)US, and (G + C + E)US were also increased (0.918 vs. 0.909, 0.889 vs. 0.858, 0.928 vs. 0.906), but without statistical differences (p = 0.57, p = 0.22, and p = 0.28; Table 2). The highest accuracy of 86.2%, sensitivity of 86.9%, and PPV of 87.8 were achieved by the DL model based on (G + C + M)US, and the highest specificity of 89.5% and NPV of 87.7% were achieved based on (G + C + E)US. The DL model using (G + C + E + M)US images achieved the best performance (AUC of 0.928), with an increase of 10.3% compared with that using a single GSU (p< 0.001).

3.3 Deep learning performance compared with radiologists

The diagnostic performance of radiologists with different levels of experience in differentiating malignant from benign TNs is shown in Table 3 and Figure 6. When independently evaluating the TNs without DL assistance, the diagnosis of senior radiologists showed higher accuracy, specificity, and AUC than that of juniors (80.6% vs. 72.7%, p = 0.008; 81.7% vs. 72.0%, p = 0.018; 0.794 vs. 0.720, p = 0.002, respectively). The sensitivity (79.5%) of US diagnosis by senior radiologists was also better than that (73.5%) of junior radiologists (p = 0.079).

TABLE 3
www.frontiersin.org

Table 3 The diagnostic performance of deep learning (DL), radiologists alone, and DL-assisted radiologists.

FIGURE 6
www.frontiersin.org

Figure 6 The ROC curves of DL and radiologists with different degrees of experience. ROC, receiver operating characteristics; DL, deep learning.

When the DL method was added for the second diagnosis in the test set, the diagnostic performance of junior radiologists significantly increased from 0.720 to 0.796 (p< 0.001). The AUC of the junior radiologists in the second diagnosis was similar to that of the first diagnosis by senior radiologists (0.796 vs. 0.794) but was considered inferior (0.796 vs. 0.822), and the differences were not statistically significant (p > 0.05). Moreover, the DL model also had a certain auxiliary diagnostic effect and could slightly improve the diagnostic performance of senior radiologists in terms of accuracy (from 80.6% to 83.4%), which was higher than that of DL based on GSU (83.4% vs. 78.9%, p = 0.041). However, the AUC of senior radiologists with DL assistance was only comparable to that of DL based on a single GSU (0.822 vs. 0.825, p = 0.512) and significantly less than that of DL based on multimodal US images (0.822 vs. 0.858–0.928, p< 0.05).

4 Discussion

Thyroid cancer has recently become one of the most common malignancies in Chinese women (2). US was the first choice for the examination of thyroid lesions, and TNs were diagnosed on US imaging by radiologists according to TI-RADS. Each guideline has its strengths and weaknesses; for example, the American Thyroid Association 2015 guideline showed better diagnostic efficiency in evaluating TNs >1 cm, the TIRADS issued by the American College of Radiology in 2017 had more advantages in reducing unnecessary biopsy operations, and TNs were well diagnosed by radiologists according to C-TIRADS, achieving a higher performance (68). However, the diagnostic results were susceptible to operator dependency, probe, and US equipment variability. FNA is a comparatively accurate method for differentiating TNs preoperatively, but it was reported that approximately 20% of FNA samples obtained had ambiguous results (1517). AI not only solves the complex problem of the US risk stratification system but also reduces intra- and interobserver variability in US diagnosis (23, 30).

Nevertheless, most applications of DL in the diagnosis of TNs have been conducted based on single GSU imaging or single-view sections, limiting access to image information to a certain extent (21, 3133). In addition to GSU, radiologists also referred to CDFI and elastography for obtaining the blood flow and hardness information of TNs to assist the GSU diagnosis clinically and make a diagnosis after a comprehensive analysis. In terms of the statistical analysis in our study, a higher elastic score was markedly correlated with malignant TNs, confirming that malignant TNs tend to be hard. The differences in vascular distribution pattern and Adler grade were statistically significant in TNs, and malignant TNs tended to be less or lacked blood flow. In addition, many studies have verified the effectiveness of combined or multimodal US imaging in the differentiation of TNs visually (11, 13, 24, 25). Therefore, based on multimodal US images of TNs obtained from transverse and longitudinal sections, new DL models were used to distinguish benign from malignant TNs in our study.

In our study, the diagnostic performance of GSU (0.825) was comparable to that of previous studies (AUC of 0.788 and 0.829) (32, 33). However, the DL models using combined or multimodal US images achieved a better performance (0.858–0.928) than those using GSU alone (0.825) (p< 0.05). Notably, the AUC of the DL model based on GSU alone was also greatly improved after adding CDFI (0.825 vs. 0.909, p< 0.001). In a related study, Baig et al. quantified the regional blood flow indices of TNs, and the diagnostic accuracy of GSU features was increased from 58.6% to 79.3% when combined with CDFI (p< 0.05) (34). The DL models in our study provided consistent and repeatable results and outperformed conventional machine learning-based methods with a specificity of 86.9%, a PPV of 85.2%, and an accuracy of 83.8%. As for the improvement in diagnostic efficiency after adding CDFI, we found that it may be due to the attention mechanism algorithm applied in this study, which could obtain richer and more objective features that were previously unrecognized visually by learning the information of CDFI images autonomously. We also demonstrated that SE imaging helped improve the diagnosis of DL based on GSU (0.825 vs. 0.858, p< 0.05). However, the AUC of the DL model based on (G + E)US was markedly less than that based on (G + C)US (0.858 vs. 0.909, p = 0.001). Additionally, there were no significant differences between (G + C + E)US and (G + C)US (0.906 vs. 0.909, p = 0.294). Therefore, our study confirmed that CDFI played a more substantial role in distinguishing TNs than SE in our study, and the less obvious advantages of SE may be associated with the subjectivity of the collecting process of SE images.

Adding a Mask containing the contour information of the TNs was found to help improve the diagnostic performance of DL models based on GSU (from 0.825 to 0.881), (G + C)US (from 0.909 to 0.918), (G + E)US (from 0.858 to 0.889), and (G + C + E)US (from 0.906 to 0.928), indicating that effective delineation of the nodular boundaries in US images played an important role in characterizing TNs. The best AUC of 0.928 was achieved by DL using (G + C + E + M)US. The highest specificity (89.5%) and PPV (87.8%) were achieved by DL based on (G + C + E)US, which could play a primary role in avoiding overdiagnosis and helping reduce unnecessary biopsies for the diagnosis of TNs, whereas the highest sensitivity (86.9%) and NPV (87.7%) of great clinical significance for screening out malignant TNs and avoiding misdiagnosis were achieved by DL based on (G + C + M)US. In summary, the performance of DL models based on multimodal US imaging was superior to that based on a single GSU, which supports our assumption that multimodal US could provide more comprehensive and effective information for TN diagnosis.

In clinical practice, US diagnosis by radiologists cannot be completely replaced by AI technology, and a final diagnosis should be made by radiologists. Therefore, we compared the performance of the DL method for differentiating TNs with that of visual diagnosis by radiologists and further explored the auxiliary role of DL for radiologists’ diagnosis. Compared with the first diagnosis of TNs visually by junior radiologists, there was a significant improvement in the second diagnosis with DL assistance (0.720 vs. 0.796, p< 0.001), which could be comparable to that of the seniors in the first diagnosis (0.796 vs. 0.794, p > 0.05). Moreover, the DL method could also provide an auxiliary diagnosis for senior radiologists in terms of accuracy (from 80.6% to 83.4%), which was superior to DL based on GSU alone (83.4% vs. 78.9%, p = 0.041). It has been proven that DL can assist clinical radiologists in improving diagnostic ability and increasing confidence, especially for juniors with less experience. In a study by Peng et al. (35), the DL-assisted method also improved the AUC of radiologists in diagnosing TNs from 0.837 to 0.875 (p< 0.001).

Nevertheless, there were no significant diagnostic differences with and without DL assistance for senior radiologists (0.794 vs. 0.822, p = 0.141). It seems that DL-aided diagnosis was less effective for senior than junior radiologists, which may be related to the fact that senior radiologists were more likely to rely on their own clinical experience. Our analysis may also be due to the fact that, by using DL models, we sacrificed interpretability for robust and complex imaging features with greater generalizability. Furthermore, DL technology obtained results based on features that were learned and extracted independently rather than on predefined handcrafted features, where the process was abstract and incomprehensible, leading to distrust by the radiologists. To resolve the visualization of DL learning and decision processes, Kim et al. (36) applied Grad-CAM to generate output images overlaid with heat maps to achieve visual interpretability. Meanwhile, Zhou et al. (20) found that the adjacent parenchyma of TNs is critical for classification by visual interpretability of DL.

By comparing the radiologists’ and DL’s diagnostic efficacy, we found that senior radiologists with DL assistance only had a diagnosis comparable to the DL model based on GSU in terms of AUC (0.822 vs. 0.825, p = 0.512), which could not be compared with the diagnosis based on multimodal US imaging (0.822 vs. 0.858–0.928, p< 0.05), effectively demonstrating the excellent clinical value of the DL method, especially for multimodal US imaging, with potential for further development and application. Radiologists may be affected by fatigue and other factors in daily work, whereas AI can run on its own and has the characteristics of indefatigability with stable and high diagnostic efficiency.

Our study has several advantages. To our knowledge, this is the first study in which the attention mechanism-guided residual network was used to construct a variety of DL models based on different US imaging combinations. The objects in our study were TNs of the C-TIRADS 3–5 categories, which are more suitable for clinical diagnosis difficulties and extend the scope of clinical application. We have verified that DL models based on multimodal US can assist radiologists in improving diagnostic performance, especially for those with less experience, and postoperative pathological results were used as the gold standard for statistical analysis in this study, which was more objective than the studies using cytological pathological results (20, 31). Our case set included relatively more samples and achieved a balance between benign and malignant TNs, which could effectively reduce the diagnostic bias compared with a previous study (18).

This study had some limitations. First, the main limitation of our study was that the data were retrospectively derived from a single center, and additional external validation or multicenter studies are needed to refine our study. Second, the images in this study were static images stored in a compressed format, which may have led to some potential image features not being mined. Therefore, dynamic images or raw radiofrequency signals should be included in future studies. Third, the visualization of the DL proposed in this study was not achieved. Visualization of the DL process could be conducted to make the results more reliable in subsequent studies. More technologies could be included, such as shear wave elastography, superb microvascular imaging, and contrast-enhanced US.

In conclusion, the DL model based on multimodal US images can achieve a high diagnostic value in the differential diagnosis of benign and malignant TNs of C-TIRADS 3–5 categories, aid second-opinion provision, and improve the diagnostic ability for radiologists, which is of great significance for clinical decision-making.

Data availability statement

The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding authors.

Ethics statement

The studies involving human participants were reviewed and approved by the Ethics Committee of the Second Affiliated Hospital of Harbin Medical University. Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.

Author contributions

YY and JT conceived and designed the study. YT, QD, and HK collected the clinical and image data and performed image preprocessing. YT, LZ, and WY analyzed and evaluated the ultrasonographic features of thyroid nodules. YY and WQ provided a deep learning algorithm and built the models. YY analyzed the image data and performed statistical analysis. YT, YY, and TW wrote the manuscript. JT, XX, and XL reviewed and edited the manuscript. All authors have contributed to the manuscript and approved the submitted version.

Funding

This work was supported by the National Key R&D Program of China (2018YFA0701400), the Shenzhen Universities Stabilization Support Program (20200812144239001), the National Nature Science Foundation Grants of China (12174267, 62022086, 11874382, and 81827802), the Shenzhen Foundation (JCYJ20200109114237902, SGDX2020110309400200, and ZDSYS201802061806314), CAS research projects (KFJ-PTXM-012 and 2011DP173015), the Natural Science Foundation of Guangdong Province (2022A1515011343, 2020B1111130002, and 2020B1212060051), and the Youth Innovation Promotion Association (CAS 2018391).

Acknowledgments

We would like to thank the ultrasound doctors in the Department of Ultrasound, The Second Affiliated Hospital of Harbin Medical University, Harbin, for participating in the diagnosis of thyroid nodules.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2022.1012724/full#supplementary-material

References

1. Haugen BR, Alexander EK, Bible KC, Doherty GM, Mandel SJ, Nikiforov YE, et al. American Thyroid association management guidelines for adult patients with thyroid nodules and differentiated thyroid cancer the American thyroid association guidelines task force on thyroid nodules and differentiated thyroid cancer. Thyroid (2016) 26:1–133. doi: 10.1089/thy.2015.0020

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Chen W, Zheng R, Baade PD, Zhang S, Zeng H, Bray F, et al. Cancer statistics in China, 2015. CA Cancer J Clin (2016) 66:115–32. doi: 10.3322/caac.21338

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Pacini F, Castagna MG, Brilli L, Pentheroudakis G, Grp EGW. Thyroid cancer: ESMO clinical practice guidelines for diagnosis, treatment and follow-up. Ann Oncol (2012) 23:110–9. doi: 10.1093/annonc/mds230

CrossRef Full Text | Google Scholar

4. Kwak JY, Han KH, Yoon JH, Moon HJ, Son EJ, Park SH, et al. Thyroid imaging reporting and data system for US features of nodules: A step in establishing better stratification of cancer risk. Radiology (2011) 260:892–9. doi: 10.1148/radiol.11110206

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Tessler FN, Middleton WD, Grant EG, Hoang JK, Berland LL, Teefey SA, et al. ACR thyroid imaging, reporting and data system (TI-RADS): White paper of the ACR TI-RADS committee. J Am Coll Radiol (2017) 14:587–95. doi: 10.1016/j.jacr.2017.01.046

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Zhou J, Song Y, Zhan W, Wei X, Zhang S, Zhang R, et al. Thyroid imaging reporting and data system (TIRADS) for ultrasound features of nodules: Multicentric retrospective study in China. Endocrine (2021) 72:157–70. doi: 10.1007/s12020-020-02442-x

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Liu H, Ma AL, Zhou YS, Yang DH, Ruan JL, Liu XD, et al. Variability in the interpretation of grey-scale ultrasound features in assessing thyroid nodules: A systematic review and meta-analysis. Eur J Radiol (2020) 129:109050. doi: 10.1016/j.ejrad.2020.109050

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Choi SH, Kim EK, Kwak JY, Kim MJ, Son EJ. Interobserver and intraobserver variations in ultrasound assessment of thyroid nodules. Thyroid (2010) 20:167–72. doi: 10.1089/thy.2008.0354

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Park CS, Kim SH, Jung SL, Kang BJ, Kim JY, Choi JJ, et al. Observer variability in the sonographic evaluation of thyroid nodules. J Clin Ultrasound (2010) 38:287–93. doi: 10.1002/jcu.20689

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Lee HJ, Yoon DY, Seo YL, Kim JH, Baek S, Lim KJ, et al. Intraobserver and interobserver variability in ultrasound measurements of thyroid nodules. J Ultrasound Med (2018) 37:173–8. doi: 10.1002/jum.14316

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Chaigneau E, Russ G, Royer B, Bigorgne C, Bienvenu-Perrard M, Rouxel A, et al. TIRADS score is of limited clinical value for risk stratification of indeterminate cytological results. Eur J Endocrinol (2018) 179:13–20. doi: 10.1530/EJE-18-0078

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Seo H, Na DG, Kim JH, Kim KW, Yoon JW. Ultrasound-based risk stratification for malignancy in thyroid nodules: A four-tier categorization system. Eur Radiol (2015) 25:2153–62. doi: 10.1007/s00330-015-3621-7

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Pei SF, Cong SZ, Zhang B, Liang CH, Zhang L, Liu JJ, et al. Diagnostic value of multimodal ultrasound imaging in differentiating benign and malignant TI-RADS category 4 nodules. Int J Clin Oncol (2019) 24:632–9. doi: 10.1007/s10147-019-01397-y

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Ha EJ, Na DG, Baek JH, Sung JY, Kim JH, Kang SY. US Fine-needle aspiration biopsy for thyroid malignancy: Diagnostic performance of seven society guidelines applied to 2000 thyroid nodules. Radiology (2018) 287:893–900. doi: 10.1148/radiol.2018171074

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Singh Ospina N, Brito JP, Maraka S, Espinosa de Ycaza AE, Rodriguez-Gutierrez R, Gionfriddo MR, et al. Diagnostic accuracy of ultrasound-guided fine needle aspiration biopsy for thyroid malignancy: Systematic review and meta-analysis. Endocrine (2016) 53:651–61. doi: 10.1007/s12020-016-0921-x

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Theoharis CGA, Schofield KM, Hammers L, Udelsman R, Chhieng DC. The Bethesda thyroid fine-needle aspiration classification system: Year 1 at an academic institution. Thyroid (2009) 19:1215–23. doi: 10.1089/thy.2009.0155

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Mathur A, Weng J, Moses W, Steinberg SM, Rahbari R, Kitano M, et al. A prospective study evaluating the accuracy of using combined clinical factors and candidate diagnostic markers to refine the accuracy of thyroid fine needle aspiration biopsy. Surgery (2010) 148:1170–6. doi: 10.1016/j.surg.2010.09.025

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Liu Z, Zhong SB, Liu Q, Xie CX, Dai YZ, Peng C, et al. Thyroid nodule recognition using a joint convolutional neural network with information fusion of ultrasound images and radiofrequency data. Eur Radiol (2021) 31:5001–11. doi: 10.1007/s00330-020-07585-z

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Wang K, Lu X, Zhou H, Gao Y, Zheng J, Tong M, et al. Deep learning radiomics of shear wave elastography significantly improved diagnostic performance for assessing liver fibrosis in chronic hepatitis b: A prospective multicentre study. Gut (2019) 68:729–41. doi: 10.1136/gutjnl-2018-316204

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Zhou H, Jin YH, Dai L, Zhang MW, Qiu YQ, Wang K, et al. Differential diagnosis of benign and malignant thyroid nodules using deep learning radiomics of thyroid ultrasound images. Eur J Radiol (2020) 127:108992. doi: 10.1016/j.ejrad.2020.108992

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Kwon SW, Choi IJ, Kang JY, Jang WI, Lee GH, Lee MC. Ultrasonographic thyroid nodule classification using a deep convolutional neural network with surgical pathology. J Digit Imaging (2020) 33:1202–8. doi: 10.1007/s10278-020-00362-w

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Akkus Z, Cai J, Boonrod A, Zeinoddini A, Weston AD, Philbrick KA, et al. A survey of deep-learning applications in ultrasound: Artificial intelligence-powered ultrasound for improving clinical workflow. J Am Coll Radiol (2019) 16:1318–28. doi: 10.1016/j.jacr.2019.06.004

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Sorrenti S, Dolcetti V, Radzina M, Bellini MI, Frezza F, Munir K, et al. Artificial intelligence for thyroid nodule characterization: Where are we standing? Cancers (2022) 14(14). doi: 10.3390/cancers14143357

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Colakoglu B, Yildirim D, Alis D, Ucar G, Samanci C, Ustabasioglu FE, et al. Elastography in distinguishing benign from malignant thyroid nodules. J Clin Imag Sci (2016) 6:51. doi: 10.4103/2156-7514.197074

CrossRef Full Text | Google Scholar

25. Sorrenti S, Dolcetti V, Fresilli D, Del Gaudio G, Pacini P, Huang PT, et al. The role of CEUS in the evaluation of thyroid cancer: From diagnosis to local staging. J Clin Med (2021) 10(19). doi: 10.3390/jcm10194559

CrossRef Full Text | Google Scholar

26. Adler DD, Carson PL, Rubin JM, Quinn-Reid D. Doppler Ultrasound color flow imaging in the study of breast cancer: Preliminary findings. Ultrasound Med Biol (1990) 16:553–9. doi: 10.1016/0301-5629(90)90020-d

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Asteria C, Giovanardi A, Pizzocaro A, Cozzaglio L, Morabito A, Somalvico F, et al. US-Elastography in the differential diagnosis of benign and malignant thyroid nodules. Thyroid (2008) 18:523–31. doi: 10.1089/thy.2007.0323

PubMed Abstract | CrossRef Full Text | Google Scholar

28. He KM, Zhang XY, Ren SQ, Sun J, Ieee. (2016). Deep residual learning for image recognition, in: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR) (Seattle, WA) 27-30. doi: 10.1109/cvpr.2016.90

CrossRef Full Text | Google Scholar

29. Woo SH, Park J, Lee JY, Kweon IS eds. (2018). CBAM: convolutional block attention module, in: Lecture Notes in Computer Science. (Munich, Germany: 15th Eur Conference on Computer Vision (ECCV)) (2018):3–19. doi: 10.1007/978-3-030-01234-2_1

CrossRef Full Text | Google Scholar

30. Ha EJ, Baek JH, Na DG. Risk stratification of thyroid nodules on ultrasonography: Current status and perspectives. Thyroid (2017) 27:1463–8. doi: 10.1089/thy.2016.0654

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Li XC, Zhang S, Zhang Q, Wei X, Pan Y, Zhao J, et al. Diagnosis of thyroid cancer using deep convolutional neural network models applied to sonographic images: A retrospective, multicohort, diagnostic study. Lancet Oncol (2019) 20:193–201. doi: 10.1016/S1470-2045(18)30762-9

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Wu GG, Lv WZ, Yin R, Xu JW, Yan YJ, Chen RX, et al. Deep learning based on ACR TI-RADS can improve the differential diagnosis of thyroid nodules. Front Oncol (2021) 11:575166. doi: 10.3389/fonc.2021.575166

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Zhang YC, Wu Q, Chen YT, Wang Y. A clinical assessment of an ultrasound computer-aided diagnosis system in differentiating thyroid nodules with radiologists of different diagnostic experience. Front Oncol (2020) 10:557169. doi: 10.3389/fonc.2020.557169

PubMed Abstract | CrossRef Full Text | Google Scholar

34. Baig FN, van Lunenburg JTJV, Liu SYW, Yip SP, Law HKW, Ying M. Computer-aided assessment of regional vascularity of thyroid nodules for prediction of malignancy. Sci Rep (2017) 7:14350. doi: 10.1038/s41598-017-14432-7

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Peng S, Liu YH, Lv WM, Liu LZ, Zhou Q, Yang H, et al. Deep learning-based artificial intelligence model to assist thyroid nodule diagnosis and management: A multicentre diagnostic study. Lancet Digit Health (2021) 3:e250–9. doi: 10.1016/S2589-7500(21)00041-8

PubMed Abstract | CrossRef Full Text | Google Scholar

36. Kim YJ, Choi Y, Hur SJ, Park KS, Kim HJ, Seo M, et al. Deep convolutional neural network for classification of thyroid nodules on ultrasound: Comparison of the diagnostic performance with that of radiologists. Eur J Radiol (2022) 152:110335. doi: 10.1016/j.ejrad.2022.110335

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: thyroid nodule, deep learning, multimodal, ultrasound, diagnosis

Citation: Tao Y, Yu Y, Wu T, Xu X, Dai Q, Kong H, Zhang L, Yu W, Leng X, Qiu W and Tian J (2022) Deep learning for the diagnosis of suspicious thyroid nodules based on multimodal ultrasound images. Front. Oncol. 12:1012724. doi: 10.3389/fonc.2022.1012724

Received: 05 August 2022; Accepted: 18 October 2022;
Published: 08 November 2022.

Edited by:

Lorenz Kadletz-Wanke, Medical University of Vienna, Austria

Reviewed by:

Salvatore Sorrenti, Department of Surgical Sciences, Sapienza University of Rome, Italy
Fajin Dong, Jinan University, China

Copyright © 2022 Tao, Yu, Wu, Xu, Dai, Kong, Zhang, Yu, Leng, Qiu and Tian. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Jiawei Tian, jwtian2004@163.com; Yanyan Yu, yy.yu@szu.edu.cn

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.