- 1Cancer Center and Department of Breast and Thyroid Surgery, Xiang’an Hospital, School of Medicine, Xiamen University, Xiamen, China
- 2Fujian Key Laboratory of Precision Diagnosis and Treatment in Breast Cancer, Xiamen, China
- 3Xiamen Key Laboratory of Endocrine-Related Cancer Precision Medicine, Xiamen, China
- 4Xiamen Research Center of Clinical Medicine in Breast and Thyroid Cancers, Xiang’an Hospital of Xiamen University, School of Medicine, Xiamen University, Xiamen, China
- 5Department of Ultrasonic Medicine Affiliated Hospital of Xizang Minzu University, Xianyang, China
- 6National Institute for Data Science in Health and Medicine, Xiamen University, Xiamen, China
- 7Department of Ultrasonic Medicine, Xiang’an Hospital, School of Medicine, Xiamen University, Xiamen, China
- 8Department of Ultrasonic Medicine of Shantou Central Hospital, Shantou, China
- 9School of Informatics, Xiamen University, Xiamen, China
Objective: To develop a deep learning (DL) model for predicting axillary lymph node (ALN) metastasis using dynamic ultrasound (US) videos in breast cancer patients.
Methods: A total of 271 US videos from 271 early breast cancer patients collected from Xiang’an Hospital of Xiamen University andShantou Central Hospitabetween September 2019 and June 2021 were used as the training, validation, and internal testing set (testing set A). Additionally, an independent dataset of 49 US videos from 49 patients with breast cancer, collected from Shanghai 10th Hospital of Tongji University from July 2021 to May 2022, was used as an external testing set (testing set B). All ALN metastases were confirmed using pathological examination. Three different convolutional neural networks (CNNs) with R2 + 1D, TIN, and ResNet-3D architectures were used to build the models. The performance of the US video DL models was compared with that of US static image DL models and axillary US examination performed by ultra-sonographers. The performances of the DL models and ultra-sonographers were evaluated based on accuracy, sensitivity, specificity, and area under the receiver operating characteristic curve (AUC). Additionally, gradient class activation mapping (Grad-CAM) technology was also used to enhance the interpretability of the models.
Results: Among the three US video DL models, TIN showed the best performance, achieving an AUC of 0.914 (95% CI: 0.843-0.985) in predicting ALN metastasis in testing set A. The model achieved an accuracy of 85.25% (52/61), with a sensitivity of 76.19% (16/21) and a specificity of 90.00% (36/40). The AUC of the US video DL model was superior to that of the US static image DL model (0.856, 95% CI: 0.753-0.959, P<0.05). The Grad-CAM technology confirmed the heatmap of the model, which highlighted important subregions of the keyframe for ultra-sonographers’ review.
Conclusion: A feasible and improved DL model to predict ALN metastasis from breast cancer US video images was developed. The DL model in this study with reliable interpretability would provide an early diagnostic strategy for the appropriate management of axillary in the early breast cancer patients.
Introduction
Breast cancer is the most commonly diagnosed cancer among female malignancies worldwide and has become the leading cause of cancer-related deaths (1). The involvement of axillary lymph nodes (ALNs) is a crucial prognostic factor for breast cancer patients, and accurate axillary staging is critical for evaluating the disease status. Once ALN metastasis occurs, the clinical stage and treatment regimen may change, which can significantly affect patient prognosis (2, 3). However, the diagnosis of metastatic ALNs from imaging can vary greatly depending on the imaging results of different-level physicians/radiologists. Currently, the standard procedure for diagnosing ALN status before surgery is pathological examination of the lymph node obtained through biopsy. Clearly, lymph node biopsy is an invasive procedure with relatively complicated operation and is time-consuming (4–6). Lymph nodes biopsy can probably cause complications such as soft tissue infection, bleeding, seroma. Occasionally, it will cause the subcutaneous effusion due to lymphatic fistula. In addition, the technique has a false-negative rate ranging from 7.8% to 27.3% (7–9).
General imaging studies a variety of non-invasive diagnostic tools, including ultrasound (US) (10), magnetic resonance imaging (MRI) (11), and mammography (12). Asian women, especially young women, have denser breast tissue (13), and US has been shown to be more suitable for the detection of breast lesions. Thus, US is widely used as the first choice for breast disease screening due to its convenience, non-invasiveness, real-time capabilities, absence of radiation, and flexibility in performing ultrasound-guided biopsies (14, 15). Usually, the US diagnosis of ALN is based on the size or shape of the lymph node and the status of the lymphatic gate, which can lead to significant variability due to the subjectivity of examination and the interpretation skills of ultra-sonographers (16). Previously, our group combined US findings and clinico-pathological factors and accurately predicted probability of ALN metastases. The model was further validated in a Dutch population with predictive probability less than 12%. However, this approach is still based on histopathological information including histological grade, hormone receptor and Her2 status obtained from either core needle biopsy or excision. Therefore, a new noninvasive method for rapid, accurate, and objective evaluation of ALN metastasis in breast cancer is urgently needed.
In recent years, artificial intelligence (AI) has been gradually applied in breast imaging to improve workflows, perform automatic image segmentation, enable intelligent diagnosis, and predict ALN metastasis accurately (17–19). To avoid the complicated feature extraction process and extract more abundant information from the image, the deep learning (DL) method has been widely used in medical image research in recent years (20, 21). DL transforms the original image data into a higher-level, more abstract expression through a hierarchical network and replaces the complex process of manual extraction and feature design by obtaining hierarchical feature information.
Current prediction models for ALN metastasis are based on static two-dimensional ultrasound images. Sun et al. manually delineated regions of interest (ROIs) on breast ultrasound images and constructed deep learning models based on intratumor, peritumor, and peritumor plus peritumor using the DenseNet network (22). The results showed that the three deep learning models had better prediction performance than their corresponding radiomics models. Additionally, peri-tumor microenvironment information can improve the accuracy of the ALN prediction model performance.
A recent study reported the effectiveness of DL models using three different networks to predict ALN metastasis in multicenter breast cancer patients. The results showed that the diagnostic ability of the DL model was also better than that of ultra-sonographers (23). In another study, Zheng et al. used deep learning radiomics of conventional ultrasound to extract depth features from static images and shear wave elastography images and established a deep learning image group model to predict ALN metastasis (24). The diagnostic performance in predicting ALN status between disease-free axilla and any axillary metastasis with an AUC of 0.902 (95% CI: 0.843-0.961). Therefore, deep learning has broad application prospects in predicting the status of ALN metastasis in breast cancer.
Previous studies were limited to analyzing ultrasound static (single frame) images instead of sequential image acquisition of video, which can cause the loss of many subtle or important US lesion features in the keyframe, and even some subtle lesions may be omitted (25). Additionally, hyperplastic glandular tissues or ribs may be mistaken for masses, which can mislead the deep learning model and result in misjudgment. Therefore, the aim of this study is to construct a non-invasive, rapid, accurate, and objective prediction model of ALN metastasis in breast cancer patients based on ultrasound video images. This model can guide the selection of clinical tumor staging and surgical methods, thereby avoiding unnecessary ALN biopsies in breast cancer patients. This study aims to develop a DL model using videos that demonstrate diagnostic performance comparable to that of experienced ultra-sonographers. Unlike previous studies that relied on static images, our model leverages the dynamic nature of video to capture more subtle features of the breast and surrounding tissues. By incorporating video data, we aim to improve the accuracy of ALN metastasis prediction and reduce the need for invasive procedures such as biopsy.
Materials and methods
Patients
This retrospective multicenter diagnostic study was approved by the institutional review board of Xiang’an Hospital of Xiamen University and Shantou Central Hospital, informed consent was obtained from all participants (XAHLL2022060(2022) Research 041). Breast cancer US video data were acquired from 421 patients from Shantou Central Hospital, Guangdong, China, and Xiang’an Hospital of Xiamen University, Fujian, China, from September 2019 to June 2021. The inclusion criteria for the study were as follows: (a) patients with breast cancer confirmed by pathology; (b) clinical stage of T1-T2 and no distant metastasis; (c) complete US videos of the breast lesions were obtained; (d) patients without malignant tumors other than breast cancer; and (e) no history of axillary surgery. The exclusion criteria were (a) ALN involvement due to diseases other than breast cancer and (b) treatment for breast cancer with hormonal therapy, chemotherapy, or radiation therapy before surgery. The scanning and ultrasound diagnosis of the lesions was performed by ultra-sonographers experienced in breast ultrasound diagnosis by using the cross-plane screening method. Imaging data for the longitudinal and transverse views of each lesion were stored in the ultrasound equipment and then transferred to a hard disk for storage in a picture archiving and communications system (PACS). A flowchart describing the research process is shown in Figure 1.
Figure 1 Patient recruitment workflow. In total, 271 US videos of 271 patients were included according to the inclusion criteria. The included patients were examined by conventional US and had complete clinical information for this study.
Finally, 210 breast cancer US videos of 210 patients (121 without ALN metastasis and 89 with ALN metastasis) were used as the training and validation set, and the other 61 US videos of patients (34 without ALN metastasis and 27 with ALN metastasis) were used as testing set A. Another independent testing set B consisted of 49 breast cancer US videos of patients (38 without ALN metastasis and 11 with ALN metastasis) from Shanghai’s tenth Hospital of Tongji University, China, from July 2021 to April 2022.
Data analysis
Ultrasound examinations were performed by five ultra-sonographers (two from Xiang’an Hospital of Xiamen University and three from Shantou Central Hospital) who had 8-20 years of experience in breast ultrasound. Ultrasound systems (GE Healthcare, USA; Siemens, Munich, Germany; Philips, Amsterdam, the Netherlands) were used to generate the ultrasound images. Image quality control was done for all videos, the L12-5 linear probe at a frequency of 12 MHz, the ultrasound images were adjusted by gain, focus, and zoom as needed. We used B-mode to examine the breast lesions and obtained transverse or longitudinal images. After performing 2D US, the videos and some single images with the fewest artifacts and the best quality was chosen and stored image for analysis and all breast US images extracted from systems were converted into DICOM format. Women with breast disorders were referred for breast ultrasound (BUS) and those with Breast Imaging Reporting And Data System, BI-RADS) ≥4 had ultrasound-guided biopsy and surgical resection.
Deep learning model
All the clinical and pathologic data were obtained from the medical records. Pathologic results of the breast cancer included tumor clinical stage, tumor location, pathologic results of ALN status and tumor type. Clinical data included patients age, tumor size, BI-RADS category and pathologic results of ALN status with and without metastasis were recorded.
The original breast ultrasound data obtained from the clinic was relatively rough and there was a large amount of redundant information such as equipment parameters and other information around the imaging position. A series of pre-processing steps were adopted to normalize the data before training models. The steps were as follows: 1. Only the intermediate lesion position of the images was retained by cropping and excluding marginal irrelevant information and patient text data; 2. After adjusting the video to 256×256 pixels, various complex sampling situations are simulated through random horizontal flipping, elastic transformation, random cropping and other data enhancement methods to enhance the data variety and model generalization; 3. Scaled the video to 224×224 while the pixel values were normalized to 0 ~ 1. Through a series of data augmentation methods, the possibility that the input of the network was the same as the image seen by the network before each training was significantly reduced, and the diversity of data samples was further increased.
Interpretability of DL model
At present, CNNs are the most well-known type of DL in medical image analysis (26). Our DL model mainly consisted of three modules including feature extraction, feature fusion, and final classification prediction, using the classical CNN network ResNet-3D as the backbone (27), as shown in Figure 2. The basic network contained five stages. The first stage was the preprocessing stage for the input, including a simple convolutional layer and a max-pooling layer, and the subsequent four stages were composed of multiple residual blocks to realize the core residual structure in ResNet. In the Feature Fusion Module, unique interlaced offsets were provided for different frames to fuse spatiotemporal information (28). In the training, we utilized the widely adopted CrossEntropy Loss as the loss function in this study, given its extensive application in various classification tasks across the field. By employing the CrossEntropy Loss, which has become a standard choice in such scenarios, we aimed to effectively measure the disparity between predicted class probabilities and the true labels, thus facilitating accurate model optimization.
Figure 2 The overview of our DL model architecture. It mainly consists of three modules including feature extraction, feature fusion, and final classification prediction.
where p represents the true label and q represents the predict value.
The network and code of this study were written and implemented in Python using Pytorch. All deep learning algorithms were trained in the same training setting and trained on the Linux Server Ubuntu 18.04 LTS platform, GeForce RTX 3080 Ti GPU. During the training process, the network weights were randomly initialized and trained by the Adam optimizer, and the initial learning rate was set to 1.0×10−4. We also trained the model on static images for comparison, where the training process model was trained on static images, using the same testing set (testing set A: 61 patients) as the dynamic videos at inference time. 5-fold cross-validation was made to make the experiment results more reliable and accurate. The sample length T (Time channel) was set to 16 and the batch size was set to 4, for a total of 300 training epochs. We selected three CNN models of R2 + 1D (29), TIN (30) and ResNet-3D architectures for comparison and analysis. For static image training, we used three representative models ResNet, Inception V3, and VGG. The trained DL models outputted the predicted probability of the presence of lymph node metastasis according to US videos and chose the class of the highest probability as the prediction result. The results were measured based on the difference between the predicted output of the best model trained by five-fold cross-validation and the true labels of the testing set, and the corresponding accuracy, sensitivity, specificity, ROC curves, and PR curves were calculated. In addition, to better compare the performance of the models, we also compared the results of dynamic videos with static images. Temporal Interlace Network (TIN) presents a simple but effective module for video classification. TIN does not learn temporal features, but fuses spatio-temporal information by interleaving spatial features from past to future and future to past. A differentiable sub-module can calculate the offset of the interleaved features in the time dimension, and can rearrange the features according to the offset, so that each group of features is displaced by a different distance in the time dimension. Thus, 3D convolution is replaced by convenient and fast feature displacement operation to realize information exchange between adjacent frames. This makes the number of parameters and computation of the network much lower than that of ordinary 3D convolutional networks, making the network as a whole quite lightweight.
Clinical application of DL model
To obtain ultra-sonographers’ predictive performance on the independent testing set, we performed an independent evaluation in testing set B. Three ultra-sonographers of different levels of experience in breast ultrasound were trained in how to perform predictive analysis based on some typical characteristics, such as the size of the lesions and the presence of lymphatic gate invasion (31), calcifications (32), architectural distortions (33), margin, cortical morphologic features, cortical medulla thickness, and blood flow (34, 35). The clinical real-time workflow consisted of three components. First, the ultra-sonographers were blinded to the pathological information of the axillary lymph node while they were scanning the axilla. Second, they analyzed some typical characteristics of breast cancer US images with the use of the American College of Radiology Breast Imaging Reporting and Data System (BI-RADS) (36). Third, the ultra-sonographers conducted a reexamination of the patient’s US video. They utilized the trained DL model to directly input the test image into the model, which provided an immediate output indicating the metastatic status of the ALNs. Finally, we compared the performance of ultra-sonographers with the DL model.
Model interpretation
In this study, we utilized Grad-CAM technology to improve the interpretability of our models, which was crucial for clinical applications. To generate the heatmap, we evaluated sites of interest for subsequent clinical examination. For each video and static image, we fed them into the fully trained model and obtained the feature map of the final convolutional layer. Then, we calculated the heatmap by running Grad-CAM on this feature map. We mapped the resulting heatmaps to visualize the areas of the keyframe of videos that were most indicative of lesions of metastasis. As show in Figure 3 the best TIN model focuses more on the location of the nodules than the other two models and the characteristics of the nodules are the most important basis for predicting whether the nodules will metastasize. The technology and details are presented in the Supplementary Information section.
Figure 3 The visual class activation diagram of the prediction results of dynamic video models. The brighter the color, the better the visibility of the model. The best TIN model focuses more on the location of the nodules than the other two models, and the characteristics of the nodules are the most important basis for predicting whether the nodules will metastasize.
Statistical analysis
Differences between variable groups were analyzed using the Mann-Whitney U test. The chi-square test was used to compare rates between different groups. The performances of the three algorithms were evaluated by AUC, as well as the accuracy (ACC), sensitivity (SEN), specificity (SPE), positive predictive value (PPV), and negative predictive value (NPV). Intraobserver and interobserver agreements were compared using the Kappa value. All statistical analyses were two-sided, and P< 0.05 was considered statistically significant. We performed all statistical analyses using MedCalc (v.20.0.26; https://www.medcalc.org, Copyright MedCalc Software Ltd), Python software (v.3.8.5150.0; http://www.python.org), and SPSS software for Windows (v.20.0).
Results
Patient population and clinical characteristics
We studied a total of 421 breast lesions and finally enrolled 271 women with 271 malignant breast lesions for analysis. The average patient age was 52.90 years (range, 26–87 years) for the training and validation set and 53.31 years (range, 26–78 years) for testing set A. Among them, 9 (3.32%) patients had noninvasive carcinoma, 242 (89.3%) patients had invasive ductal carcinomas, and 9 (3.32%) had invasive lobular carcinomas. The mean tumor size was 0.23 cm (0.21-0.24 cm) for the training and validation set and 0.20 cm (0.18 cm-0.22 cm) for testing set A. Of the 271 patients, 133 (49.08%) had T1-stage tumors, and 138 (50.92%) had T2-stage tumors. The ALN of 116 (42.80%) patients was metastasized, and 155 (57.20%) patients had no lymph node metastasis. A detailed summary of patient demographics and tumor characteristics is provided in Table 1.
DL model-based US videos showed better performance than DL model-based static images
We used three deep learning networks, TIN, R2 + 1D, and ResNet3D, to build the prediction model based on US videos. The change curve of the loss function during the training of the dynamic video model is shown in labels "A-D" of Figure 4. Additionally, we also built the prediction model based on US static images using ResNet50, InceptionV3, and VGG19 networks. Among the three US video DL models, the TIN network showed the best performance, with an accuracy, sensitivity, and specificity of 85.25% (52/61), 76.19% (16/21), and 90.00% (36/40), respectively. The AUC of the TIN model was better than that of the ResNet50 model (0.914, 95% CI: 0.843-0.985 vs. 0.856, 95% CI: 0.753-0.959, P< 0.05), as shown in Figure 5. The ResNet50 network showed the best performance among the US static image DL models, with an accuracy, sensitivity, and specificity of 80.33% (49/61), 66.67% (14/21), and 87.50% (35/40), respectively. The ResNet network is widely recognized for its extensive adoption as a classical model, primarily due to its incorporation of residual connections. These connections play a vital role in facilitating the smooth propagation of gradients and enhancing information flow across deeper layers. Consequently, this effectively addresses the issue of vanishing gradients. By leveraging this advantage, ResNet50 is capable of training deeper networks while maintaining performance levels, thereby capturing intricate features and achieving enhanced accuracy. In general the training of static images ignores the temporal correlation information between image frames, which is crucial for the prediction of whether the lymph node is metastatic or not, resulting in worse training results for the static image model. Detailed performance of the US video DL model and US static image model are provided in Table 2.
Figure 4 The figure shows the change curve of the loss function during the training of the dynamic video model. Subfigure (A, B) are Acc and loss curve for 300 training epoches. Subfigure (C, D) are Acc and loss curve for 300 validating epoches. TIN and ResNet converged well and stabilized at a small value of 0.2-0.3 in 300 training rounds. R2 + 1D converges relatively, but the trend is not obvious enough and the convergence value is about 0.6, which is a big gap. This also explains why its results on the test set are significantly lower than those of the other two models.
Figure 5 Comparison of receiver operating characteristic (ROC) curves between two models (video vs. static image) for predicting ALN metastasis.
DL model-based US videos achieved equal performance compared with ultra-sonographers
The performance of the US video DL model (TIN) in predicting lymph node metastasis was compared with that of three breasts ultra-sonographers with varying levels of experience using the pathological reference standard. For independent testing set B, consisting of 49 lesions, the senior ultrasonographer had an accuracy of 79.59% (39/49), the sensitivity of 63.64% (7/11), and specificity of 84.21% (32/38), while the DL model had an accuracy of 83.67% (41/49), the sensitivity of 54.55% (6/11), and specificity of 92.11% (35/38). As shown in Table 3, the ROC curves for the DL model and the ultra-sonographers, as seen in Figure 6, demonstrated that the ultra-sonographers achieved AUCs of 0.609 to 0.726 based on image classification alone, while the AUC of the DL model (TIN) was 0.773 (95% CI: 0.630-0.915, P<0.05), which was equal to or slightly higher than that of the ultra-sonographers with different levels of experience. However, there was no statistically significant difference between the DL model and ultra-sonographers. The external test set exhibits a distinct data distribution from the original test set, while test set B has a smaller data volume in comparison to the initial test set. As a result, there is a slight decrease in performance when evaluating test set B as compared to test set A.
Figure 6 Comparison of receiver operating characteristic (ROC) curves between the ultra-sonographers and DL model.
Table 3 The performance of the DL models for predicting lymph node metastasis compared with the ultrasonographers.
Model interpretation
The heatmaps generated by the US video DL model and US static image model are shown in Figure 7. The green and blue backgrounds represent the low forecast range, while the darker the feature color, the higher the attention of the model. This indicates that the trained model accurately identifies the areas that contribute the most to the final prediction result, which is also confirmed by clinicians. The focus heatmap continues to highlight the key elements in the current image, demonstrating that the characteristics of malignant lesions have been well studied and serve as the basis for LNM classification. These heatmaps have great clinical value in guiding subsequent clinical examination and treatment planning.
Figure 7 The visual activation diagram of the prediction results of the static image model. The brighter the place, the higher the model’s attention. It can be seen that all models basically pay attention to the characteristics of the nodule area, but they are not concentrated enough. Among them, Inception V3 and VGG19 models focus on the left side of the nodule, which also affects the final prediction results to some extent.
Discussion
For the past three decades, sentinel lymph node biopsy (SLNB) has been established as a standard of care for clinical node-negative breast cancer patients, allowing for the avoidance of axillary lymph node dissection (ALND) when SLNs are not involved with metastases (37). Moreover, ALND can also be omitted in selected patients with T1 or T2 breast cancer and fewer than two positive sentinel lymph nodes who undergo breast-conserving surgery (38). However, accurately predicting the metastatic status of ALN pre-operatively and/or intra-operatively remains challenging, despite numerous studies evaluating the risk of LN metastases. Therefore, there is a critical need to accurately predict the metastatic status of ALNs in a non-invasive manner. In this study, we aimed to utilize DL models with US video images in combination with clinicopathological factors to achieve the goal of predicting lymph node metastasis.
US has shown great promise in the assessment of ALN status in patients with early breast cancer, however, with the features of unclear margins, irregular shapes, or loss of fatty hilum, only visualized nodes can be analyzed and most of negative ALN with micro metastasis or without suspicious imaging features are missed (5). The accurate detection of lymph node micro metastasis is essential for guiding surgical decision making, and adjuvant therapy. Recent advances in AI models, such as DL technology, have been applied to breast US imaging (18, 39), and some studies have reported the effectiveness of DL models using a convolutional neural network (CNN) for the prediction of clinical ALN metastasis, with AUCs ranging from 0.72 to 0.89 (23, 40). However, these studies were based solely on ultrasound static (single frame) images, which may have resulted in the loss of many subtle or important lesion features and even caused the neglect of small lesions, as reported previously. In contrast, our study analyzed US videos containing multiple frames of images from the same patient while also considering the correlation between different frames. For video processing, additional temporal convolutions are employed. These convolutions help capture motions and temporal dependencies across consecutive frames. They typically involve 3D convolutions, where filters have both spatial and temporal dimensions. These additional operations enable deep learning models to capture temporal patterns, motion information, and long-term dependencies (41, 42). By combining the features between frames, the model to capture temporal relationships, dependencies, and patterns that may not be apparent when considering individual frames in isolation. As a result, the inclusion of temporal information through feature fusion enhances the overall performance of video-based tasks.
We demonstrated that our DL model achieved better performance than static images previously studied, resulting in higher accuracy (85.25% vs. 80.33%) and specificity (90.00% vs. 87.50%). Furthermore, compared to senior ultra-sonographers in testing set B, the DL model also showed higher accuracy (83.67% vs. 79.59%) and specificity (92.11% vs. 84.21%). These significant improvements suggest that this model may be useful in a clinical setting for predicting lymph node metastasis, potentially assisting ultra-sonographers in making decisions regarding appropriate axillary rescanning. This DL model might have great potential to serve as a noninvasively imaging biomarker to replace SLND and/or ALND for patients with early-stage breast cancer. Especially, the model could assist breast surgeons to make clinical decisions for appropriate axillary management, e.g., omission of SLN biopsy in the node negative patients.
To the best of our knowledge, our study is the first to demonstrate the effectiveness of DL model in predicting ALN metastases using ultrasound (US) video images. While previous studies have used deep learning techniques for lymph node classification, they mainly focused on magnetic resonance imaging (MRI) images and CT (40, 43). Yu et al (44). developed an effective preoperative MRI radiomics evaluation methodology for ALN status in patients with early-stage invasive breast cancer, utilizing machine learning methods. A multiomic signature, which combined tumor and lymph node MRI radiomics, clinical and pathologic traits, and molecular subtypes, exhibited superior ALN status prediction performance, with AUCs of 0.90, 0.91, and 0.93 across the training, external validation, and prospective-retrospective validation cohorts respectively. Zhang et al (45), on the other hand, established and verified a multiparametric MRI-based radiomics nomogram for predicting axillary sentinel lymph node (SLN) burden in early-stage breast cancer before treatment. The developed radiomics nomogram, which integrated radiomics signature and MRI-determined ALN burden, demonstrated good calibration in predicting SLN burden, with AUCs of 0.82, 0.81, and 0.81 across the training, validation, and test cohorts respectively. Wu et al (46). formulated and assessed non-invasive models leveraging contrast-enhanced spectral mammography (CESM) to estimate the risk of non-sentinel lymph node (NSLN) metastasis and axillary tumor burden among breast cancer patients with 1-2 positive sentinel lymph nodes (SLNs). The radiomics nomogram for NSLN metastasis status prediction yielded an AUC of 0.85 in the test set and 0.82 in the temporal validation set. Yang et al (46). created a deep learning signature based on staging CT to preoperatively predict sentinel lymph node metastasis in breast cancer. The deep learning signature presented an impressive discriminative capability, with an AUC of 0.801 in the primary cohort and 0.817 in the validation cohort.
The use of US video images offers several advantages over MRI, such as cost-effectiveness, wider availability, and no need for contrast agent administration. Additionally, since the usual practice involves ultrasound-guided biopsy after MRI findings, our study further supports the potential of US images in developing robust DL algorithms. These algorithms have practical value in predicting lymph node status, assisting in decision-making, and improving patient management.
Importantly, this DL model offers several advantages for evaluating the status of ALNs. First, it can provide a probability of ALN metastasis for input breast US video images, potentially preventing misdiagnosis of axillary malignant lymph nodes by junior ultra-sonographers with less than 5 years of experience. Second, the interpretability of the DL model is increased through the attention heatmap generated by CAM, which is a tool for visualizing CNN networks. The attention heatmap allows us to see which regions in the image were relevant to this class (47). Therefore, we can infer which keyframe of the input videos is focused on by the DL model using the attention heatmaps. Third, the DL model’s performance allows rescanning of patients with negative axillary lymph node ultrasound. The attention heatmap of the metastatic nodes represented evidence of DL model classification and could assist in clinical decision making by directly identifying the ROI.
However, this study had several limitations. First, the DL model was developed for predicting the probability of ALN metastasis in early breast cancer patients with lesion ultrasonic videos. Ultra-sonographers still need to identify lymph nodes on US and input the videos. Second, this was a retrospective study, and the dataset was limited. Some of the patients with negative lymph nodes may have positive lymph nodes if followed up for a long enough time. Third, this study was based on three centers, and more external validation studies are necessary.
In conclusion, based on US video images, we developed a prediction model of ALN metastasis in early breast cancer patients using deep learning approach and obtained
good performance in the independent external validation cohort. The model may provide an effective diagnostic reference for metastasis in early breast cancer patients clinical use.
Data availability statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Ethics statement
The studies involving human participants were reviewed and approved by Xiang’an Hospital of Xiamen University. The patients/participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.
Author contributions
W-BL and Z-CD contributed equally to this work and share first authorship. W-BL collected data. W-BL and Z-CD drafted the manuscript. QD participated the data processing. Y-JL, J-XG, J-GW performed the analysis. W-HH conceived the study. All authors contributed to the article and approved the submitted version.
Funding
This work was supported by the Natural Science Foundation Committee of China (Grant No.32171363), Fujian Major Scientific and Technological Special Project for Social Development (Grant No.2020YZ016002), the Science and Technology Project of Xiamen Municipal Bureau of Science and Technology (Grant No. 3502Z20199047) and an open project of Xiamen’s Key Laboratory of Precision Medicine for Endocrine-Related Cancers (Grant No. XKLEC 2020KF02), The Fujian Provincial Medical Innovation Project (2020CXB053), Natural Science Foundation of Fujian Province,(No. 2021D035), Natural Science Foundation of Tibet, China (XZ202101ZR0087G).
Acknowledgments
We thank all our authors listed in this manuscript and thank to the Guojun Zhang’s team and LianSheng Wang’s team for providing help.
Conflict of interest
The handling editor SGW declared a shared parent affiliation with the authors W-BL, Z-CD, J-XG, QD, WH at the time of review.
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2023.1219838/full#supplementary-material
References
1. Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A, et al. Global Cancer Statistics 2020: GLOBOCAN estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA Cancer J Clin (2021) 71(3):209–49. doi: 10.3322/caac.21660
2. Allemani C, Matsuda T, Di Carlo V, Harewood R, Matz M, Nikšić M, et al. Global surveillance of trends in cancer survival 2000-14 (CONCORD-3): analysis of individual records for 37 513 025 patients diagnosed with one of 18 cancers from 322 population-based registries in 71 countries. Lancet (2018) 391(10125):1023–75. doi: 10.1016/S0140-6736(17)33326-3
3. Chang JM, Leung JWT, Moy L, Ha SM, Moon WK. Axillary nodal evaluation in breast cancer: state of the art. Radiology (2020) 295(3):500–15. doi: 10.1148/radiol.2020192534
4. Connor C, McGinness M, Mammen J, Ranallo L, Lafaver S, Klemp J, et al. Axillary reverse mapping: a prospective study in women with clinically node negative and node positive breast cancer. Ann Surg Oncol (2013) 20(10):3303–7. doi: 10.1245/s10434-013-3113-4
5. Sclafani LM, Baron RH. Sentinel lymph node biopsy and axillary dissection: added morbidity of the arm, shoulder and chest wall after mastectomy and reconstruction. Cancer J (2008) 14(4):216–22. doi: 10.1097/PPO.0b013e31817fbe5e
6. Boughey JC, Moriarty JP, Degnim AC, Gregg MS, Egginton JS, Long KH. Cost modeling of preoperative axillary ultrasound and fine-needle aspiration to guide surgery for invasive breast cancer. Ann Surg Oncol (2010) 17(4):953–8. doi: 10.1245/s10434-010-0919-1
7. Krag D, Weaver D, Ashikaga T, Moffat F, Klimberg VS, Shriver C, et al. The sentinel node in breast cancer–a multicenter validation study. N Engl J Med (1998) 339(14):941–6. doi: 10.1056/NEJM199810013391401
8. Krag DN, Anderson SJ, Julian TB, Brown AM, Harlow SP, Ashikaga T, et al. Technical outcomes of sentinel-lymph-node resection and conventional axillary-lymph-node dissection in patients with clinically node-negative breast cancer: results from the NSABP B-32 randomised phase III trial. Lancet Oncol (2007) 8(10):881–8. doi: 10.1016/S1470-2045(07)70278-4
9. Pesek S, Ashikaga T, Krag LE, Krag D. The false-negative rate of sentinel node biopsy in patients with breast cancer: a meta-analysis. World J Surg (2012) 36(9):2239–51. doi: 10.1007/s00268-012-1623-z
10. Fujioka T, Mori M, Kubota K, Yamaga E, Yashima Y, Oda G, et al. Clinical Usefulness of Ultrasound-Guided Fine Needle Aspiration and Core Needle Biopsy for Patients with Axillary Lymphadenopathy. Medicina (Kaunas) (2021) 57(7):3295–9.
11. Kikuchi Y, Mori M, Fujioka T, Yamaga E, Oda G, Nakagawa T, et al. Feasibility of ultrafast dynamic magnetic resonance imaging for the diagnosis of axillary lymph node metastasis: A case report. Eur J Radiol Open (2020) 7:100261. doi: 10.1016/j.ejro.2020.100261
12. Yang J, Wang T, Yang L, Wang Y, Li H, Zhou X, et al. Preoperative prediction of axillary lymph node metastasis in breast cancer using mammography-based radiomics method. Sci Rep (2019) 9(1):4429. doi: 10.1038/s41598-019-40831-z
13. Garayoa J, Chevalier M, Castillo M, Mahillo-Fernández I, Amallal El Ouahabi N, Estrada C, et al. Diagnostic value of the stand-alone synthetic image in digital breast tomosynthesis examinations. Eur Radiol (2018) 28(2):565–72. doi: 10.1007/s00330-017-4991-9
14. Kornecki A. Current status of breast ultrasound. Can Assoc Radiol J (2011) 62(1):31–40. doi: 10.1016/j.carj.2010.07.006
15. Winkler NS. Ultrasound guided core breast biopsies. Tech Vasc Interv Radiol (2021) 24(3):100776. doi: 10.1016/j.tvir.2021.100776
16. Youk JH, Son EJ, Gweon HM, Kim H, Park YJ, Kim JA. Comparison of strain and shear wave elastography for the differentiation of benign from malignant breast lesions, combined with B-mode ultrasonography: qualitative and quantitative assessments. Ultrasound Med Biol (2014) 40(10):2336–44. doi: 10.1016/j.ultrasmedbio.2014.05.020
17. Asaka K, Akai H, Kunimatsu A, Kiryu S, Abe O. Deep learning with convolutional neural network in radiology. Jpn J Radiol (2018) 36(4):257–72. doi: 10.1007/s11604-018-0726-3
18. Fujioka T, Mori M, Kubota K, Oyama J, Yamaga E, Yashima Y, et al. The utility of deep learning in breast ultrasonic imaging: A review. Diagnostics (Basel) (2020) 10(12). doi: 10.3390/diagnostics10121055
19. Bahl M. Artificial intelligence: A primer for breast imaging radiologists. J Breast Imaging. (2020) 2(4):304–14. doi: 10.1093/jbi/wbaa033
20. Burt JR, Torosdagli N, Khosravan N, RaviPrakash H, Mortazi A, Tissavirasingham F, et al. Deep learning beyond cats and dogs: recent advances in diagnosing breast cancer with deep neural networks. Br J Radiol (2018) 91(1089):20170545. doi: 10.1259/bjr.20170545
21. Ha R, Chang P, Karcich J, Mutasa S, Fardanesh R, Wynn RT, et al. Axillary lymph node evaluation utilizing convolutional neural networks using MRI dataset. J Digit Imaging. (2018) 31(6):851–6. doi: 10.1007/s10278-018-0086-7
22. Sun Q, Lin X, Zhao Y, Li L, Yan K, Liang D, et al. Deep learning vs. Radiomics for predicting axillary lymph node metastasis of breast cancer using ultrasound images: don't forget the peritumoral region. Front Oncol (2020) 10:53. doi: 10.3389/fonc.2020.00053
23. Zhou LQ, Wu XL, Huang SY, Wu GG, Ye HR, Wei Q, et al. Lymph node metastasis prediction from primary breast cancer US images using deep learning. Radiology (2020) 294(1):19–28. doi: 10.1148/radiol.2019190372
24. Zheng X, Yao Z, Huang Y, Yu Y, Wang Y, Liu Y, et al. Deep learning radiomics can predict axillary lymph node status in early-stage breast cancer. Nat Commun (2020) 11(1):1236. doi: 10.1038/s41467-020-15027-z
25. Guo X, Liu Z, Sun C, Zhang L, Wang Y, Li Z, et al. Deep learning radiomics of ultrasonography: Identifying the risk of axillary non-sentinel lymph node involvement in primary breast cancer. EBioMedicine (2020) 60:103018. doi: 10.1016/j.ebiom.2020.103018
26. Litjens G, Kooi T, Bejnordi BE, Setio AAA, Ciompi F, Ghafoorian M, et al. A survey on deep learning in medical image analysis. Med Image Anal (2017) 42:60–88. doi: 10.1016/j.media.2017.07.005
27. He K, Zhang X, Ren S, Sun J. Deep Residual Learning for Image Recognition. IEEE (2016) doi: 10.1109/CVPR.2016.90
28. Shao H, Qian S, Liu Y. Temporal Interlacing Network. Proceedings of the AAAI Conference on Artificial Intelligence. (2020)
29. Tran D, Wang H, Torresani L, Ray J, LeCun Y, Paluri M. A Closer Look at Spatiotemporal Convolutions for Action Recognition. (2018). Conference on Computer Vision and Pattern Recognition, Salt Lake City, UT, USA. 6450–6459.
30. Shao H, Qian S, Liu Y. Temporal Interlacing Network. Proceedings of the AAAI Conference on Artificial Intelligence. (2020).
31. Cornwell LB, McMasters KM, Chagpar AB. The impact of lymphovascular invasion on lymph node status in patients with breast cancer. Am Surg (2011) 77(7):874–7. doi: 10.1177/000313481107700722
32. Bae MS, Shin SU, Song SE, Ryu HS, Han W, Moon WK. Association between US features of primary tumor and axillary lymph node metastasis in patients with clinical T1-T2N0 breast cancer. Acta Radiol (2018) 59(4):402–8. doi: 10.1177/0284185117723039
33. Cho N, Moon WK, Han W, Park IA, Cho J, Noh DY. Preoperative sonographic classification of axillary lymph nodes in patients with breast cancer: node-to-node correlation with surgical histology and sentinel node biopsy results. AJR Am J Roentgenol. (2009) 193(6):1731–7. doi: 10.2214/AJR.09.3122
34. Yang WT, Chang J, Metreweli C. Patients with breast cancer: differences in color Doppler flow and gray-scale US features of benign and malignant axillary lymph nodes. Radiology (2000) 215(2):568–73. doi: 10.1148/radiology.215.2.r00ap20568
35. Bedi DG, Krishnamurthy R, Krishnamurthy S, Edeiken BS, Le-Petross H, Fornage BD, et al. Cortical morphologic features of axillary lymph nodes as a predictor of metastasis in breast cancer: in vitro sonographic study. AJR Am J Roentgenol. (2008) 191(3):646–52. doi: 10.2214/AJR.07.2460
36. Magny SJ, Shikhman R, Keppke AL. Breast Imaging Reporting and Data System. StatPearls. Treasure Island (FL: StatPearls Publishing LLC (2023).
37. Morrow M. Management of the node-positive axilla in breast cancer in 2017: selecting the right option. JAMA Oncol (2018) 4(2):250–1. doi: 10.1001/jamaoncol.2017.3625
38. Giuliano AE, Ballman KV, McCall L, Beitsch PD, Brennan MB, Kelemen PR, et al. Effect of axillary dissection vs no axillary dissection on 10-year overall survival among women with invasive breast cancer and sentinel node metastasis: the ACOSOG Z0011 (Alliance) randomized clinical trial. Jama (2017) 318(10):918–26. doi: 10.1001/jama.2017.11470
39. Chartrand G, Cheng PM, Vorontsov E, Drozdzal M, Turcotte S, Pal CJ, et al. Deep learning: A primer for radiologists. Radiographics (2017) 37(7):2113–31. doi: 10.1148/rg.2017170077
40. Sun S, Mutasa S, Liu MZ, Nemer J, Sun M, Siddique M, et al. Deep learning prediction of axillary lymph node status using ultrasound images. Comput Biol Med (2022) 143:105250. doi: 10.1016/j.compbiomed.2022.105250
41. Qiu SQ, Aarnink M, Van Maaren MC, Dorrius MD, Bhattacharya A, Veltman J, et al. Validation and update of a lymph node metastasis prediction model for breast cancer. Eur J Surg Oncol (2018) 44(5):700–7. doi: 10.1016/j.ejso.2017.12.008
42. Qiu SQ, Zeng HC, Zhang F, Chen C, Huang WH, Pleijhuis RG, et al. A nomogram to predict the probability of axillary lymph node metastasis in early breast cancer patients with positive axillary ultrasound. Sci Rep (2016) 6:21196. doi: 10.1038/srep21196
43. Ren T, Cattell R, Duanmu H, Huang P, Li H, Vanguri R, et al. Convolutional neural network detection of axillary lymph node metastasis using standard clinical breast MRI. Clin Breast Cancer. (2020) 20(3):e301–e8. doi: 10.1016/j.clbc.2019.11.009
44. Yu Y, He Z, Ouyang J, Tan Y, Chen Y, Gu Y, et al. Magnetic resonance imaging radiomics predicts preoperative axillary lymph node metastasis to support surgical decisions and is associated with tumor microenvironment in invasive breast cancer: A machine learning, multicenter study. EBioMedicine (2021) 69:103460. doi: 10.1016/j.ebiom.2021.103460
45. Zhang X, Yang Z, Cui W, Zheng C, Li H, Li Y, et al. Preoperative prediction of axillary sentinel lymph node burden with multiparametric MRI-based radiomics nomogram in early-stage breast cancer. Eur Radiol (2021) 31(8):5924–39. doi: 10.1007/s00330-020-07674-z
46. Yang X, Wu L, Ye W, Zhao K, Wang Y, Liu W, et al. Deep learning signature based on staging CT for preoperative prediction of sentinel lymph node metastasis in breast cancer. Acad Radiol (2020) 27(9):1226–33. doi: 10.1016/j.acra.2019.11.007
Keywords: axillary lymph node metastasis, artificial intelligence, ultrasound video image, breast lesion, deep learning model
Citation: Li W-B, Du Z-C, Liu Y-J, Gao J-X, Wang JG, Dai Q and Huang W-H (2023) Prediction of axillary lymph node metastasis in early breast cancer patients with ultrasonic videos based deep learning. Front. Oncol. 13:1219838. doi: 10.3389/fonc.2023.1219838
Received: 09 May 2023; Accepted: 06 July 2023;
Published: 01 September 2023.
Edited by:
San-Gang Wu, First Affiliated Hospital of Xiamen University, ChinaReviewed by:
Hua-Rong Ye, China Resources & Wisco General Hospital Affiliated to Wuhan University of Science and Technology, ChinaSubathra Adithan, Jawaharlal Institute of Postgraduate Medical Education and Research (JIPMER), India
Copyright © 2023 Li, Du, Liu, Gao, Wang, Dai and Huang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Wen-He Huang, aHVhbmd3ZW5oZTIwMDlAMTYzLmNvbQ==
†These authors have contributed equally to this work and share first authorship