- 1Department of Ultrasound, Shanghai General Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
- 2Shanghai Key Lab of Digital Media Processing and Transmission, Shanghai Jiao Tong University, Shanghai, China
- 3Department of Ultrasound, Zhongshan Hospital Wusong Branch, Fudan University, Shanghai, China
Purpose: A convolutional neural network (CNN) can perform well in either of two independent tasks [classification and axillary lymph-node metastasis (ALNM) prediction] based on breast ultrasound (US) images. This study is aimed to investigate the feasibility of performing the two tasks simultaneously.
Methods: We developed a multi-task CNN model based on a self-built dataset containing 5911 breast US images from 2131 patients. A hierarchical loss (HL) function was designed to relate the two tasks. Sensitivity, specificity, accuracy, precision, F1-score, and analyses of receiver operating characteristic (ROC) curves and heatmaps were calculated. A radiomics model was built by the PyRadiomics package.
Results: The sensitivity, specificity and area under the ROC curve (AUC) of our CNN model for classification and ALNM tasks were 83.5%, 71.6%, 0.878 and 76.9%, 78.3%, 0.836, respectively. The inconsistency error of ALNM prediction corrected by HL function decreased from 7.5% to 4.2%. Predictive ability of the CNN model for ALNM burden (≥3 or ≥4) was 77.3%, 62.7%, and 0.752, and 66.6%, 76.8%, and 0.768, respectively, for sensitivity, specificity and AUC.
Conclusion: The proposed multi-task CNN model highlights its novelty in simultaneously distinguishing breast lesions and indicating nodal burden through US, which is valuable for “personalized” treatment.
Introduction
The accurate diagnosis and precise staging of breast cancer (BC) is crucial for clinical decision-making, and is based on imaging examinations (Plevritis et al., 2018; Guo et al., 2018a; da Costa Vieira et al., 2017). According to the Tumor–Node–Metastasis (TNM) staging system, axillary lymph node (ALN) status is the principal factor for the clinical staging of BC (Giuliano et al., 2017). Ultrasound (US) is a routine imaging modality used to evaluate breast masses and is recommended for ALN assessment according to guidelines (Chang et al., 2020). Studies have shown the pooled sensitivity and specificity to be 0.87 and 0.80, respectively, and the area under the receiver operating characteristic curve (AUC) to be 0.9049, for conventional US in breast-lesion classification (Wubulihasimu et al., 2018). For ALNs, US can detect axillary lymph node metastases (ALNM) with sensitivity and specificity ranging from 26% to 76%, and 88% to 98%, respectively, based on morphologic criteria (Houssami et al., 2014). However, operator dependence and unclear diagnostic criteria for US of the breast limit its progress towards a precise diagnosis. These knowledge gaps call for confirmation of the presence/absence and extent of ALNM preoperatively for suspicious breast lesions concurrently for “tailored” treatment plans.
Several studies have shown that the US findings of primary BC might be associated with its malignancy and ALNM (Akissue de Camargo Teixeira et al., 2017; Bae et al., 2018; Guo et al., 2018b; Yu et al., 2018). Thus, one can predict malignancy and ALNM status simultaneously based on the morphologic characteristics of a primary breast mass. These relevant morphologic features include the tumor diameter, indistinct margins, and architectural distortion (Akissue de Camargo Teixeira et al., 2017; Bae et al., 2018; Guo et al., 2018b; Yu et al., 2018). However, some of these morphologic features are subjective and even invisible to the naked eyes of radiologists. Therefore, completing such sophisticated work is challenging for radiologists.
The advent of powerful artificial intelligence (AI) technology [particularly deep learning (DL) algorithms] could help to reduce the number of hospital visits and financial costs for patients because only US examination would be used (Sadoughi et al., 2018; Le et al., 2019). With self-learned features from US images of the breast, Zhao and others reported that the diagnostic performance of deep convolutional neural networks (CNNs) was comparable with that of expert radiologists, and also improved inter-observer agreement among radiologists (Zhao et al., 2019). Other studies have demonstrated the feasibility of predicting ALNM through the characteristics of US images of primary BC using DL models (Moon et al., 2018; Sun et al., 2020). However, AI studies have not explored the dual-task performance of breast-mass classification and ALNM prediction on US images.
We explored the feasibility of predicting both malignancy and ALNM using a unified CNN model based on US images of primary breast masses and exploring the underlying relationship between these two tasks. Such feasibility indicates that it is potential to build a multi-functional AI that could perform comprehensive analysis given US images. Specifically, the relationship among the multiple clinical tasks could be potentially learned as prior knowledge to guide feature optimization procedure and thus constrain the proposed CNN model to obtain a more accurate diagnosis result. Such a working mechanism could be considered as an extension of the diagnosis process of radiologist and could not be built by a radiomics model. Actually, it can be expected that we extend the model for more related AI functions by simply adding more task-specific heads, for example, molecular subtypes predictor.
Materials and Methods
Study Cohort
For this retrospective study, Institutional Review Board approval was obtained from our ethics committee (No. 2019KY055). The requirement for obtaining written informed consent from the subjects was waived by the Institutional Review Board. From August 2011 to December 2019, 5911 US images of breast masses were collected from 2131 patients (2120 females and 11 males). The nodules included in our study matched those detected by pathology after surgical resection. The inclusion criteria of this study were as follows: 1) All the breast masses were pathologically proven; 2) The patients had not undergone radiotherapy, chemotherapy and other anti-tumor treatments; 3) Breast cancer with ALNM were all ipsilateral with one malignant tumor. The flowchart of our study is shown as Figure 1. The dataset was split into sets (training, validation and testing), where images from the same patient were confirmed in the same part, and 30% of patients were chosen randomly once for all as the test set. All hyperparameters were chosen based on the performance on validation parts.
FIGURE 1. Flowchart of the procedures in the development and evaluation of our deep-learning model and radiomics model for malignancy prediction and ALNM prediction.
US images were acquired using US equipment obtained from Aixplorer (Super Imagine; Aix-en-Provence, France), Philips (IU22; Amsterdam, Netherlands), GE Healthcare (LOGIQ E9; Pittsburgh, PA, United States), Hitachi (EUB 8500; Tokyo, Japan), Esaote (MyLab™Twice; Genoa, Italy) and Siemens (Sequia512 and ACUSON S3000; Munich, Germany) with linear transducers of frequency 5–12 MHz.
Data Preprocessing
An annotation tool for graphical images, LabelImg (https://github.com/tzutalin/labelImg), was used to crop the box-level region of interest (RoI), which was delineated manually and confirmed by two radiologists with >5 years of experience in breast US interpretation. Two radiologists delineated different parts of the study cohort and therefore there is no discrepancy. To achieve data diversity and to prevent overfitting, strong data augmentation was adopted for image processing. Geometric transformations of random images (including random cropping and scaling, random flipping, and random color distortion) were used. The augmented images were resized to 256 × 256 pixels before being fed into the model. We implemented the training-and-testing framework with PyTorch (https://pytorch.org/).
Hierarchical Loss for Prediction Consistency Between Two Tasks
We proposed a multi-task framework to predict the malignancy of the tumor and ALNM simultaneously. Tumors with ALNM are always malignant. Accordingly, HL was designed to constrain the model to achieve consistent predictions between two tasks (if there was a high/low score for ALNM, there should also be a high/low score for malignancy) and formed as:
where
Multi-Task Convolutional Neural Network Framework
The overall learning architecture of our model is illustrated in Figure 2. We used ResNet as feature extractor. The classifiers were fully connected layers with a sigmoid function as the activation. The sigmoid function scaled the output of classifiers to a range of [0,1], which could be explained as the probability of a malignant tumor with ALNM. Binary cross entropy (BCE) loss was used for classification loss because both tasks could be formulated as a binary classification. The overall loss function for training was the weighted sum of BCE loss and HL, as follows:
where the hyperparameter
Radiomics Model
We also built an advanced radiomics model proposed by Sun et al. (2020) as a baseline to compare with our proposed multitask CNN. The radiomics model extracts statistical features from ultrasound images and performs classification with random forest algorithms. The advantages of our proposed multitask CNN framework against the radiomics model are twofold: 1) The features extracted by CNN model are optimized by the classification objective, while those extracted by the radiomics model are manually designed, which are difficult to be generalized for extensively varied image contents and inclined to be non-optimal towards classification; 2) In our framework, the two tasks could be performed simultaneously without re-training and more importantly, the relationship between the two tasks could be utilized to enhance the feature representation and results in classification performance improvements. For each US image, 102 radiomics features were extracted using Pyradiomics (https://pyradiomics.readthedocs.io) (Supplementary Table S1), including intensity features and texture features. When extracting these features, data augmentation is not utilized for the reproduction of extracted features. Shape features were not considered because pixel-level labeling of RoIs required very expensive manual labor. The Boruta feature-selection algorithm was used for dimension reduction to avoid overfitting. The selected features for each task are summarized in Supplementary Table S2. We used the Python package Boruta for implementation (https://pypi.org/project/Boruta/). The classifier of our radiomics model was the Random Forest (RF) classifier with a maximum depth of 300. The scikit-learn package (https://scikit-learn.org/stable/) in Python was used for implementation.
Evaluation and Statistical Analyses
The difference in the characteristics of patients and lesions were assessed with the chi-square test or Student’s t-test, where appropriate. Descriptive data are the mean ± SD. Statistical tests were two-sided and undertaken with the scipy package (https://www.scipy.org/) in Python. p < 0.05 was considered significant. Experiments were repeated five times to ensure reproducibility. We introduced gradient-weighted class activation mapping (Grad-CAM) (Selvaraju et al., 2017) to visualize the feature extracted by the model (heatmap). Accuracy, sensitivity, specificity, F1-score and precision were used for performance evaluation. It should be noted that these metrics are calculated according to the relationship between model prediction and the ground truth. Actually, these metrics are calculated according to the number of true-positive, true-negative, false-positive and false-negatives, the definition of which regarding the pathological result could be found in Appendix. It should be also noted that calculating these metrics require a discrete prediction while the output of model is a real value in range [0, 1] interpreted as the probability of being positive. Therefore, we follow the common practice of setting a threshold as 0.5 on both ALNM task and BM task. That is, if the output of model is greater than 0.5, the prediction would be considered as positive. To further evaluate the performance of model under different thresholds, receiver operating characteristic (ROC) curve would be plotted and the area under curve (AUC) would be reported.
Experimental Configuration
ResNet50 (He et al., 2016) was used for all feature extractors in our DL models. The Adam optimization algorithm was selected to train our models with initial learning rate of 1e-4. All models were trained for 200 epochs, with the learning rate reduced by a factor of 10 every 50 epochs. Two ablation experiments were conducted whereby the two tasks were undertaken independently and then together but without the HL function. In the present study, “Single-BM” denotes the algorithm using ResNet to conduct malignancy prediction, “Single-ALNM” denotes the algorithm using ResNet to conduct ALNM prediction, and “Multi-task” denotes the algorithm which uses ResNet as the shared backbone and two classifiers to predict the scores of the two tasks simultaneously without the HL term.
Results
Patient Data
The clinical and pathological results of the whole dataset are shown in Table 1. Finally, 1491 patients (mean age± SD: 54.25 ± 14.29 years; range: 17–88 years) with 4077 images were used for the training set and validation set, and 640 patients (mean age± SD: 55.04 ± 14.49 years; range: 21–91 years) with 1834 images were used for the test set. For prediction of breast-mass malignancy, the dataset comprised 3909 images of 1811 malignant breast lesions and 2002 images of 320 benign lesions. For ALNM prediction, 489 ALN-positive patients with 1067 images and 1322 ALN-negative patients with 2842 images were sorted for investigation.
Convolutional Neural Network Model Versus Radiomics Model
For the BM task, our model showed an encouraging improvement in performance, with an AUC of 0.878, sensitivity of 83.5% and specificity of 71.6% compared with those in the radiomics model, which achieved an AUC of 0.848 (p < 0.001) (Table 2; Figure 3). With respect to ALNM prediction, the AUC, sensitivity, and specificity of our model was 0.836, 76.9% and 78.3%, respectively, which were all superior to those of the radiomics model (p < 0.05 for all metrics).
TABLE 2. Performance of the deep-learning model and radiomics model on the test cohort on ALNM prediction and benign/malignant classification task.
FIGURE 3. Receiver operating characteristic curves of our deep-learning model and a traditional radiomics model on malignancy prediction and ALNM prediction. (A) and ALNM prediction (B) tasks.
Evaluation of Hierarchical Loss Function
The malignancy prediction for Multi-Task showed no significant difference from Single-BM (0.874 vs. 0.871, p = 0.379) (Table 3). The ALNM prediction for Multi-Task showed no significant difference from Single-BM (0.820 vs. 0.817, p = 0.597), and we noticed that combining the two tasks together did not necessarily improve the performance because the collaboration between the two tasks was not well-built. The AUC value of ALNM prediction with HL introduction surpassed that of Multi-Task (0.836 vs. 0.817, p = 0.009), which indicated that the prior knowledge that tumors with ALNM are always malignant imposed a valid interactive constraint between two classifiers. All p-values are listed in Supplementary Table S3. We further study the impact of HL for reducing inconsistent predictions. The percentages of two types of inconsistency error in the test set for models without HL and with HL are 7.5% vs. 4.2% and 2.0% vs. 1.1%, respectively. The impact of HL for reducing inconsistent predictions are shown in Figure 4.
Deep Learning Interpretability With Gradient-Weighted Class Activation Mapping
Heatmaps are shown as Figure 5. The color change from blue to red represents the significance of the regions for prediction of a malignant tumor or ALNM ranging from low to high, respectively. In Figures 5B,C, intratumoral regions and peritumoral regions are highlighted with hot colors, and contribute dominantly to the prediction of malignant tumors or ALNM. We also studied the average CAM response over different distances to image centers on the test set (Supplementary Figure S1). The result indicated that pixels closer to the image center (which denotes a higher probability of being the inratumoral region) contributed more to the prediction. Pixels distant from the image center tended to be the intratumoral region and background. In Figure 6, the heatmap of the model supervised only by the ALNM label (third row) showed that the model highlighted only the smooth peritumoral region rather than considering the peritumoral and intratumoral information together. However, for the model supervised by malignancy and ALNM labels (second row), the peritumoral and intratumoral regions were highlighted. These data might explain the reason for the performance gain of ALNM prediction: introduction of HL and malignancy supervision.
FIGURE 5. Heatmaps generated by Grad-CAM. The color ranges from blue to red, and represents the significance value of the region ranging from low to high, respectively, for the prediction of a malignant tumor or ALNM. (A) Benign case which was classified correctly. (B) Malignant masses without ALNM. (C) and (D) Malignant masses.
FIGURE 6. Heatmaps produced using Grad-CAM. The first row (A–D) shows US images from patients with ALNM. The second row shows heatmaps generated from our model that performed ALNM prediction and malignancy prediction. The third row shows heatmaps generated from the model performing only ALNM prediction.
Prediction of the Nodal Burden of Axillary Lymph-Node Metastasis
To calculate the nodal burden of ALNM, we built several binary classification tasks whereby positive samples were images from patients with ALNM number no less than a given number ranging from 1 to 4. The detailed performance evaluation of the CNN method and radiomics method on this set of binary classification tasks is summarized in Table 4. The score for accuracy, sensitivity, specificity, and the AUC for the number of ALNM ≥ 2, ALNM ≥ 3 and ALNM ≥ 4 by the CNN model was, respectively: 63.4%, 67.2%, 63.0%, and 0.715; 63.6%, 77.3%, 62.7%, and 0.752; 75.3%,66.6%, 76.8%, and 0.768.
TABLE 4. Performance of the radiomics model and our model on prediction of the exact number of lymph-node metastases.
Discussion
We successfully developed a multi-task architecture to integrate two traditional single-mode classification tasks to achieve prediction of ALNM and malignancy in parallel. In the practical work, the radiologist must make the classification of breast nodules firstly, and then the lymph node status assessment. This sequence is a comprehensive diagnosis. However, many previous research for the prediction of lymph node were based on selective breast cancer tumors, do not tally with the actual diagnosis process. Therefore, in our opinion, this study is consistent with the clinical diagnostic process and has good clinical relevance based on clinical issues. The HL function was utilized to embed the prior constraint between the two tasks that tumors predicted with ALNM should obtain a label of “malignancy” while there must be no metastasis to LNs in a benign breast tumor. Ablation studies were undertaken to evaluate the impact of the HL function, which was shown to be effective for reducing inconsistent prediction errors. This is an innovative study that is different from the previous task of single classification or lymph node prediction model. In clinical practice, if US examination cannot provide a correct assessment of lymph node status, patients may need MRI examination or sentinel lymph node biopsy to confirm the status of lymph nodes. Using our model, radiologists and physicians could reach a diagnosis quickly without resorting to other examinations efficiently and economically.
Numerous studies suggested that DL provides a powerful assistant tool for radiologists to reduce their workload (Xiao et al., 2018; Wu et al., 2019). The excellent ability of CNNs to extract image features was proved to be superior to radiomics using CT, US or MRI (Truhn et al., 2019; Caballo et al., 2020; Sun et al., 2020). Consistent with those results, our CNN model was also superior to the radiomics model based on the same dataset. Han and others used a CNN framework based on large-scale data to differentiate breast lesions on US images, and achieved an accuracy of ∼0.9, sensitivity of ∼0.86, and specificity of ∼0.96 (Han et al., 2017). Our model achieved a relatively lower specificity (0.72) due to the sample composition, in which the number of breast tumors was in the majority. As a retrospective study, images from multiple machines were also responsible for this classification performance. However, our model had the possibility of eliciting reproducibility and repeatability in daily practice. Moreover, based on this preliminary study, our next work is to build a comprehensive model by combining clinical data and multimodal ultrasound image data to accomplish this dual task, which may yield better diagnostic performance.
Most importantly, our study provides a valid approach to predict the nodal burden of ALNM of malignant breast tumors. Zhou and others revealed that the best-performing model significantly outperformed the three experienced radiologists in ALNM prediction, but detailed data were not provided (Zhou et al., 2020). The nodal burden of ALNM is an important parameter for clinical treatment. According to the results of the Z0011 trial from the American College of Surgeons Oncology Group, the clinical treatment for ALN has been changed further worldwide. This is because dissection of ALNs does not affect the overall survival or disease-free survival of patients with T1–T2 BC, or <3 positive sentinel LNs under different treatment regimens (Giuliano et al., 2016). In the TNM staging system, ALNM ≥ 4 increases the clinical stage of BC, which corresponds to a change in the treatment regimen (Giuliano et al., 2017). Compared with the study conducted by Zheng et al. (2020), in addition to prediction of ALN status of N0, N+(≥1),N+(1–2), N+(≥3), our study carried out more attempts at lymph node metastasis status (N ≥ 4), which is important reference information for the clinical staging of breast cancer.
Our study showed moderate performance in prediction of ALNM number, which might have been due to our small sample size. Nevertheless, we demonstrated the potential of CNN model in predicting the ALN staging of BC. Our model achieved better performance on ALNM prediction with higher nodal burden, which indicates that there were more hard samples among breast tumors with a lower nodal burden. For training the CNN model, these hard samples might provide incorrect gradients, which should be ignored to prevent the model from overfitting. A possible explanation for this scenario could be that the metastatic biological behaviors of breast tumors with a lower nodal burden are inconspicuous compared with those of a heavier breast tumor which displays well-marked morphologic features (Bae et al., 2018; Yu et al., 2018). Moreover, our results are in accordance with the fact that US has greater accuracy in detecting a heavy nodal burden of BC compared with lower nodal involvement (Rukanskienė et al., 2020).
To assist radiologists in understanding our CNN model and reveal the inter relationship of the two tasks, we visualized the heatmap (which shows the important parts of the US image of the breast during prediction). We discovered that peritumoral and intratumoral regions contributed to prediction of ALNM. This result is in accordance with studies suggesting that the biological changes of peritumoral regions might lead to metastatic spread (Wu et al., 2017; Cheon et al., 2018). This result might also explain the reason for the performance gain of ALNM prediction by introducing HL and malignancy supervision. The useful interaction between two tasks provides extra information, thereby forcing model to consider intratumoral and peritumoral regions to make more precise predictions.
Our study had two main limitations. First, the self-built dataset was of limited size. Further expansion of data with larger and multicenter datasets should be undertaken in the future. Second, our data were only derived from two-dimensional US images, future studies which focus on a combination of multi-mode US images of breast masses should be taken to improve the diagnostic performance of ALN status.
Conclusion
We developed a novel multi-task CNN model to predict malignancy and detailed LN status simultaneously based on US images of primary BC. For patients, US can be a cost-effective, convenient, and functional alternative examination. For physicians, preoperative evaluation could be accomplished with fewer screening items and shorter intervals between visits. Our CNN model has considerable potential for assisting clinically precise treatment for BC.
Data Availability Statement
The datasets presented in this article are not readily available because This data involves patient privacy information that we have chosen not to disclose. Requests to access the datasets should be directed to d3Vyb25nNzExMUAxNjMuY29t.
Author Contributions
All authors approved the final submission of the manuscript. RW and YX designed the study. CL, LJ, MY, SS, and JC performed the dataset collection. YG performed the image analysis. YX guided the data analysis. CL, YG, and YX contributed to the manuscript drafting and editing.
Funding
This work was supported by the National Natural Science Foundation of China (Grants No. 82071931, 82130057), program for Shanghai Outstanding Medical Academic Leaders (2019LJ18), the interdisciplinary program of Shanghai Jiaotong University (ZH2018ZDA17), and the program from Science and Technology Commission of Shanghai Municipality (No. 20Y11912400).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
We thank the following residents and medical students: Yi Zheng, Tianhu Zhang, Luni Zhang, Xueting Yang, Yan Liu and Qian Xu for assisting with the collection of data from patient administration system.
Supplementary Material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fphys.2022.882648/full#supplementary-material
Abbreviations
CNN, Convolutional neural network; ROC, Receiver operating characteristic; AUC, Area under the ROC curve; TNM, Tumor, Node, Metastasis; AI, Artificial intelligence; DL, Deep learning; ROI, Region of interest; SD, Standard deviation.
References
Akissue de Camargo Teixeira P., Chala L. F., Shimizu C., Filassi J. R., Maesaka J. Y., de Barros N. (2017). Axillary Lymph Node Sonographic Features and Breast Tumor Characteristics as Predictors of Malignancy: a Nomogram to Predict Risk. Ultrasound Med. Biol. 43, 1837–1845. doi:10.1016/j.ultrasmedbio.2017.05.003
Bae M. S., Shin S. U., Song S. E., Ryu H. S., Han W., Moon W. K. (2018). Association between US Features of Primary Tumor and Axillary Lymph Node Metastasis in Patients with Clinical T1-T2n0 Breast Cancer. Acta Radiol. 59, 402–408. doi:10.1177/0284185117723039
Caballo M., Pangallo D. R., Mann R. M., Sechopoulos I. (2020). Deep Learning-Based Segmentation of Breast Masses in Dedicated Breast CT Imaging: Radiomic Feature Stability between Radiologists and Artificial Intelligence. Comput. Biol. Med. 118, 103629. doi:10.1016/j.compbiomed.2020.103629
Chang J. M., Leung J. W. T., Moy L., Ha S. M., Moon W. K. (2020). Axillary Nodal Evaluation in Breast Cancer: State of the Art. Radiology 295 (3), 500–515. doi:10.1148/radiol.2020192534
Cheon H., Kim H. J., Kim T. H., Ryeom H.-K., Lee J., Kim G. C., et al. (2018). Invasive Breast Cancer: Prognostic Value of Peritumoral Edema Identified at Preoperative MR Imaging. Radiology 287, 68–75. doi:10.1148/radiol.2017171157
da Costa Vieira R. A., Biller G., Uemura G., Ruiz C. A., Curado M. P. (2017). Breast Cancer Screening in Developing Countries. Clin. (Sao Paulo) 72, 244–253. doi:10.6061/clinics/2017(04)09
Giuliano A. E., Ballman K., Mccall L., Beitsch P., Whitworth P. W., Blumencranz P., et al. (2016). Locoregional Recurrence after Sentinel Lymph Node Dissection with or without Axillary Dissection in Patients with Sentinel Lymph Node Metastases. Ann. Surg. 264, 413–420. doi:10.1097/SLA.0000000000001863
Giuliano A. E., Connolly J. L., Edge S. B., Mittendorf E. A., Rugo H. S., Solin L. J., et al. (2017). Breast Cancer-Major Changes in the American Joint Committee on Cancer Eighth Edition Cancer Staging Manual. CA A Cancer J. Clin. 67 (4), 290–303. doi:10.3322/caac.21393
Guo Q., Dong Z., Zhang L., Ning C., Li Z., Wang D., et al. (2018). Ultrasound Features of Breast Cancer for Predicting Axillary Lymph Node Metastasis. J. Ultrasound Med. 37, 1354–1353. doi:10.1002/jum.14469
Guo R., Lu G., Qin B., Fei B. (2018). Ultrasound Imaging Technologies for Breast Cancer Detection and Management: A Review. Ultrasound Med. Biol. 44, 37–70. doi:10.1016/j.ultrasmedbio.2017.09.012
Han S., Kang H.-K., Jeong J.-Y., Park M.-H., Kim W., Bang W.-C., et al. (2017). A Deep Learning Framework for Supporting the Classification of Breast Lesions in Ultrasound Images. Phys. Med. Biol. 62, 7714–7728. doi:10.1088/1361-6560/aa82ec
He K., Zhang X., Ren S., Sun J. (2016). “Deep Residual Learning for Image Recognition,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 770–778. doi:10.1109/cvpr.2016.90
Houssami N., Diepstraten S. C., Cody H. S., Turner R. M., Sever A. R. (2014). Clinical Utility of Ultrasound-Needle Biopsy for Preoperative Staging of the Axilla in Invasive Breast Cancer. Anticancer Res. 34, 1087–1097.
Le E. P. V., Wang Y., Huang Y., Hickman S., Gilbert F. J. (2019). Artificial Intelligence in Breast Imaging. Clin. Radiol. 74, 357–366. doi:10.1016/j.crad.2019.02.006
Moon W. K., Chen I.-L., Yi A., Bae M. S., Shin S. U., Chang R.-F. (2018). Computer-aided Prediction Model for Axillary Lymph Node Metastasis in Breast Cancer Using Tumor Morphological and Textural Features on Ultrasound. Comput. Methods Programs Biomed. 162, 129–137. doi:10.1016/j.cmpb.2018.05.011
Plevritis S. K., Munoz D., Kurian A. W., Stout N. K., Alagoz O., Near A. M., et al. (2018). Association of Screening and Treatment with Breast Cancer Mortality by Molecular Subtype in US Women, 2000-2012. JAMA 319, 154–164. doi:10.1001/jama.2017.19130
Rukanskienė D., Veikutis V., Jonaitienė E., Basevičiūtė M., Kunigiškis D., Paukštaitienė R., et al. (2020). Preoperative Axillary Ultrasound versus Sentinel Lymph Node Biopsy in Patients with Early Breast Cancer. Medicina 56, 127. doi:10.3390/medicina56030127
Sadoughi F., Kazemy Z., Hamedan F., Owji L., Rahmanikatigari M., Talebi Azadboni T. (2018). Artificial Intelligence Methods for the Diagnosis of Breast Cancer by Image Processing: a Review. Bctt 10, 219–230. doi:10.2147/BCTT.S175311
Selvaraju R. R., Cogswell M., Das A., Vedantam R., Parikh D., Batra D. (2017). “Grad-cam: Visual Explanations from Deep Networks via Gradient-Based Localization,” in Proceedings of the IEEE international conference on computer vision, 618–626. doi:10.1109/iccv.2017.74
Sun Q., Lin X., Zhao Y., Li L., Yan K., Liang D., et al. (2020). Deep Learning vs. Radiomics for Predicting Axillary Lymph Node Metastasis of Breast Cancer Using Ultrasound Images: Don't Forget the Peritumoral Region. Front. Oncol. 10, 10–53. doi:10.3389/fonc.2020.00053
Truhn D., Schrading S., Haarburger C., Schneider H., Merhof D., Kuhl C. (2019). Radiomic versus Convolutional Neural Networks Analysis for Classification of Contrast-Enhancing Lesions at Multiparametric Breast MRI. Radiology 290, 290–297. doi:10.1148/radiol.2018181352
Wu G.-G., Zhou L.-Q., Xu J.-W., Wang J.-Y., Wei Q., Deng Y.-B., et al. (2019). Artificial Intelligence in Breast Ultrasound. Wjr 11, 19–26. doi:10.4329/wjr.v11.i2.19
Wu J., Li B., Sun X., Cao G., Rubin D. L., Napel S., et al. (2017). Heterogeneous Enhancement Patterns of Tumor-Adjacent Parenchyma at MR Imaging Are Associated with Dysregulated Signaling Pathways and Poor Survival in Breast Cancer. Radiology 285, 401–413. doi:10.1148/radiol.2017162823
Wubulihasimu M., Maimaitusun M., Xu X.-L., Liu X.-D., Luo B.-M. (2018). The Added Value of Contrast-Enhanced Ultrasound to Conventional Ultrasound in Differentiating Benign and Malignant Solid Breast Lesions: a Systematic Review and Meta-Analysis. Clin. Radiol. 73, 936–943. doi:10.1016/j.crad.2018.06.004
Xiao T., Liu L., Li K., Qin W., Yu S., Li Z. (2018). Comparison of Transferred Deep Neural Networks in Ultrasonic Breast Masses Discrimination. BioMed Res. Int. 2018, 1–9. doi:10.1155/2018/4605191
Yu X., Hao X., Wan J., Wang Y. Y., Yu L., Liu B. Y. (2018). Correlation between Ultrasound Appearance of Small Breast Cancer and Axillary Lymph Node Metastasis. Ultrasound Med. Biol. 44, 342–349. doi:10.1016/j.ultrasmedbio.2017.09.020
Zhao C., Xiao M., Jiang Y., Liu H., Wang M., Wang H., et al. (2019). Feasibility of Computer-Assisted Diagnosis for Breast Ultrasound: the Results of the Diagnostic Performance of S-Detect from a Single Center in China. Cmar 11, 921–930. doi:10.2147/CMAR.S190966
Zheng X., Yao Z., Huang Y., Yu Y., Wang Y., Liu Y., et al. (2020). Deep Learning Radiomics Can Predict Axillary Lymph Node Status in Early-Stage Breast Cancer. Nat. Commun. 11 (1), 1236. doi:10.1038/s41467-020-15027-z
Keywords: breast neoplasms, ultrasonography, neural networks, lymph nodes, multi-task
Citation: Li C, Guo Y, Jia L, Yao M, Shao S, Chen J, Xu Y and Wu R (2022) A Convolutional Neural Network Based on Ultrasound Images of Primary Breast Masses: Prediction of Lymph-Node Metastasis in Collaboration With Classification of Benign and Malignant Tumors. Front. Physiol. 13:882648. doi: 10.3389/fphys.2022.882648
Received: 24 February 2022; Accepted: 10 May 2022;
Published: 02 June 2022.
Edited by:
Lipeng Ning, Brigham and Women’s Hospital and Harvard Medical School, United StatesCopyright © 2022 Li, Guo, Jia, Yao, Shao, Chen, Xu and Wu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Rong Wu, d3Vyb25nNzExMUAxNjMuY29t; Yi Xu, eHV5aUBzanR1LmVkdS5jbg==
†These authors have contributed equally to this work and share first authorship