Skip to main content

ORIGINAL RESEARCH article

Front. Oncol. , 16 February 2022

Sec. Breast Cancer

Volume 12 - 2022 | https://doi.org/10.3389/fonc.2022.829041

This article is part of the Research Topic Quantitative Imaging and Artificial Intelligence in Breast Tumor Diagnosis View all 27 articles

Predicting HER2 Status in Breast Cancer on Ultrasound Images Using Deep Learning Method

Zilong XuZilong Xu1Qiwei YangQiwei Yang2Minghao LiMinghao Li2Jiabing GuJiabing Gu1Changping DuChangping Du1Yang ChenYang Chen1Baosheng Li,*Baosheng Li1,3*
  • 1Laboratory of Image Science and Technology, School of Computer Science and Engineering, Southeast University, Nanjing, China
  • 2Laboratory of Radiation Oncology, School of Medicine, Shandong University, Jinan, China
  • 3Department of Radiation Oncology, Shandong Cancer Hospital and Institute, Shandong First Medical University and Shandong Academy of Medical Sciences, Jinan, China

Purpose: The expression of human epidermal growth factor receptor 2 (HER2) in breast cancer is critical in the treatment with targeted therapy. A 3-block-DenseNet-based deep learning model was developed to predict the expression of HER2 in breast cancer by ultrasound images.

Methods: The data from 144 breast cancer patients with preoperative ultrasound images and clinical information were retrospectively collected from the Shandong Province Tumor Hospital. An end-to-end 3-block-DenseNet deep learning classifier was built to predict the expression of human epidermal growth factor receptor 2 by ultrasound images. The patients were randomly divided into a training (n = 108) and a validation set (n = 36).

Results: Our proposed deep learning model achieved an encouraging predictive performance in the training set (accuracy = 85.79%, AUC = 0.87) and the validation set (accuracy = 80.56%, AUC = 0.84). The effectiveness of our model significantly exceeded the clinical model and the radiomics model. The score of the proposed model showed significant differences between HER2-positive and -negative expression (p < 0.001).

Conclusions: These results demonstrate that ultrasound images are predictive of HER2 expression through a deep learning classifier. Our method provides a non-invasive, simple, and feasible method for the prediction of HER2 expression without the manual delineation of the regions of interest (ROI). The performance of our deep learning model significantly exceeded the traditional texture analysis based on the radiomics model.

Introduction

Human epidermal growth factor receptor 2 (HER2) is an important biomarker and a target in the therapy used in approximately 30% of breast cancer patients (1, 2). Although HER2-enriched cancers may have a worse prognosis, they can be effectively treated with therapies targeting HER2 protein, such as Herceptin (chemical name: trastuzumab), Perjeta (chemical name: pertuzumab), and Kadcyla (chemical name: T-DM1 or ado-trastuzumab emtansine) (3). Breast cancer molecular subtypes are categorized in clinical practice by immunohistochemical markers.

The recent literature shows that radiomics features extracted from medical images may predict patient outcomes (46). Breast cancer diagnosis in clinical practice is performed using a type of radiation-free medical imaging approach, and ultrasound imaging plays a significant role (710). The association of peritumoral radiomics features extracted from magnetic resonance imaging (MRI) and the expression of HER2 was established (11).

In recent years, besides the development of compressed sensing (12, 13), wavelet transform (14), and dictionary learning techniques (1517), deep learning approaches have become popular in the field of medical image processing with the development of optimization techniques and the improvement in computational devices (18). The deep learning method-based classification has a positive impact in precision medicine, since it can improve the effectiveness of computer-assisted clinical and radiological decision (19). Existing literature describes the use of the deep learning method to predict medical targets, such as EGFR mutation status in lung cancer (20), and recurrence in high-grade serous ovarian cancer (21, 22).

Deep learning automatically generates the representations that are expressed in terms of other, simpler representations through gradient descent and back-propagation. The abstract mapping from the raw data to the target label is built as a training result (23). DenseNets developed for image tasks have several advantages: avoid the vanishing-gradient, reuse features, and reduce the number of parameters (24). DenseNet (24) exceeds AlexNet (25), GoogLeNet (26), VGG (27), and ResNet (28) in the ImageNet classification task.

In this study, a dense-block-based deep learning model was developed to predict HER2 expression based on preoperative ultrasound images. This proposed method like other supervised deep learning models is an end-to-end workflow. The model requires only a rectangle region of the tumor without the precise delineation of the tumor boundary or human-defined features, while conventional radiomics methods depend on feature engineering. The interobserver error can be reduced and the time for manual segmentation can be saved through our method. The proposed deep learning model can automatically learn HER2 expression features from ultrasound images through back-propagation and optimization algorithm (23). An ultrasound image dataset collected from the Shandong Cancer Hospital and Institute was provided to train and evaluate our deep learning model.

Material and Methods

This work used a DenseNet-based deep learning model to predict breast cancer molecular subtypes from the ultrasound images. The workflow is shown in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1 Overall structure of the developed DenseNet-based deep learning classifier. Being fed a tumor image, the deep learning model predicts the probability of the expression of HER2.

Patients

This retrospective study was approved by the ethics review board. Preoperative ultrasound images of 144 patients were collect from the Shandong Cancer Hospital and Institute. The collected ultrasound images were obtained by an experienced radiologist using a broadband 42–46-Hz transducer (Philips Healthcare System, Amsterdam, Netherlands). Most of the images were cross-section images, the remaining were longitudinal sections. For consistency, only cross-section ultrasound images were used. The whole dataset was randomly divided into a training set and a validation set through the hold-out method. The training set and the validation set were mutually exclusive.

In clinical practice, the molecular subtypes can suggest candidate drugs for the treatment of these patients (29). Immunohistochemistry (IHC) is the most common clinical approach for immunostaining. Since IHC can accurately identify the molecular subtypes of breast cancer by high specificity, in this work, the molecular subtype were identified by IHC. The resulting score of 0, 1+, 2+, and 3+ in the IHC staining represented the amount of HER2 protein on the surface of the cells in a breast cancer tissue sample. The score 0 or 1+ indicate “HER2 negative.” The score 2+ is called “borderline.” The score 3+ indicates “HER2 positive.” If the result of IHC is uncertain, the fluorescence in situ hybridization (FISH) was carried out.

Our inclusion criteria of the data were as follows: (1) pathologically confirmed breast cancer; (2) available preoperative ultrasound image data; (3) pathological IHC examination of tumor specimens; and (4) no history of preoperative therapy. Our exclusion criteria were as follows: (1) ultrasound images were too ambiguous to be analyzed and (2) invasive biopsy was performed before the ultrasound examination.

A rectangle region of interest (ROI) containing the entire tumor was manually selected by radiologists. The ROI was enough due to the strong capability of the information extraction of the deep learning model. Consequently, the precise drawing of the tumor border was not necessary.

Development of the Deep Learning Model

In comparison with previous popular network architectures, DenseNets leverage shortcut connections enhance the information flow to provide better effectiveness. The shortcut connection can be defined as follows:

xk=Hk([x0, x1,, xk1])(1)

where [x0, x1,…, xk−1] refers to the concatenation of the deduced feature maps in the kth layers (24).

Before training the network, several data preprocessing procedures were carried out: ROI selection, image cropping, and image resizing. In each ultrasound image, a rectangle region containing the whole tumor and the tumor borderline was selected. Although tumors have different sizes, all ROI including tumors were scaled to the same size (64 × 64 pixels) by bilinear interpolation before being fed into the network.

The structure of the model was designed in an attempt to achieve better results. Our proposed network was composed of three dense blocks. Two types of dense block were present among these three dense blocks (see Figure 1) in our network. Block type 1 contained 4 layers, while block type 2 contained 32 layers. Both the two types of dense block employed shortcut connections from each layer to all subsequent layers. The details of the two types of dense blocks are shown in Figure 1. The detailed structure of the entire network is indicated in Supplementary Table 1.

The deep learning model was implemented based on the TensorFlow (30) framework and Python 3.5 (31). The trained model gradually becomes stable as the batch size increases, resulting in less overfitting. The weighting coefficient for the classification was adjusted for the imbalance of the classes. Weighting cross-entropy was used as the loss function in our implementation. This approach could help us avoid downsampling or upsampling of the original data; thus, our data distribution was close to the real clinical data. The weight coefficient was tuned, and then a series of experiments were performed. The best configuration was related to the label distribution of the training data. The detailed parameter setting for training the model is indicated in Supplementary Table 2.

Visual Analysis of the Model

The shallow convolutional layer learned low-level simple features such as the horizontal and diagonal edges. A deeper convolutional layer learned more complex features such as tumor shape. The features learned by the low-level layers were intuitive, while the learned features became more abstract with the layers deepening and could gradually be related to the molecular subtypes.

The class activation map method was used to generate an attention map of the trained model for visualization (32, 33). This method helped to visualize and highlight the discriminative image parts detected by the feature extractor, which contributed to the predicted class scores on any given image. The examples of attention map are shown in Figure 2. The positive filter tended to focus on the boundary of the tumor or the high echo region. In the HER2 case, the positive filter indicated the HER2+ category, while the negative filter corresponded to the HER2- category. The positive filter needed to collect more information from a larger area to make a decision than the negative filter.

FIGURE 2
www.frontiersin.org

Figure 2 Class activation heat map: the attention map of the trained model for predicting HER2 expression.

Statistical Analysis

Statistical analysis was performed using a Python language-based toolkit including SciPy (34), scikit-learn (35), and WORC packages. The Mann–Whitney U-test (36) was used to compare the age difference, while a chi-square test (37) was used to compare the difference in other factors. The DeLong test was used to evaluate the difference of the receiver operating characteristic (ROC) curves among different models. A p-value < 0.05 was considered statistically significant.

Results

Clinical Characteristics of the Patients

The clinical characteristics of the patients are listed in Table 1. No significant difference was found between the training and validation cohorts in terms of age, TNM stages, and BI-RADS. These clinical characteristics were also used to build a clinical model for the comparison with the proposed deep learning model.

TABLE 1
www.frontiersin.org

Table 1 Clinical characteristics of patients in the primary and validation cohorts.

Prediction Performance of the Proposed Deep Learning Model

A 3-dense-block-based deep learning model using preoperative ultrasound images was proposed in this study to predict HER2 expression in patients with breast cancer. Our deep learning model showed promising results of accurate predictions. The DL model achieved an AUC of 0.87 in the training cohort (accuracy = 85.19%, sensitivity = 75.53%, specificity = 90.54%, PPV = 78.12%, NPV = 88.16%) and AUC of 0.84 in the validation cohort (accuracy = 80.56%, sensitivity = 72.73%, specificity = 84.00%, PPV = 66.67%, NPV = 87.5%). The result of the experiment allowed us to conclude that the performance of the deep learning model significantly exceeded the traditional radiomics model. Moreover, the deep learning score between HER2+ and HER2- type groups in the training cohort and validation cohort was significantly different (p < 0.01; Figure 3). A radiomics model was also built for comparison to predict the Luminal type. The PyRadiomics toolkit was used to extract image features, and then six features were selected by the recursive feature elimination. Finally, a random forest including 90 trees was built in the radiomics model for prediction. Deep learning features were extracted from the last convolutional layer (global average pool) for cluster analysis (see Figure 4). The clustering figure suggested that the deep learning features have different responses to positive and negative cases.

FIGURE 3
www.frontiersin.org

Figure 3 Deep learning model score HER2 classifier.

FIGURE 4
www.frontiersin.org

Figure 4 Cluster analysis of deep learning features.

Comparison Between the Deep Learning Model and Other Methods

A clinical model and a radiomics model were built as a comparison to the proposed deep learning model. The clinical model considered age, stage, and BI-RADS as features and employed a support vector machine as the classifier. In the radiomics model, 961 features were extracted through the PyRadiomics toolkit. A random forest classifier was built for the prediction of HER2 expression in the radiomics model.

The quantitative effectiveness is shown in Table 2, and the ROC curves are shown in Figure 5, which suggested that our proposed deep learning model significantly exceeded the clinical model (AUC = 0.55, accuracy = 68.52%, sensitivity = 52.94%, specificity = 75.68% in the training set; AUC = 0.51, accuracy = 63.89%, sensitivity = 54.55%, specificity = 68.02% in the testing set; p < 0.05) and the radiomics model (AUC = 0.78, accuracy = 71.29%, sensitivity = 55.88%, specificity = 78.38% in the training set; AUC = 0.74, accuracy = 72.22%, sensitivity = 72.72%, specificity = 72.00% in the testing set; p < 0.05). The confusion matrix shown in Figure 6 reveals that the deep learning model achieved a lower confusion degree in comparison with the clinical model and radiomics model.

TABLE 2
www.frontiersin.org

Table 2 Predictive performance of each model for HER2.

FIGURE 5
www.frontiersin.org

Figure 5 The receiver operating characteristic curve (ROC) of the HER2 on the training set and the testing set.

FIGURE 6
www.frontiersin.org

Figure 6 Confusion matrix: (A) clinical model; (B) radiomics model; (C) DL model.

Discussion

This work proposed a DenseNet-based deep learning model to predict HER2 expression in patients with breast cancer through preoperative non-invasive ultrasound images. The deep learning model was trained in the training cohort, which included 108 patients and was validated in the validated cohort, which included 36 patients. The proposed model was highly effective in the training cohort (accuracy = 85.79%, AUC = 0.87) and the validation cohort (accuracy = 80.56%, AUC = 0.84), exceeding the clinical model and radiomics model. The related tumor area representing HER2 expression status could be obtained by our model using the class activation map.

HER2 is a critical biomarker and its expression helps to make personalized treatments for breast cancer patients. Patients whose HER2 is positive should receive trastuzumab (marketed as Herceptin) which is effective only in cancers where HER2 is overexpressed (38). In clinical practice, IHC is widely used to evaluate the expression of HER2. HER2 expression is positive when the result of IHC is 3+, while HER2 is negative when IHC is 0 or 1+. In 2+ cases by IHC, fluorescence in situ hybridization (FISH) should be employed to confirm the final expression of HER2 (39). However, the IHC and FISH methods require an invasive approach to collect a sample and they are time-consuming. Due to the possibility of positional deviation, an invasive biopsy may fail, and wrong results may be obtained. The prediction of HER2 through preoperative ultrasound images using deep learning could compensate for the above lack.

Recently, the texture analysis-based radiomics method has been used for the diagnosis of the breast cancer (40). Before building a predictive model, ROI must be delineated by radiologists, and then texture features should be manually extracted. However, the delineation of the tumor boundary influences the extracted feature values. The deep learning model needs only an approximate rectangle ROI of the tumors rather than the accurate delineation of the boundaries compared to the conventional texture analysis-based radiomics which requires feature engineering (41). The proposed deep learning model with a multi-block structure and shortcut connections extracts features from raw image pixels to abstract maps without time-consuming handcrafted feature engineering. The model takes raw ultrasound images as input and then predicts HER2 expression.

Despite the promising effectiveness of the proposed deep learning method, this study has some limitations. First, the ultrasound images to build the model were collected from only one manufacture (Philips). Ultrasound signals emitted from different transducers produced by different manufacturers may lead to distinct image features. Hence, building a more general model should be considered in the future. Second, only one type of ultrasound image was used to build the model. In the future, the feature concatenation of convolutional operation in the neural network should be explored to build a two-branch model. Other types of images such as the color Doppler ultrasound or mammography may be considered for the two-modal model to increase the predictive performance. The combination of the deep learning-based tumor auto-detection and deep learning-based radiomics will be considered in the future to obtain a complete clinical diagnostic software.

Conclusions

The above results demonstrate that features of pretreatment ultrasound images are related to HER2 expression. Our proposed deep learning model significantly exceeded the traditional texture analysis-based radiomics model. Our method without manual delineation of ROI is non-invasive, simple, and feasible.

Data Availability Statement

The original contributions presented in the study are included in the article/supplementary material Further inquiries can be directed to the corresponding author.

Ethics Statement

Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author Contributions

ZX contributed to the study design, data acquisition, data analysis, data interpretation, software development, and manuscript drafting. QY contributed to the study design and data acquisition. ML contributed to the manuscript drafting. JG contributed to the manuscript drafting. CD contributed to the manuscript drafting. YC contributed to the data analysis and data interpretation. BL contributed to the study concept, study design, data acquisition, data analysis, data interpretation, and manuscript drafting. He is the PI of the study and oversaw the entirety of the project. All authors contributed to the article and approved the submitted version.

Funding

This paper was supported in part by the National Key Research and Development Program of China (2016YFC0105106), in part by the academic promotion program of Shandong First Medical University (2019LJ004), in part by the National Natural Science Foundation of China (81530060), in part by the State’s Key Project of Research and Development Plan under Grant 2017YFC0109202 and Grant 2017YFA0104302, and in part by the National Natural Science Foundation under 61871117.

Conflict of Interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary Material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2022.829041/full#supplementary-material

References

1. Mitri Z, Constantine T, O’Regan R. The HER2 Receptor in Breast Cancer: Pathophysiology, Clinical Use, and New Advances in Therapy. Chemother Res Pract (2012) 2012:743193–200. doi: 10.1155/2012/743193

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Yin L, Duan JJ, Bian XW, Yu SC. Triple-Negative Breast Cancer Molecular Subtyping and Treatment Progress. Breast Cancer Res (2020) 22:1–13. doi: 10.1186/s13058-020-01296-5

CrossRef Full Text | Google Scholar

3. Quartino AL, Li H, Kirschbrown WP, Mangat R, Wada DR, Garg A, et al. Population Pharmacokinetic and Covariate Analyses of Intravenous Trastuzumab (Herceptin R ), a HER2-Targeted Monoclonal Antibody, in Patients With a Variety of Solid Tumors. Cancer Chemother Pharmacol (2019) 83:329–40. doi: 10.1007/s00280-018-3728-z

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Shang W, Zeng C, Du Y, Hui H, Liang X, Chi C, et al. Core–shell Gold Nanorod@ Metal–Organic Framework Nanoprobes for Multimodality Diagnosis of Glioma. Adv Mater (2017) 29:1604381. doi: 10.1002/adma.201604381

CrossRef Full Text | Google Scholar

5. Lv R, Yang P, Hu B, Xu J, Shang W, Tian J. In Situ Growth Strategy to Integrate Up-Conversion Nanoparticles With Ultrasmall Cus for Photothermal Theranostics. ACS Nano (2017) 11:1064–72. doi: 10.1021/acsnano.6b07990

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Huang YQ, Liang CH, He L, Tian J, Liang CS, Chen X, et al. Development and Validation of a Radiomics Nomogram for Preoperative Prediction of Lymph Node Metastasis in Colorectal Cancer. J Clin Oncol (2016) 2157–64. doi: 10.1200/JCO.2015.65.9128

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Kornecki A. Current Status of Breast Ultrasound. Can Assoc Radiol J (2011) 62:31–40. doi: 10.1016/j.carj.2010.07.006

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Jensen JA. Medical Ultrasound Imaging. Prog Biophys Mol Biol (2007) 93:153–65. doi: 10.1016/j.pbiomolbio.2006.07.025

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Wu T, Sultan LR, Tian J, Cary TW, Sehgal CM. Machine Learning for Diagnostic Ultrasound of Triplenegative Breast Cancer. Breast Cancer Res Treat (2019) 173:365–73. doi: 10.1007/s10549-018-4984-7

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Cui H, Zhang D, Peng F, Kong H, Guo Q, Wu T, et al. Identifying Ultrasound Features of Positive Expression of Ki67 and P53 in Breast Cancer Using Radiomics. Asia Pac J Clin Oncol (2020) e176–84. doi: 10.1111/ajco.13397

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Braman N, Prasanna P, Whitney J, Singh S, Beig N, Etesami M, et al. Association of Peritumoral Radiomics With Tumor Biology and Pathologic Response to Preoperative Targeted Therapy for Her2 (Erbb2)–Positive Breast Cancer. JAMA Netw Open (2019) 2:e192561–e192561. doi: 10.1001/jamanetworkopen.2019.2561

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Donoho DL. Compressed Sensing. IEEE Trans Inf Theory (2006) 52:1289–306. doi: 10.1109/TIT.2006.871582

CrossRef Full Text | Google Scholar

13. Yu H, Wang G. Compressed Sensing Based Interior Tomography. Phys Med Biol (2009) 54:2791. doi: 10.1088/0031-9155/54/9/014

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Garduno E, Herman GT, Davidi R. Reconstruction From a Few Projections by ?1-Minimization of The˜ Haar Transform. Inverse Probl (2011) 27:055006. doi: 10.1088/0266-5611/27/5/055006

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Liu J, Hu Y, Yang J, Chen Y, Shu H, Luo L, et al. 3D Feature Constrained Reconstruction for Low-Dose CT Imaging. IEEE Trans Circuits Syst Video Technol (2016) 28:1232–47. doi: 10.1109/TCSVT.2016.2643009

CrossRef Full Text | Google Scholar

16. Liu J, Ma J, Zhang Y, Chen Y, Yang J, Shu H, et al. Discriminative Feature Representation to Improve Projection Data Inconsistency for Low Dose Ct Imaging. IEEE Trans Med Imaging (2017) 36:2499–509. doi: 10.1109/TMI.2017.2739841

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Chen Y, Shi L, Feng Q, Yang J, Shu H, Luo L, et al. Artifact Suppressed Dictionary Learning for Low-Dose Ct Image Processing. IEEE Trans Med Imaging (2014) 33:2271–92. doi: 10.1109/TMI.2014.2336860

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Yin X, Zhao Q, Liu J, Yang W, Yang J, Quan G, et al. Domain Progressive 3D Residual Convolution Network to Improve Low-Dose Ct Imaging. IEEE Trans Med Imaging (2019) 38:2903– 2913. doi: 10.1109/TMI.2019.2917258

PubMed Abstract | CrossRef Full Text | Google Scholar

19. Parekh VS, Jacobs MA. Deep Learning and Radiomics in Precision Medicine. Expert Rev Precis Med Drug Dev (2019) 4:59–72. doi: 10.1080/23808993.2019.1585805

PubMed Abstract | CrossRef Full Text | Google Scholar

20. Wang S, Shi J, Ye Z, Dong D, Yu D, Zhou M, et al. Predicting EGFR Mutation Status in Lung Adenocarcinoma on Computed Tomography Image Using Deep Learning. Eur Respir J (2019) 53:1800986. doi: 10.1183/13993003.00986-2018

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Han S, Kang HK, Jeong JY, Park MH, Kim W, Bang WC, et al. A Deep Learning Framework for Supporting the Classification of Breast Lesions in Ultrasound Images. Phys Med Biol (2017) 62:7714–28. doi: 10.1088/1361-6560/aa82ec

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Wang S, Liu Z, Rong Y, Zhou B, Bai Y, Wei W, et al. Deep Learning Provides a New Computed Tomography-Based Prognostic Biomarker for Recurrence Prediction in High-Grade Serous Ovarian Cancer. Radiother Oncol (2019) 132:171–7. doi: 10.1016/j.radonc.2018.10.019

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Goodfellow I, Bengio Y, Courville A. Deep Learning. Cambridge: MIT Press (2016).

Google Scholar

24. Huang G, Liu Z, van der Maaten L, Weinberger KQ. Densely Connected Convolutional Networks. Proc IEEE Conf Comput Vis Pattern Recognit (2017) 4700–8. doi: 10.1109/CVPR.2017.243

CrossRef Full Text | Google Scholar

25. Krizhevsky A, Sutskever I, Hinton GE. Imagenet Classification With Deep Convolutional Neural Networks. Adv Neural Inf Process Syst (2012) 1097–105. doi: 10.1145/3065386

CrossRef Full Text | Google Scholar

26. Szegedy C, Liu W, Jia Y, Sermanet P, Reed S, Anguelov D, et al. Going Deeper With Convolutions. Proc IEEE Conf Comput Vis Pattern Recognit (2015) 1–9. doi: 10.1109/CVPR.2015.7298594

CrossRef Full Text | Google Scholar

27. Simonyan K, Zisserman A. Very Deep Convolutional Networks for Large-Scale Image Recognition. arXiv preprint arXiv:1409.1556 (2014).

Google Scholar

28. He K, Ren S, Sun J, Zhang X. Deep Residual Learning for Image Recognition. Proc IEEE Conf Comput Vis Pattern Recognit (2016) 770–8. doi: 10.1109/CVPR.2016.90

CrossRef Full Text | Google Scholar

29. Nielsen TO, Hsu FD, Jensen K, Cheang M, Karaca G, Hu Z, et al. Immunohistochemical and Clinical Characterization of the Basal-Like Subtype of Invasive Breast Carcinoma. Clin Cancer Res (2004) 10:5367–74. doi: 10.1158/1078-0432.CCR-04-0220

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Software Google. Tensorflow. (2021).

Google Scholar

31. Software Python Software Foundation. Python. (2021).

Google Scholar

32. Zhou B, Khosla A, Lapedriza A, Oliva A, Torralba A. Learning Deep Features for Discriminative Localization. Proc IEEE Conf Comput Vis Pattern Recognit (2016) 2921–9. doi: 10.1109/CVPR.2016.319

CrossRef Full Text | Google Scholar

33. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-Cam: Visual Explanations From Deep Networks via Gradient-Based Localization. Proc IEEE Int Conf Comput Vis (2017) 618–26. doi: 10.1109/ICCV.2017.74

CrossRef Full Text | Google Scholar

34. Bressert E. Scipy and Numpy: An Overview for Developers. Sebastopol: O’REILLY (2012).

Google Scholar

35. Pedregosa F, Varoquaux G, Gramfort A, Michel V, Thirion B, Grisel O, et al. Scikit-Learn: Machine Learning in Python. J Mach Learn Res (2011) 12:2825–30.

Google Scholar

36. McKnight PE, Najab J. Mann-Whitney U Test. In: The Corsini Encyclopedia of Psychology. Hoboken: Wiley. (2010). p. 1–1.

Google Scholar

37. McHugh ML. The Chi-Square Test of Independence. Biochem Med (2013) 23:143–9. doi: 10.11613/BM.2013.018

CrossRef Full Text | Google Scholar

38. Van Steenhoven JE, Kuijer A, van Maaren MC, Roos M, Elias SG, van Diest PJ, et al. Quantifying the Mitigating Effects of Whole-Breast Radiotherapy and Systemic Treatments on Regional Recurrence Incidence Among Breast Cancer Patients. Ann Surg Oncol (2020) 27:1–10. doi: 10.1245/s10434-020-08356-2

CrossRef Full Text | Google Scholar

39. Ellis C, Dyson M, Stephenson T, Maltby E. Her2 Amplification Status in Breast Cancer: A Comparison Between Immunohistochemical Staining and Fluorescence in Situ Hybridisation Using Manual and Automated Quantitative Image Analysis Scoring Techniques. J Clin Pathol (2005) 58:710–4. doi: 10.1136/jcp.2004.023424

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Conti A, Duggento A, Indovina I, Guerrisi M, Toschi N. Radiomics in Breast Cancer Classification and Prediction. Semin Cancer Biol (2020) 238–50. doi: 10.1016/j.semcancer.2020.04.002

PubMed Abstract | CrossRef Full Text | Google Scholar

41. Krizhevsky A, Sutskever I, Hinton GE. Imagenet Classification With Deep Convolutional Neural Networks. Commun ACM (2017) 60:84–90. doi: 10.1145/3065386

CrossRef Full Text | Google Scholar

Keywords: breast cancer, ultrasound, deep learning, DenseNet, human epidermal growth factor receptor 2

Citation: Xu Z, Yang Q, Li M, Gu J, Du C, Chen Y and Li B (2022) Predicting HER2 Status in Breast Cancer on Ultrasound Images Using Deep Learning Method. Front. Oncol. 12:829041. doi: 10.3389/fonc.2022.829041

Received: 04 December 2021; Accepted: 24 January 2022;
Published: 16 February 2022.

Edited by:

Jun Shen, Sun Yat-sen University, China

Reviewed by:

Ze-Hong Yang, Sun Yat-sen University, China
Zi-Xian Wang, Sun Yat-sen University Cancer Center (SYSUCC), China

Copyright © 2022 Xu, Yang, Li, Gu, Du, Chen and Li. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Baosheng Li, YnNsaUBzZGZtdS5lZHUuY24=

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.

Research integrity at Frontiers

Man ultramarathon runner in the mountains he trains at sunset

95% of researchers rate our articles as excellent or good

Learn more about the work of our research integrity team to safeguard the quality of each article we publish.


Find out more