- 1Department of Medical Imaging Center, The First Affiliated Hospital, Jinan University, Guangzhou, China
- 2Department of Ultrasound, Remote Consultation Center of ABUS, The Second Affiliated Hospital, Guangzhou University of Chinese Medicine, Guangzhou, China
- 3Medical Ultrasound Image Computing Lab, Shenzhen University, Shenzhen, China
Objective: This study aimed to evaluate a convolution neural network algorithm for breast lesion detection with multi-center ABUS image data developed based on ABUS image and Yolo v5.
Methods: A total of 741 cases with 2,538 volume data of ABUS examinations were analyzed, which were recruited from 7 hospitals between October 2016 and December 2020. A total of 452 volume data of 413 cases were used as internal validation data, and 2,086 volume data from 328 cases were used as external validation data. There were 1,178 breast lesions in 413 patients (161 malignant and 1,017 benign) and 1,936 lesions in 328 patients (57 malignant and 1,879 benign). The efficiency and accuracy of the algorithm were analyzed in detecting lesions with different allowable false positive values and lesion sizes, and the differences were compared and analyzed, which included the various indicators in internal validation and external validation data.
Results: The study found that the algorithm had high sensitivity for all categories of lesions, even when using internal or external validation data. The overall detection rate of the algorithm was as high as 78.1 and 71.2% in the internal and external validation sets, respectively. The algorithm could detect more lesions with increasing nodule size (87.4% in ≥10 mm lesions but less than 50% in <10 mm). The detection rate of BI-RADS 4/5 lesions was higher than that of BI-RADS 3 or 2 (96.5% vs 79.7% vs 74.7% internal, 95.8% vs 74.7% vs 88.4% external). Furthermore, the detection performance was better for malignant nodules than benign (98.1% vs 74.9% internal, 98.2% vs 70.4% external).
Conclusions: This algorithm showed good detection efficiency in the internal and external validation sets, especially for category 4/5 lesions and malignant lesions. However, there are still some deficiencies in detecting category 2 and 3 lesions and lesions smaller than 10 mm.
Introduction
Breast cancer is the most common cancer and a leading cause of cancer death in women worldwide, but precise detection can provide an opportunity for timely treatment (1). Among the various detection methods, B-mode ultrasound screening technology is favored and recommended as a routine diagnostic tool because of its low cost and rapid imaging (2). Although breast ultrasound imaging can characterize the suspicious tumor area of the breast tissue, it has high technical dependence on the operator, poor diagnostic repeatability, long time-consuming and low accuracy, and massive daily image analysis aggravates the burden of clinical radiologists (3). Furthermore, the inconsistency of different radiologists on the same image may lead to severe false-positive problems, thereby delaying effective treatments (4).
Automatic breast ultrasound (ABUS) imaging has become an essential tool in breast cancer diagnosis. ABUS is considered to have high repeatability, low operator dependence, less time consuming by radiologists in image acquisition, automatic three-dimensional reconstruction of the whole breast, coronal information, and a relatively wide observation field. Studies have shown that mammography (MG) plus ABUS examination can increase the detection rate of breast cancer in women with dense breasts, particularly the detection rate of small lesions (5). A multi-center study on Chinese women showed that ABUS had good reliability compared with handhold ultrasound (HHUS) and MG (6). The other study conducted in the United States showed that it could help improve the detection rate of breast cancer by adding ABUS to breast cancer screening (7).
Although ABUS has many advantages, it also inevitably aggravates the workload of screening and diagnostic examiners. Different computer-aided diagnosis (CAD) systems have been developed to standardize and accelerate diagnostic procedures (8). Relevant studies suggest that computer-aided diagnosis software could effectively improve the detection of lesions and the speed of diagnosis (9). However, in breast cancer imaging research, deep learning (AI or computer-aided diagnosis CAD) has mainly focused on mammography or ultrasound 2D/3D imaging combined with deep learning (10–12). In lesion detection and diagnosis, accurate segmentation of the breast mass in a 3D ABUS image is an essential task in ABUS image analysis. It also plays a vital role in designing a computer-aided detection or diagnostic system (13–15). Recently, deep learning techniques have made significant progress in medical image segmentation (16–18). The convolution neural network (CNN) has become a promising choice in breast ultrasound image segmentation (19–22).
The diagnostic process of breast lesions includes detection, diagnosis, and treatment. Lesion detection is the premise of this diagnosis. A single ABUS volume image could have 320 frames with a layer thickness of 0.5 mm. The amount of data in ABUS images is more than that in most natural images. The cost of manually marking breast lumps is high. Direct training of large-scale segmentation networks with tens of millions of parameters may introduce potential over-fitting (23). This study aimed to evaluate a convolution neural network algorithm for breast disease detection, developed based on ABUS image and Yolo v5. The efficiency and accuracy of the algorithm were analyzed in detecting lesions with different allowable false positive values and lesion sizes, and the differences were compared and analyzed, which included the various indicators in internal validation and external validation data.
Materials and Methods
Patients
This multi-center retrospective study was conducted following the Declaration of Helsinki and was approved by the local institutional review board (ZE2020-232). A total of 32,493 cases of ABUS examination were analyzed, which were collected from 7 hospitals (GDHCM, BHLY, CFFIC, PHHKS, YDHSZ, LKMC, and JBMH) between October 2016 and December 2020. The inclusion criteria of this study were: ① The quality control requirements of ABUS were met; ② Malignant lesions were confirmed by pathology; ③ The clinical information of the case was complete; and ④ benign lesions required a biopsy or follow-up for more than 2 years by ultrasound, mammography or MRI. The exclusion criteria in this study were: ① the quality control requirements of the image were not met; ② the patient had a history of breast trauma, surgery, mastitis, etc.; ③ suspected malignant tumor without pathological results; ④ benign lesions failed to follow up as required; and ⑤ other conditions could affect diagnosis. Finally, 30,515 cases were excluded from the study.
All training and test data were from the Guangdong Province Hospital of Chinese Medicine because each case needed a bilateral breast examination, and some lesions could be displayed in two or three-volume data. The training set collected 3,457 ABUS volume data from 936 cases (age range 18–70 years, average 41.54 ± 11.22 years), 221 malignant lesions and 3,662 benign lesions, and 791 confirmed lesions by pathology. Another 1,406 ABUS volume data of 301 cases (age range 26–69 years, average 42.37 ± 12.48 years) were used for validation, including 57 malignant lesions and 885 benign lesions, and 247 lesions were confirmed by pathology.
The other 741 cases with 2,538 volume data were included in the validation data. A total of 452 volume data of 413 cases (age range 23–67 years, average 41.21 ± 11.93 years) were used as internal validation data (IVD), including 161 malignant lesions and 1,017 benign lesions. A total of 214 lesions of BI-RADS 4 or 5 and 47 lesions of BI-RADS 3 were confirmed by pathology. The other lesions were followed up to rule out the possibility of malignancy. A total of 328 cases (age range 22–65 years, average 40.32 ± 13.44 years) with 1,086 volume data were used as external validation data (EVD), which included 57 malignant lesions and 1,879 benign lesions. A total of 96 lesions of BI-RADS 4 or 5 and 29 lesions of BI-RADS 3 were confirmed by pathology, and the other lesions were followed up as required. All cases with dense breasts were collected in this study (Figure 1).
Quality Control and Reference Standard
All cases were scanned according to the scanning specifications recommended by the ABUS use manual. The standard images of unilateral breast scanning are AP, LAT, and MED. SUP and INF can be added when the breast is large. So each case had 4 to 8 volumes of data. The quality control standard recommended by the GE ABUS use manual was used for quality control, and the volume data with lesions were collected. In this study, two doctors (XW and YD) with more than 5 years of experience in ABUS were used for quality control.
Two senior radiologists (ZS and CL, with 10 and 15 years of experience in breast ultrasound diagnosis) confirmed all lesions regarding pathological or follow-up results. Cases with differences were judged by another senior radiologist (ZS, with 25 years of experience in breast ultrasound diagnosis).
Methods and Data Annotation
The algorithm (Volume-Breast Ultrasound Intelligent Lesion Detection System, V-BUILDS) was confirmed to have a data enhancement strategy of Mosaic data enhancement ratio of 0.5 and a mixed data enhancement ratio of 0, which was based on Yolo V5. It adopts WBF for model fusion, adds the detection model of the transformer encoder module method, and the algorithm obtained the 3D RESNET reducing false-positive method to detect, verify, and compare the internal and external data. The YOLOv5 classifiers were trained using the open-source Python library Image AI. In the development of the algorithm, the training and testing set were applied to the training and testing of this algorithm. The flowchart is shown in Figure 2.
Figure 2 The detection network architecture of our proposed framework. It included a detection model and a three-dimensional false positive reduction model. The ABUS volume was sliced along the cross-section in the detection model training to obtain a two-dimensional image. Four training set images were randomly selected in each study. The improved Mosaic data enhancement method was input into the network for training to develop the three-dimensional false positive reduction model. This model took the lesion as the center, cut the tumour region, inputted it into the three-dimensional classification network, and classified the multi-scale features of the classification network after ROI pooling. In the stage of network reasoning, the slice data of volume are input into the network in turn, and the detection frames of adjacent slices are combined through NMS to obtain a three-dimensional detection frame. According to the three-dimensional detection frame, the ROI area was cut from the original volume data and inputted into the false positive reduction network to obtain the probability that the location was a lesion.
In this study, PAIR was used to annotate ABUS images (PAIR was a multifunctional labeling software developed by the Medical Ultrasonic Image Computing Lab Music of the medical department of Shenzhen University based on C++). The pair has the advantages of supporting multiple data formats, labeling tasks, custom feature attributes, integrating deep learning semi-automatic labeling, and ensuring data security.
Equipment and Computation Platform
All images were acquired by INVENIA ABUS 1.0 (model 5500-4400-01, GE Healthcare, USA), using a C15-6 × W arc probe with a central frequency of 10 MHz. The examination depth was adjusted according to the size of the breast volume of the patient. The pixel size of the ABUS images was 0.27 × 0.27 × 0.5 mm.
The proposed method was implemented on an NVIDIA RT × 2080TiCPU, Intel(R) × eon(R) Silver 4210 CPU, PyTorch1. 7.0. The model construction of fast RCNN, Retinanet, and Fcos was carried out on the detectron2 framework.
Classification Performance Evaluation and Statistical Analysis
To evaluate the performance of the breast lesion detection model, accuracy (ACC), sensitivity (SEN), specificity (SPE), false positive (FPS), negative–positive (NPS), positive predictive value (PPV), negative predictive value (NPV), and Youden index (Yi) were evaluated by F1 score (F1). An independent sample t-test was used for inter-group comparison, and the rank-sum test (Mann–Whitney U test) was used for those who did not obey normal distribution or uneven variance. Inspection-level α = 0.05 (normality test, α = 0.10).
Additionally, this study also used the FROC curve to study the relationship between model sensitivity and false-positive ratio, in which vertical sitting represents sensitivity, and the abscissa represents the ratio of the number of false-positive lesions to the number of true positive lesions.
Results
Pathology and Follow-Up Results of Different Data Sets
Each data set contained multiple pathological types of lesions confirmed by pathology. As shown in Table 1, invasive carcinoma (non-special type) is the most common breast malignant tumor, while fibroadenoma is the most common benign breast tumor. Most lesions were benign and confirmed after more than 2 years of follow-up.
Lesion Detection in Internal and External Validation Data Based on Different False Positive Values
As shown in Table 2, the sensitivity indicators of different categories of data to false-positive values were analyzed based on the comparison of internal and external validation data. The overall detection rate of the algorithm was as high as 78.1 and 71.2% in the internal and external validation sets. The study showed that 0.5 false-positive values per frame were susceptible to all categories of data, whether internal or external validation data. With the increase in the number of false positives allowed per frame, the detection sensitivity of each lesion also increased slightly. However, the sensitivity was 93.2% for detecting category 4/5 lesions in the internal validation set when 1.5 false positives were allowed per frame. Moreover, when 4 false positives were allowed per frame, the sensitivity was 96.5%. In the external validation set, the detection performance of category 4/5 lesions was similar. When 3 false positives were allowed per frame, it reached its highest value, and the sensitivity was 95.8%. However, the detection of category 2 and 3 lesions in the internal set and category 3 lesions in the internal and external set failed to reach 80%. With the increase in the number of the false positives allowed per frame, the detection sensitivity increased slightly. The detection rate of BI-RADS 4/5 lesions was higher than that of BI-RADS 3 or 2 (96.5% vs 79.7% vs 74.7% internal, 95.8% vs 74.7% vs 88.4% external) (P <0.001). This relationship could be more visually expressed in Figure 3.
Table 2 When false positives (FPS) are allowed in different frames, the detection rates of different BI-RADS categories of lesions in different validation sets (IVD and EVD) were shown.
Figure 3 When false positives(FPS) are allowed in different frames, the detection rates of different BI-RADS categories of lesions in different validation sets were shown in the figure. The detection rate of malignant lesion in different validation set was although shown in the figure.
Analysis of Lesion Size and Missed Diagnosis of Internal and External Validation Data
In this group of cases (Table 3), for the internal validation data set, the non-detection rate of lesions less than 5 mm was 55.6%, and the non-detection rate of 5–10 mm lesions was 25.3%. The non-detection rate of lesions more significant than 10 mm was 12.6%. In the external validation data set, the non-detection rate of lesions less than 5 mm in this algorithm was 61.4%, the non-detection rate of 5–10 mm lesions was 28.8%, and the non-detection rate of lesions more significant than 10 mm was 15.6%.The algorithm could detect more lesions with the increasing nodule size (87.4% in ≥10 mm lesions, but less than 50% in <10 mm). However, there was no difference between the two sets (P >0.05). Figure 4 shows the composition of the size detection relationship of each category.
Table 3 In the internal (IVD) and external validation (EVD) sets, the number of possible benign (category 2 or 3) and suspicious malignant lesions (category 4 or 5) of different sizes and the number of detected lesions of different sizes.
Figure 4 Histogram of different categories of lesions in internal validation data (A) and external validation data (C) sets and lesions of different sizes was shown in the figure. The detection rate of different sizes of lesion in internal validation data (B) and external validation data (D) sets was shown in the figure.
Study on the Detection Rate of Malignant Lesions by Internal and External Validation Data
As shown in Table 4 and Figure 3, the detection rate of malignant lesions was higher than that of category 4/5 lesions (Table 2) in the internal validation set. Under the condition of allowing an average of 0.5 false positives per frame, the current detection rate reaches 91.3%. When three false positives per frame were allowed, the detection rate reached 96.9%. The current detection rate reaches 93.0% in the external verification data when the condition of allowing an average of 0.5 false positives per frame is met. When 1.5 false positives per frame were allowed, the detection rate reached 94.7%, and when 2.5 false positives per frame were allowed, the detection rate reached 98.2%.
Table 4 The detection rate of malignant lesion in different data set was listed when different false positives (FPS) were allowed.
Discussion
To our knowledge, this is the first study to develop the lesion detection algorithm based on ABUS volume data and evaluate its internal and external detection efficiency. In recent years, ABUS has become a popular imaging modality in breast cancer detection and diagnosis because ABUS improves the sensitivity of dense breast detection and can be successfully applied to the visualization and characterization of breast lesions (24, 25). Therefore, the analysis of ABUS images has attracted the attention of more and more attention of radiologists and researchers (9). This study applied the deep learning algorithm (V-BUILDS) to the completed model training and 3D false positive reduction network. Based on the ABUS data, the false-positive allowable values of different frames were externally verified for the data from multiple centers. The differences between internal and external verification and the differences in various indicators such as lesion detection efficiency and accuracy were compared and analyzed. The research showed that this algorithm had high detection efficiency for different categories of lesions, and the total detection efficiency for all lesions was slightly lower. However, it had a high detection efficiency for category 4 and 5 lesions, especially malignant lesions. The detection rate for category 4 and 5 lesions differed from that of category 2 or 3 in each dataset (P <0.001). The detection rate can reach 0.963 when 1.5 false positives per frame are allowed in the internal verification set.
Some cases are difficult to diagnose because of the characteristics of low signal-to-noise ratio, severe artifacts, blurred margin, and the height change of the shape (9), the small volume of the lesion, and the insufficient number of images. Such lesions may occur in different categories of lesions. However, it is difficult to compare with the surrounding tissues in the detection process, which can easily lead to missed detection or misjudgment. These conditions also occurred in lung lesions detected by CNN (26–28). In this study, when the lesion diameter was less than 10 mm, the detection rate of the lesion was low. However, when the diameter of the lesion was greater than 10 mm, the detection rate of the lesion increased significantly with the increase in the diameter of the lesion. These results were consistent with those reported in the literature (29). It is important to point out that the small lesion is usually the early stage of cancer or a benign lesion. Furthermore, it was difficult to detect with all kinds of images. There were a certain proportion of large undetected and some medium-sized lesions in the external validation data. After analyzing the causes, it was found that the edge of the lesion was blurred, the lesion (a benign lesion) was too large (almost occupying the whole breast), and the lesion scope was uneven and resembled the echo of normal gland tissue. The lesion image (pathological result: malignant lesion) was a non-mass breast lesion with an unclear boundary between the scope and surrounding tissues, heterogeneous echo, ductal hypoechoic, or only localized hypoechoic with distortion of surrounding structures.
In the external validation, the detection of category 3 lesions was always less than that of category 2 lesions. The possible reason was that category 2 lesions were mainly manifested as breast cysts (30, 31), so they had a high detection rate. The hyperplasia lesion in category 3 belonged to uncertain lesions without a clear margin (32). The detection of this lesion needs to be determined by normal glands. When there are numerous hyperplasia lesions in the dataset, it is difficult to identify them by an artificial intelligence algorithm. In this study, there was no statistical difference in detecting malignant lesions in different data sets. The BUILDS achieved high detection sensitivity of 91.3% (IVD) and 93.0% (EVD) at 0.5 false positives per scan. The detection rate in IVD was 98.1% at 3.5 false positives per scan, and the detection rate in EVD was 98.2% at 2.5 false positives per scan. Compared to the reported performance of recently published studies on different datasets, this result was comparable to the numbers reported in a recent multi-view convolutional network study (33) and was much better than the recently published 3D CNN lung lesion detection work (34).
In this group of cases, to ensure that the training and testing set data were widely representative, the included cases were universal and often accompanied by multiple lesions. At the same time, there was the coexistence of benign and malignant lesions, which included numerous hyperplasia lesions. The detection rate of type 2 and 3 lesions in this study was lower, but it was significantly higher than that in previous studies (32, 35–37).
There were several limitations to this study. Firstly, there were some limitations in the amount of research data and some differences in image quality in this multi-center retrospective study. Secondly, the different locations of the lesion may affect the detection of the lesion. When the lesion was located at the edge of the image, it may lead to false judgment due to insufficient local pressure or incomplete display of the lesion. Moreover, if the lesion was located behind the nipple, the lesion was often unclear or incomplete because of the acoustic shadow of the nipple. Additionally, some lesions in this group were too small and lacked pixels, which will also lead to detection difficulties.
This algorithm showed good detection efficiency in internal and external validation, especially for category 4/5 lesions and malignant lesions. However, there are still some deficiencies in detecting category 2 and 3 lesions and lesions smaller than 10 mm. This study showed that this algorithm could be an effective auxiliary tool for lesion detection in ABUS.
Data Availability Statement
The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.
Author Contributions
JXZ: Experimental design, organization and implementation, and thesis writing. XT: Algorithm design and verification and literature novelty search. YHJ: Image annotation and data collection. XXW: Image annotation and data collection. DY: Image annotation and data collection. WX: Image annotation and data collection. SLZ: Image annotation and statistical analysis. LC: Image annotation and experimental verification. LPL: Project acquisition and guide implementation. DN: Algorithm guidance. All authors listed have made a substantial, direct, and intellectual contribution to the work and approved it for publication.
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Acknowledgments
The authors thank all the hospitals that provided cases for this study, and all the radiologists and ABUS technicians of the breast remote consultation team of Guangdong Province Hospital of Chinese Medicine. This study was supported by the National Natural Science Foundation of China (No. 62171290), Shenzhen-Hong Kong Joint Research Program (No. SGDX20201103095613036) and the Shenzhen Science and Technology Research and Development Fund for Sustainable Development project (No. KCXFZ20201221173613036).
References
1. Zhou LQ, Wu XL, Huang SY, Wu GG, Ye HR, Wei Q, et al. Lymph Node Metastasis Prediction From Primary Breast Cancer US Images Using Deep Learning. Radiology (2020) 294(1):19–28. doi: 10.1148/radiol.2019190372
2. Sun Q, Lin X, Zhao Y, Li L, Yan K, Liang D, et al. Deep Learning vs. Radiomics for Predicting Axillary Lymph Node Metastasis of Breast Cancer Using Ultrasound Images: Don't Forget the Peritumoral Region. Front Oncol (2020) 10:53. doi: 10.3389/fonc.2020.00053
3. Rajpurkar P, Irvin J, Ball RL, Zhu K, Yang B, Mehta H, et al. Deep Learning for Chest Radiograph Diagnosis: A Retrospective Comparison of the CheXNeXt Algorithm to Practicing Radiologists. PLoS Med (2018) 15(11):e1002686. doi: 10.1371/journal.pmed.1002686
4. Kim YG, Cho Y, Wu CJ, Park S, Jung KH, Seo JB, et al. Short-Term Reproducibility of Pulmonary Nodule and Mass Detection in Chest Radiographs: Comparison Among Radiologists and Four Different Computer-Aided Detections With Convolutional Neural Net. Sci Rep (2019) 9(1):18738. doi: 10.1038/s41598-019-55373-7
5. Brem RF, Tabár L, Duffy SW, Inciardi MF, Guingrich JA, Hashimoto BE, et al. Assessing Improvement in Detection of Breast Cancer With Three-Dimensional Automated Breast US in Women With Dense Breast Tissue: The SomoInsight Study. Radiology (2015) 274(3):663–73. doi: 10.1148/radiol.14132832
6. Hooley RJ, Greenberg KL, Stackhouse RM, Geisel JL, Butler RS, Philpotts LE. Screening US in Patients With Mammographically Dense Breasts: Initial Experience With Connecticut Public Act 09-41. Radiology (2012) 265(1):59–69. doi: 10.1148/radiol.12120621
7. Wilczek B, Wilczek HE, Rasouliyan L, Leifland K. Adding 3D Automated Breast Ultrasound to Mammography Screening in Women With Heterogeneously and Extremely Dense Breasts: Report From a Hospital-Based, High-Volume, Single-Center Breast Cancer Screening Program. Eur J Radiol (2016) 85(9):1554–63. doi: 10.1016/j.ejrad.2016.06.004
8. Hamet P, Tremblay J. Artificial Intelligence in Medicine. Metabolism (2017) 69s:S36–s40. doi: 10.1016/j.metabol.2017.01.011
9. Zhang X, Lin X, Zhang Z, Dong L, Sun X, Sun D, et al. Artificial Intelligence Medical Ultrasound Equipment: Application of Breast Lesions Detection. Ultrason Imaging (2020) 42(4-5):191–202. doi: 10.1177/0161734620928453
10. Schaefgen B, Heil J, Barr RG, Radicke M, Harcos A, Gomez C, et al. Initial Results of the FUSION-X-US Prototype Combining 3D Automated Breast Ultrasound and Digital Breast Tomosynthesis. Eur Radiol (2018) 28(6):2499–506. doi: 10.1007/s00330-017-5235-8
11. Venkatesh SS, Levenback BJ, Sultan LR, Bouzghar G, Sehgal CM. Going Beyond a First Reader: A Machine Learning Methodology for Optimizing Cost and Performance in Breast Ultrasound Diagnosis. Ultrasound Med Biol (2015) 41(12):3148–62. doi: 10.1016/j.ultrasmedbio.2015.07.020
12. Chen DR, Chien CL, Kuo YF. Computer-Aided Assessment of Tumor Grade for Breast Cancer in Ultrasound Images. Comput Math Methods Med (2015) 2015:914091. doi: 10.1155/2015/914091
13. Xi X, Shi H, Han L, Wang T, Ding HY, Zhang G, et al. Breast Tumor Segmentation With Prior Knowledge Learning. Neurocomputing (2016) 237:145–57. doi: 10.1016/j.neucom.2016.09.067
14. Cao X, Chen H, Li Y, Peng Y, Wang S, Cheng L. Dilated Densely Connected U-Net With Uncertainty Focus Loss for 3D ABUS Mass Segmentation. Comput Methods Programs Biomed (2021) 209:106313. doi: 10.1016/j.cmpb.2021.106313
15. Agarwal R, Diaz O, Lladó X, Gubern-Mérida A, Vilanova JC, Martí R. Lesion Segmentation in Automated 3d Breast Ultrasound: Volumetric Analysis. Ultrason Imaging (2018) 40(2):97–112. doi: 10.1177/0161734617737733
16. Ji Y, Zhang R, Li Z, Ren J, Luo P. UXNet: Searching Multi-Level Feature Aggregation for 3D Medical Image Segmentation. arXiv (2020) 32(5):2891–900. doi: 10.1007/978-3-030-59710-8_34
17. Xue Y, Xu T, Zhang H, Long LR, Huang X. SegAN: Adversarial Network With Multi-Scale L(1) Loss for Medical Image Segmentation. Neuroinformatics (2018) 16(3-4):383–92. doi: 10.1007/s12021-018-9377-x
18. Shen T, Gou C, Wang FY, He Z, Chen W. Learning From Adversarial Medical Images for X-Ray Breast Mass Segmentation. Comput Methods Programs Biomed (2019) 180:105012. doi: 10.1016/j.cmpb.2019.105012
19. Xing J, Li Z, Wang B, Qi Y, Yu B, Zanjani FG, et al. Lesion Segmentation in Ultrasound Using Semi-Pixel-Wise Cycle Generative Adversarial Nets. IEEE/ACM Trans Comput Biol Bioinf (2021) 18(6):2555–65. doi: 10.1109/TCBB.2020.2978470
20. Hu Y, Guo Y, Wang Y, Yu J, Li J, Zhou S, et al. Automatic Tumor Segmentation in Breast Ultrasound Images Using a Dilated Fully Convolutional Network Combined With an Active Contour Model. Med Phys (2019) 46(1):215–28. doi: 10.1002/mp.13268
21. Xu Y, Wang Y, Yuan J, Cheng Q, Wang X, Carson PL. Medical Breast Ultrasound Image Segmentation by Machine Learning. Ultrasonics (2019) 91:1–9. doi: 10.1016/j.ultras.2018.07.006
22. Wang Y, Wang N, Xu M, Yu J, Qin C, Luo X, et al. Deeply-Supervised Networks With Threshold Loss for Cancer Detection in Automated Breast Ultrasound. IEEE Trans Med Imaging (2020) 39(4):866–76. doi: 10.1109/TMI.2019.2936500
23. Jing L, Tian Y. Self-Supervised Visual Feature Learning With Deep Neural Networks: A Survey. IEEE Trans Pattern Anal Mach Intell (2021) 43(11):4037–58. doi: 10.1109/TPAMI.2020.2992393
24. Vourtsis A, Kachulis A. The Performance of 3D ABUS Versus HHUS in the Visualisation and BI-RADS Characterisation of Breast Lesions in a Large Cohort of 1,886 Women. Eur Radiol (2018) 28(2):592–601. doi: 10.1007/s00330-017-5011-9
25. van Zelst JCM, Mann RM. Automated Three-Dimensional Breast US for Screening: Technique, Artifacts, and Lesion Characterization. Radiographics (2018) 38(3):663–83. doi: 10.1148/rg.2018170162
26. Pezeshk A, Hamidian S, Petrick N, Sahiner B. 3-D Convolutional Neural Networks for Automatic Detection of Pulmonary Nodules in Chest Ct. IEEE J BioMed Health Inform (2019) 23(5):2080–90. doi: 10.1109/JBHI.2018.2879449
27. Cao H, Liu H, Song E, Ma G, Xu X, Jin R, et al. A Two-Stage Convolutional Neural Networks for Lung Nodule Detection. IEEE J BioMed Health Inform (2020) 24(7):2006–15. doi: 10.1109/JBHI.2019.2963720
28. Nasrullah N, Sang J, Alam MS, Mateen M, Cai B, Hu H. Automated Lung Nodule Detection and Classification Using Deep Learning Combined With Multiple Strategies. Sensors (Basel) (2019) 19(17):3722. doi: 10.3390/s19173722
29. Ban K, Tsunoda H, Suzuki S, Takaki R, Sasaki K, Nakagawa M. Verifification of Recall Criteria for Masses Detected on Ultrasound Breast Cancer Screening. J Med Ultrasonics (2018) 45:65–73. doi: 10.1007/s10396-017-0778-5
30. Mainiero MB, Goldkamp A, Lazarus E, Livingston L, Koelliker SL, Schepps B, et al. Characterization of Breast Masses With Sonography: Can Biopsy of Some Solid Masses Be Deferred? J Ultrasound Med (2005) 24(2):161–7. doi: 10.7863/jum.2005.24.2.161
31. Graf O, Helbich TH, Hopf G, Graf C, Sickles EA. Probably Benign Breast Masses at US: Is Follow-Up an Acceptable Alternative to Biopsy? Radiology (2007) 244(1):87–93. doi: 10.1148/radiol.2441060258
32. Spak DA, Plaxco JS, Santiago L, Dryden MJ, Dogan BE. BI-RADS(®) Fifth Edition: A Summary of Changes. Diagn Interv Imaging (2017) 98(3):179–90. doi: 10.1016/j.diii.2017.01.001
33. Setio AA, Ciompi F, Litjens G, Gerke P, Jacobs C, van Riel SJ, et al. Pulmonary Nodule Detection in CT Images: False Positive Reduction Using Multi-View Convolutional Networks. IEEE Trans Med Imaging (2016) 35(5):1160–9. doi: 10.1109/TMI.2016.2536809
34. Anirudh R, Thiagarajan J, Bremer T, Kim H. Lung Nodule Detection Using 3D Convolutional Neural Networks Trained on Weakly Labeled Data. Spie (2016) 9785:978532–6. doi: 10.1117/12.2214876
35. Magnuska ZA, Theek B, Darguzyte M, Palmowski M, Stickeler E, Schulz V, et al. Influence of the Computer-Aided Decision Support System Design on Ultrasound-Based Breast Cancer Classification. Cancers (2022) 14(2):277. doi: 10.3390/cancers14020277
36. Hsieh YH, Hsu FR, Dai ST, Huang HY, Chen DR, Shia WC. Incorporating the Breast Imaging Reporting and Data System Lexicon With a Fully Convolutional Network for Malignancy Detection on Breast Ultrasound. Diagnostics (Basel, Switzerland (2021) 12(1):66. doi: 10.3390/diagnostics12010066
Keywords: automatic breast ultrasound (ABUS), convolution neural network, breast cancer, detection, validation data
Citation: Zhang JX, Tao X, Jiang YH, Wu XX, Yan D, Xue W, Zhuang SL, Chen L, Luo LP and Ni D (2022) Application of Convolution Neural Network Algorithm Based on Multicenter ABUS Images in Breast Lesion Detection. Front. Oncol. 12:938413. doi: 10.3389/fonc.2022.938413
Received: 07 May 2022; Accepted: 30 May 2022;
Published: 04 July 2022.
Edited by:
San-Gang Wu, First Affiliated Hospital of Xiamen University, ChinaReviewed by:
Fei Yan, Institutes of Advanced Technology, (CAS), ChinaLihua Li, Hangzhou Dianzi University, China
Zhiyi Chen, Third Affiliated Hospital of Guangzhou Medical University, China
Copyright © 2022 Zhang, Tao, Jiang, Wu, Yan, Xue, Zhuang, Chen, Luo and Ni. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Jianxing Zhang, zhangjx@gzucm.edu.cn; Liangping Luo, tluolp@jnu.edu.cn; Dong Ni, nidong@szu.edu.cn