- 1Department of Radiology, The Second Affiliated Hospital of Guangzhou University of Chinese Medicine, Guangzhou, China
- 2Department of Engineering, Shanghai Yanghe Huajian Artificial Intelligence Technology Co., Ltd, Shanghai, China
- 3Artificial Intelligence (AI), Research Lab, Boston Meditech Group, Burlington, MA, United States
Background: Architectural distortion (AD) is a common imaging manifestation of breast cancer, but is also seen in benign lesions. This study aimed to construct deep learning models using mask regional convolutional neural network (Mask-RCNN) for AD identification in full-field digital mammography (FFDM) and evaluate the performance of models for malignant AD diagnosis.
Methods: This retrospective diagnostic study was conducted at the Second Affiliated Hospital of Guangzhou University of Chinese Medicine between January 2011 and December 2020. Patients with AD in the breast in FFDM were included. Machine learning models for AD identification were developed using the Mask RCNN method. Receiver operating characteristics (ROC) curves, their areas under the curve (AUCs), and recall/sensitivity were used to evaluate the models. Models with the highest AUCs were selected for malignant AD diagnosis.
Results: A total of 349 AD patients (190 with malignant AD) were enrolled. EfficientNetV2, EfficientNetV1, ResNext, and ResNet were developed for AD identification, with AUCs of 0.89, 0.87, 0.81 and 0.79. The AUC of EfficientNetV2 was significantly higher than EfficientNetV1 (0.89 vs. 0.78, P=0.001) for malignant AD diagnosis, and the recall/sensitivity of the EfficientNetV2 model was 0.93.
Conclusion: The Mask-RCNN-based EfficientNetV2 model has a good diagnostic value for malignant AD.
1 Introduction
Breast cancer is the most common malignancy in women, ranking the highest incidence among women in the world (1, 2). According to the latest global cancer burden data, the number of new breast cancer cases worldwide reached 2.26 million in 2020, surpassing lung cancer and becoming the most common cancer in the world (1, 2).
Breast cancer screening with mammography is considered effective at reducing breast cancer-related mortality (3, 4). The common imaging manifestations of breast cancer are masses and calcification, followed by architectural distortion (AD). It is sometimes the only manifestation of breast cancer and has important imaging value. On the other hand, AD is seen in malignant and benign lesions (such as sclerosing adenosis, radial scar, postoperative scar, and fat necrosis after trauma, among others). AD is a structural deformation without a defined mass in breast tissue as a fine line or protrusion radiating from a point and a focal contraction, twisted or stiff at the edge of the parenchymal gland (5). AD is more subtle relative to masses and calcification with clear boundaries. AD lesions are often poorly defined, overlap and cover with normal glands, and the lesions are not easy to identify.
There are no established noninvasive standards for the distinction between benign and malignant AD. The literature reported different diagnostic efficacy of various imaging methods for cancer manifesting as AD, such as magnetic resonance imaging (MR) (6), digital breast tomosynthesis (DBT) (7), contrast enhanced spectral mammography (CESM) (8), and the positive predictive value (PPV) varies 34%-88% (8, 9). Amitai et al. reported that the specificity of MRI for AD diagnosis was 68%, and the overall accuracy was 73% (6). Goh et al. reported that the accuracy of CEDM for malignant AD was 72.5% (10). Patel et al. reported that the accuracy of CESM for malignant AD was 82% (8). However the judgment of benign and malignant AD lesions is still difficult for radiologists.
Recently, artificial intelligence (AI) algorithms have been extensively applied in the medical field. Available artificial intelligence (AI) including radiomics and deep learning have been applied to analyze images for detection and diagnosis of lesions in various clinical applications (11–13). Available AI is advanced and approach radiologists’ performance, especially for mammography (7).
The Mask Regional CNN (Mask-RCNN) (14, 15) is a deep neural network aimed at solving instance segmentation problems in machine learning. It effectively combines the two tasks of target detection and image segmentation. However, limited Mask RCNN-based deep learning have been developed for AD identification and malignant AD diagnosis. Rehman et al. recently developed an automated computer-aided diagnostic system using computer vision and deep learning to predict breast cancer based on the architectural distortion on DM and reported great accuracy (16). Xiao et al. proposed two AI methods, radiomics and deep learning, to build diagnostic models for patients presenting with architectural distortion on Digital Breast Tomosynthesis (DBT) images (17).
Although DBT can provide better spatial information for detection and characterization of architectural distortion, many hospital breast X-ray examination are still 2D mammography because of economic reasons not updated equipment timely. Fortunately, AI can be applied to develop fully-automatic computer-aided diagnostic systems (18, 19) and help radiologists improve the diagnostic efficiency (20). Yun et al. underscore the potential of using deep learning methods to enhance the overall accuracy of pretest mammography for malignant AD (20).
Therefore, this study aimed to construct and optimize deep learning models by Mask-RCNN for AD identification in FFDM and evaluate the performance of models for the diagnosis of malignant AD.
2 Materials and methods
2.1 Study design and participants
This retrospective diagnostic study was conducted at the Second Affiliated Hospital of Guangzhou University of Chinese Medicine between January 2011 and December 2020. Patients with breast AD in full-field digital mammography (FFDM) were included. The inclusion criteria were 1) AD according to the fifth edition of the breast imaging reporting and data system (BI-RADS) diagnostic criteria for architectural distortion (5) and 2) available surgical or biopsy pathological results. The exclusion criteria were 1) patients with obvious mass in the breast or 2) incomplete clinical data or images. Finally, a total of 349 patients were included in this study. Each patient on one side of the breast image contains at least one AD lesions.
2.2 Data collection and image annotation
The demographic information and clinical characteristics of the patients were collected, including age, menopause status, pathology, childbirth history, menopausal age, and surgical history. The mammogram images were collected from the Giotto FFDM system (internazionale medico scientifica, IMS, Bologna, Italy). The images for each patient were taken in four standard views: right craniocaudal (R-CC), left craniocaudal (L-CC), right mediolateral oblique (R-MLO), and left mediolateral oblique (L-MLO). The images from the eligible patients were exported to the computer in medical digital imaging and communication (DICOM) format for data anonymization. All images were re-screened to remove substandard images (i.e., incomplete image sequences, poor image quality, artifacts, and cases with clear masses). A total of six radiologists participated in the image processing, and they were divided into three groups, each group including a junior and a senior radiologist. The AD structures on images were outlined and annotated by a group of two experienced radiologists with the ITK-SNAP software (version 3.8). The outlined scope had to include all lesions (such as AD with calcification or asymmetry). The breast fibroglandular tissue (FGT) and imaging features of the lesions including size of lesion, calcification and BI-RADS classification were recorded.
The segmented images were saved in the “.nii” format, and the Python code was used to calculate the regions of interest (ROI) and intersection over union (IOU) to assess the consistency of focal delineation. IOU was calculated according to Equation 1, with A and B representing the ROI area delineated by different radiologists. Images with IOU >0.5 were included in the study (Figure 1A), and images with IOU <0.5 were redelineated (Figure 1B). All six radiologists participated in the redelineation process, reviewed the film again, and delineated the image again after reaching an agreement (Figure 1C). Typical malignant and benign AD images are shown in Figures 1D, E.
Figure 1 Typical images and image annotation of architectural distortion. (A) With intersection over union (IOU) > 0.5, the image was successfully delineated and included in the model training. The red outlined area was marked by the senior radiologist, and the green outlined area was marked by the junior radiologist. (B) With IOU <0.5, the image was re-delineated. The red outlined area was marked by the senior radiologist, and the green outlined area was marked by the junior radiologist. (C) The green outlined area represents the lesion area delineated by the radiologists, the yellow area represents the machine identification area, and the red box is the lesion range identified by machine learning models. (D) FFDM images of a 45-year-old female diagnosed with malignant AD (write arrow). The BI-RADS score is 4C. The pathology was invasive ductal carcinoma. Limited stiffness was seen on the right breast andspiculated margins. (E) FFDM images of 64-years-old female with benign AD (write arrow). The BI-RADS score is 4C.The pathology was Complex sclerosinglesion. Stellate shadow and scattered cord were seen on the left breast.
2.3 Deep learning model construction
The Mask RCNN was used to construct deep learning models in this study. A combination of training-aware neural architecture search and scaling was used to optimize training speed jointly and to develop these models. The image size of the mammograms used in this study was 2816×3584 pixels. In order to better detect and classify the benign and malignant AD lesions, the images were first scaled to 1024×1024, and Mask RCNN was used to detect and segment the breast AD lesions. There are two stages of Mask RCNN. Both stages are connected to the backbone structure, which is an FPN neural network that consists of a bottom-up pathway, a top-bottom pathway, and lateral connections. The bottom-up pathway can be any ConvNet and Transformer, such as ResNet, ResNext (21, 22), EfficientNetV1 and EfficientNetV2, which extracts the features from raw images. As a network for extracting features, the performance of ResNet and EfficientNet is widely recognized. The ResNet network solves the problem of gradient explosion and training overfitting caused by too deep a network. The ResNext network adds more branches on the basis of ResNet, thereby improving the network’s ability to learn features. EfficientNet V1 and V2 networks use the function of network search on the basis of ResNet, and further enhance the performance and efficiency of the network through parameter combinations. The top-bottom pathway generates a feature pyramid map similar in size to the bottom-up pathway. Lateral connections are convolutions and adding operations between two corresponding levels of the two pathways.
After extracting the ROI of the AD lesions, the ROI was restored to its original image size. A square area containing the lesion area was extracted from the center of the lesion (Figure 1C), and the network was used for benign and malignant classification.
2.4 Data preprocessing and model training
First, the DICOM data was converted into 16-bit PNG. Then, the image was processed by data augmentation and normalization. Data augmentation includes geometric transformations, color space augmentations, kernel filters, mixing images, and random erasing. Data augmentation can expand samples, prevent overfitting, and improve model robustness. During training, the random erasing method randomly selects a rectangular area in the original image. Then replace the pixels in that area with random values. For the detected and segmented network, the augmented and normalized image and label were scaled to 1024×1024. The ROI of AD lesions is realized by the segmentation algorithm. For the classification of the benign and malignant lesions, lesions often only occupy a small part of the image, so AD lesions are extracted according to the detected ROI, and then the extracted image is scaled to 512×512. Then, 349 patients with 1396 valid images were obtained (349×2×2 = 1396). There were 698 images on one side of the breast with 708 AD lesions in total, with more than one AD lesion detected in some images, and 698 images of contralateral breast without AD lesion. Among all patients, 60% were randomly selected as the training set,20% as the test set, and the other 20% as the validation set. A total of 209 patients with 836 validated images were included in the training set, while 70 patients with 280 images were included in the test set and the same images for the validation set. The training set was used for model training and parameter learning of the model and automatically saved the best model at any time and processed all training data during training. The test set was used for evaluating the diagnostic performance of models. The validation set was applied to save the best model parameters during training and guide the choice of parameters and models.
When training the Mask RCNN, four feature extraction networks were selected for comparison: EfficientNetV2, EfficientNetV1, ResNet, and ResNext. During training, Mask RCNN used the multi-task loss, including CrossEntropy Loss for category loss, L1 loss for regression box loss function, and CrossEntropy Loss for the mask loss function. After 50 epochs of training, four network model weights with the best performance in the validation set were obtained. The models with the highest AUC for AD identification were selected as the backbone of Mask RCNN for malignant AD identification. The pathological diagnosis was used as the gold standard for malignant AD diagnosis. The process of Mask RCNN model training and model validation for benign and malignant AD classification is shown in Figure 2.
Figure 2 Flow chart of Mask RCNN model training and validation for benign and malignant AD classification. (A) The process of Mask RCNN model construction and training. (B) The validation of Mask RCNN models for AD classification.
2.5 Statistical analysis
Statistical analysis was performed with SPSS 22.0 (IBM, Armonk, NY, USA). The normality test was performed using the Kolmogorov-Smirnov test. Continuous variables with a normal distribution were described as means ± standard deviation (SD) and compared using Student’s t-test. Continuous variables with a skewed distribution were described as median (interquartile range) and compared using Mann-Whitney U-test. Categorical data were presented as n (%) and compared with the chi-square test or Fisher’s exact test. The diagnostic performance of the models was evaluated in the validation set. The receiver operating characteristic (ROC) curve and area under the ROC curve (AUC) were determined, and the Delong test was used for the comparison of AUC. Accuracy, specificity, precision, recall/sensitivity, F1-score, Dice, and Jacc were selected as performance metrics of the deep learning models (Supplementary Material). In addition, 95% confidence intervals (CI) were calculated. Two-sided P-values <0.05 were considered statistically significant.
3 Results
A total of 349 patients were included: 159 with benign AD and 190 with malignant AD. Patients with malignant AD were aged 49.0 (43.0-56.0) years, and 102 (53.7%) patients were menopausal. Patients with benign AD were aged 48.0 (43.0-52.0) years, and 62 (36.7%) patients were menopausal. There were no differences between the two groups for age (P=0.26), but significantly more patients with malignant AD were menopausal (P=0.01) (Table 1). The maximum diameter and vertical diameter of malignant AD are greater than benign AD. The median of maximum diameter and vertical diameter of malignant AD was 1.7cm/1.3cm, the benign AD was 1.1cm/0.9cm. Benign and malignant AD patients had different distributions of FGT (P=0.02) and BI-RADS classification (P<0.001). There were no differences in calcification between the two groups (P=0.23) (Table 1). In the malignant group, 116 (61.1%) patients had invasive ductal carcinoma (IDC), 12 (6.3%) had invasive lobular carcinoma (ILC), eight (4.2%) had ductal carcinoma in situ (DCIS), one (0.5%) had mucinous carcinoma, two had medullary carcinoma (1.1%), 12 had mixed carcinomas (6.3%), and 39 cases were recorded as “breast malignancy” without detailed pathology (the malignant pathology of the 39 cases were further confirmed by a telephone follow-up) (Table 2). In the benign group, 92 (57.9%) patients had breast fibrocystic hyperplasia, 17 (10.7%) had postoperative scars, 34 (21.4%) had a radial scar or sclerosing lesion, nine (5.7%) had benign tumors, one (0.6%) had chronic mastitis, and six (3.8%) had breast fibromatosis (Table 2).
The accuracy, precision, recall/sensitivity, F1-score, Dice, Jacc of EfficientNetV2 for AD identification were 0.80, 0.91, 0.66, 0.77, 0.64, and 0.51, while those of EfficientNetV1 were 0.79, 0.84, 0.69, 0.76, 0.63, and 0.50, respectively (Table 3). The accuracy of ResNext, and ResNet for AD identification were 0.73 and 0.72, respectively (Table 3). The AUCs of EfficientNetV2, EfficientNetV1, ResNext, and ResNet for AD identification were 0.89 (95% CI: 0.83-0.96), 0.87 (95% CI: 0.78-0.93), 0.81 (95% CI: 0.71-0.88), and 0.79 (95% CI: 0.70-0.88), respectively (Figure 3A). The EfficientNetV2 model had significantly higher AUC for AD identification than ResNet and ResNext (P=0.005 and P=0.028, respectively), and there was no significant difference in AUC between EfficientNetV2 and EfficientNetV1 in AD identification (P=0.125). Therefore, EfficientNetV2 and EfficientNetV1 were selected as the models for malignant AD diagnosis.
Figure 3 Comparison of receiver operating characteristics (ROC) curves of deep learning models. (A) ROC curves of the EfficientNetV1, EfficientNetV2, ResNet, and ResNext models for architectural distortion identification. (B) ROC curves of the EfficientNetV1 and EfficientNetV2 models for malignant architectural distortion diagnosis.
The AUC of EfficientNetV2 (AUC=0.89, 95% CI: 0.81-0.95) was significantly higher than that of EfficientNetV1 (AUC=0.78, 95% CI: 0.68-0.86) for malignant AD diagnosis (P=0.001) (Figure 3B). The accuracy, precision, recall/sensitivity, F1 score, and specificity of EfficientNetV2 for malignant AD diagnosis were 0.84, 0.83, 0.93, 0.84, and 0.74, while those of EfficientNetV1 were 0.82, 0.87, 0.77, 0.82, and 0.87, respectively (Table 4).
4 Discussion
This study constructed four deep learning models based on the Mask RCNN method for AD identification, and the EfficientNetV2 model has a great diagnostic value for malignant AD, with an AUC of 0.89 and recall/sensitivity of 0.93. The EfficientNetV2 model might help radiologists in malignant AD diagnosis, decreasing the need for invasive diagnostic procedures.
Identifying subtle lesions in mammography screening is challenging, with 12.5% of malignancies missed in clinical practice (23, 24). In this study, 110 out of 190 malignant ADs were diagnosed as BI-RADS 5 grade, mainly because the images included typical malignant morphological features or malignant calcification with obvious malignant features. On the other hand, 132 of 159 benign AD were diagnosed with BI-RADS with grade 4, with 115 diagnosed with 4c and nine with grade 5, with a high rate of misdiagnosis as malignant lesions (Figures 1D, E). From these two images, both lesions had typical spiculated margins which the radiologists considered malignancy. Because spiculated margins of radiologically detected masses have been well-known morphologic criteria for breast malignancy (25, 26). However, the pathological findings were IDC and sclerosing lesion. It suggested that radiologists could find AD signs in FFDM, but the accuracy of benign vs. malignant differentiation was low, with many cases misdiagnosed as malignant lesions. Therefore, it is of great difficulty for radiologists to identify the AD lesions and differentiate between benign and malignant AD simply based on the morphological characteristics of FFDM observed by the naked eyes.
In routine clinical work, it is necessary to confirm imaging lesions using other imaging methods, such as DBT, Contrast-enhanced mammography (CEM), ultrasound, and MRI, and often an invasive diagnostic method must be performed in case of doubt. It is reported that the negative predictive value of contrast-enhanced MRI (CEMR) is 100% (6), which can help exclude malignant lesions, but CEMR had a low positive predictive value of 30%. It is also reported that the sensitivity, specificity, PPV, and NPV of Contrast-enhanced digital mammography (CEDM) are 100%, 42.6%, 48.5%, and 100% (10) for the diagnosis of malignant AD, respectively. Although the sensitivity of CEMR and CEDM examination is high, they had low positive predictive values, which means that not all the enhanced lesions are malignant, and some are benign lesions. Therefore, even the use of complementary imaging can remain inconclusive. MRI and CEM in routine breast diagnostic tests or screening are not currently standard for AD testing, given the lack of cost-effectiveness and exact diagnostic efficacy (9). A biopsy can provide pathological results, but as an invasive procedure with risks, it brings anxiety to patients and has a chance of a missed diagnosis of malignant foci. Therefore, it is clinically meaningful to find other methods to improve the diagnosis of AD without biopsy.
The fifth edition of the BI-RADS recommends that malignant lesions should be suspected without a definite trauma or surgical history, and further biopsy is recommended (5). Surgical excision has long been advocated for managing DM-detected AD, but more recent evidence suggests that surgical excision of DM-detected AD is not necessary in certain nonmalignant cases, such as when needle biopsy yields radial scar without associated atypia (27–29). Some authors have also proposed that managing nonmalignant architectural distortion on DBT remains controversial; imaging surveillance can be considered for AD on DBT yielding radial scar without atypia or other concordant benign pathologies without atypia at biopsy (30). A biopsy can sometimes provide error pathological results because the needle can sample tissues besides the malignant foci. Therefore, it is clinically meaningful to find other methods to improve the diagnosis of malignant AD without biopsy.
This study aimed to construct an AI model with good diagnostic value for malignant AD. In this study, we built a lesion extraction algorithm corresponding to malignant AD and computed its performance, and at the same time, we used the advanced EfficientNet convolutional neural network with higher computing efficiency and better generalization ability. EfficientNet is based on a lightweight convolutional neural network, which has better model compression ability. It can be widely used in mobile image recognition, object detection, image segmentation, and other tasks, and can meet the more stringent limitations of computing resources (31–34). The EfficientNetV2 model had an accuracy of 0.84, precision of 0.83, recall/sensitivity of 0.93, and F1 score of 0.84 for malignant AD diagnosis, indicating a huge application potential of deep learning models in the diagnosis of malignant AD.
Hand-craft feature extraction techniques showed some value for determining malignant ADs, but these methods still rely on an operator detecting the image features (35). We hope to use deep learning method to establish computer automatic recognition of image features and diagnostic process, to help radiologists improve work efficiency and diagnostic accuracy. Murali et al. used a support vector machine and achieved an accuracy of 90% using 150 AD ROIs (36). Banik et al. examined 4224 ROIs using the gobar filter and phase portrait analysis method and achieved a sensitivity of 90% (37). Jasionowska et al. used a complicated two-step approach (Gobar filter followed by 2-D Fourier transform) and achieved 84% accuracy (38). Other models also achieved relatively good accuracies (39, 40). Still, an issue with deep learning is that the AI model and source of the data can influence the outcomes and that a model that achieves high accuracy with one database might have lower performance with another database. Rehman et al. achieved accuracies of 0.95, 0.97, and 0.98 for the diagnosis of malignant using a larger amount of mammographies from three different databases (16). The reason for their better results may be a database based on larger amounts of data. Fortunately, the increasing availability of mammography databases will help with the development of AI. Even though the present study did not compare multiple imaging methods, the diagnostic accuracy of the EfficientNetV2 model appears promising. It is hoped that AI can be continuously developed and improved in the future, and they may assist radiologists in improving diagnostic accuracy.
This study has some limitations. First, the work is limited by its retrospective design, which leads to some degree of selection bias. Second, only one center was involved, limiting the number of cases. External validation studies are needed. Even with deficiencies, it is still believed that maximizing the clinical application of AI remains an ultimate goal for improving breast cancer screening.
In conclusion, the EfficientNetV2 established based on the Mask RCNN deep learning method has a good diagnostic value for malignant AD. It might help radiologists with malignant AD diagnoses. The results underscore the potential of using deep learning methods to enhance the overall accuracy of mammography for malignant AD.
Data availability statement
The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding authors.
Ethics statement
The studies involving human participants were reviewed and approved by the Second Affiliated Hospital of Guangzhou University of Traditional Chinese Medicine (Ethical No. ZE2020-200-01). Written informed consent for participation was not required for this study in accordance with the national legislation and the institutional requirements.
Author contributions
Conceptualization, BL and DC; Methodology, YL and YT; Software, YT; Validation, YL and YT; Formal Analysis, YT; Investigation, ZX; Resources, GY; Data Curation, GY; Writing – Original Draft Preparation, YL; Writing – Review and Editing, YL; Visualization, YT; Supervision, BL; Project Administration, YW; Funding Acquisition, DC. All authors contributed to the article and approved the submitted version.
Funding
This work was supported by the following projects: Guangdong Province Medical Research Foundation project of Health Commission of Guangdong Province (C2022092), Traditional Chinese Medicine Science and Technology Project of Guangdong Hospital of Traditional Chinese Medicine (YN2020MS09), Science and Technology Program of Guangzhou (202102010260).
Conflict of interest
Author YT was employed by the Shanghai Yanghe Huajian Artificial Intelligence Technology Co., Ltd, Shanghai, China. Author DC was employed by AI Research Lab, Boston Meditech Group, Burlington, USA.
The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
Supplementary material
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2023.1119743/full#supplementary-material
Glossary
References
1. Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A, et al. Global cancer statistics 2020: Globocan estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA: Cancer J Clin (2021) 71(3):209–49. doi: 10.3322/caac.21660
2. Siegel RL, Miller KD, Fuchs HE, Jemal A. Cancer statistics, 2022. CA: Cancer J Clin (2022) 72(1):7–33. doi: 10.3322/caac.21708
3. Broeders M, Moss S, Nystrom L, Njor S, Jonsson H, Paap E, et al. The impact of mammographic screening on breast cancer mortality in Europe: A review of observational studies. J Med Screen (2012) 19 Suppl 1:14–25. doi: 10.1258/jms.2012.012078
4. Duffy SW, Vulkan D, Cuckle H, Parmar D, Sheikh S, Smith RA, et al. Effect of mammographic screening from age 40 years on breast cancer mortality (Uk age trial): Final results of a randomised, controlled trial. Lancet Oncol (2020) 21(9):1165–72. doi: 10.1016/S1470-2045(20)30398-3
5. D’Orsi CJ, Sickles EA, Mendelson EB, Morris EA. Acr bi-rads atlas, breast imaging reporting and data system. 5th Edition. Reston: American College of Radiology (2013).
6. Amitai Y, Scaranelo A, Menes TS, Fleming R, Kulkarni S, Ghai S, et al. Can breast mri accurately exclude malignancy in mammographic architectural distortion? Eur Radiol (2020) 30(5):2751–60. doi: 10.1007/s00330-019-06586-x
7. Geras KJ, Mann RM, Moy L. Artificial intelligence for mammography and digital breast tomosynthesis: Current concepts and future perspectives. Radiology (2019) 293(2):246–59. doi: 10.1148/radiol.2019182627
8. Patel BK, Naylor ME, Kosiorek HE, Lopez-Alvarez YM, Miller AM, Pizzitola VJ, et al. Clinical utility of contrast-enhanced spectral mammography as an adjunct for tomosynthesis-detected architectural distortion. Clin Imaging (2017) 46:44–52. doi: 10.1016/j.clinimag.2017.07.003
9. Choudhery S, Johnson MP, Larson NB, Anderson T. Malignant outcomes of architectural distortion on tomosynthesis: A systematic review and meta-analysis. AJR Am J Roentgenol (2021) 217(2):295–303. doi: 10.2214/AJR.20.23935
10. Goh Y, Chan CW, Pillay P, Lee HS, Pan HB, Hung BH, et al. Architecture distortion score (Ads) in malignancy risk stratification of architecture distortion on contrast-enhanced digital mammography. Eur Radiol (2021) 31(5):2657–66. doi: 10.1007/s00330-020-07395-3
11. Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, et al. Dermatologist-level classification of skin cancer with deep neural networks. Nature (2017) 542(7639):115–8. doi: 10.1038/nature21056
12. Franck C, Snoeckx A, Spinhoven M, El Addouli H, Nicolay S, Van Hoyweghen A, et al. Pulmonary nodule detection in chest ct using a deep learning-based reconstruction algorithm. Radiat Prot dosimetry (2021) 195(3-4):158–63. doi: 10.1093/rpd/ncab025
13. Truhn D, Schrading S, Haarburger C, Schneider H, Merhof D, Kuhl C. Radiomic versus convolutional neural networks analysis for classification of contrast-enhancing lesions at multiparametric breast mri. Radiology (2019) 290(2):290–7. doi: 10.1148/radiol.2018181352
14. He K, Gkioxari G, Dollar P, Girshick R. Mask r-cnn. IEEE Trans Pattern Anal Mach Intell (2020) 42(2):386–97. doi: 10.1109/TPAMI.2018.2844175
15. Zhang Y, Chu J, Leng L, Miao J. Mask-refined r-cnn: A network for refining object details in instance segmentation. Sens (Basel) (2020) 20(4):1010–1015. doi: 10.3390/s20041010
16. Rehman KU, Li J, Pei Y, Yasin A, Ali S, Saeed Y. Architectural distortion-based digital mammograms classification using depth wise convolutional neural network. Biology(Basel) (2021) 11(1):15–43. doi: 10.3390/biology11010015
17. Chen X, Zhang Y, Zhou J, Wang X, Liu X, Nie K, et al. Diagnosis of architectural distortion on digital breast tomosynthesis using radiomics and deep learning. Front Oncol (2022) 12:991892. doi: 10.3389/fonc.2022.991892
18. Mettivier G, Ricciarci R, Sarno A, Maddaloni F, Porzio M, Staffa M. Deeplook: A deep learning computed diagnosis support for breast tomosynthesis. In: In: 16th international workshop on breast imaging (IWBI2022). Leuven, Belgium: SPIE (2022).
19. Ricciardi R, Mettivier G, Staffa M, Sarno A, Acampora G, Minelli S, et al. A deep learning classifier for digital breast tomosynthesis. Physica medica: PM: an Int J devoted to Appl Phys to Med biol: Off J Ital Assoc Biomed Phys (AIFB) (2021) 83:184–93. doi: 10.1016/j.ejmp.2021.03.021
20. Wan Y, Tong Y, Liu Y, Huang Y, Yao G, Chen DQ, et al. Evaluation of the combination of artificial intelligence and radiologist assessments to interpret malignant architectural distortion on mammography. Front Oncol (2022) 12:880150. doi: 10.3389/fonc.2022.880150
21. Abedalla A, Abdullah M, Al-Ayyoub M, Benkhelifa E. Chest X-ray pneumothorax segmentation using U-net with efficientnet and resnet architectures. PeerJ Comput Sci (2021) 7:e607. doi: 10.7717/peerj-cs.607
22. He F, Liu T, Tao D. Why resnet works? residuals generalize. IEEE Trans Neural Netw Learn Syst (2020) 31(12):5349–62. doi: 10.1109/TNNLS.2020.2966319
23. Lehman CD, Arao RF, Sprague BL, Lee JM, Buist DS, Kerlikowske K, et al. National performance benchmarks for modern screening digital mammography: Update from the breast cancer surveillance consortium. Radiology (2017) 283(1):49–58. doi: 10.1148/radiol.2016161174
24. Yankaskas BC, Schell MJ, Bird RE, Desrochers DA. Reassessment of breast cancers missed during routine screening mammography: A community-based study. AJR Am J Roentgenol (2001) 177(3):535–41. doi: 10.2214/ajr.177.3.1770535
25. Macura KJ, Ouwerkerk R, Jacobs MA, Bluemke DA. Patterns of enhancement on breast Mr images: Interpretation and imaging pitfalls. Radiograph: Rev Publ Radiol Soc North Am Inc (2006) 26(6):1719–34. doi: 10.1148/rg.266065025
26. Nunes LW, Schnall MD, Orel SG. Update of breast Mr imaging architectural interpretation model. Radiology (2001) 219(2):484–94. doi: 10.1148/radiology.219.2.r01ma44484
27. Conlon N, D’Arcy C, Kaplan JB, Bowser ZL, Cordero A, Brogi E, et al. Radial scar at image-guided needle biopsy: Is excision necessary? Am J Surg Pathol (2015) 39(6):779–85. doi: 10.1097/PAS.0000000000000393
28. Leong RY, Kohli MK, Zeizafoun N, Liang A, Tartter PI. Radial scar at percutaneous breast biopsy that does not require surgery. J Am Coll Surg (2016) 223(5):712–6. doi: 10.1016/j.jamcollsurg.2016.08.003
29. Martaindale S, Omofoye TS, Teichgraeber DC, Hess KR, Whitman GJ. Imaging follow-up versus surgical excision for radial scars identified on tomosynthesis-guided core needle biopsy. Acad Radiol (2020) 27(3):389–94. doi: 10.1016/j.acra.2019.05.012
30. Villa-Camacho JC, Bahl M. Management of architectural distortion on digital breast tomosynthesis with nonmalignant pathology at biopsy. AJR Am J Roentgenol (2022) 219(1):46–54. doi: 10.2214/AJR.21.27161
31. Chetoui M, Akhloufi MA. Explainable diabetic retinopathy using efficientnet(.). Annu Int Conf IEEE Eng Med Biol Soc (2020) 2020:1966–9. doi: 10.1109/EMBC44109.2020.9175664
32. Wang J, Liu Q, Xie H, Yang Z, Zhou H. Boosted efficientnet: Detection of lymph node metastases in breast cancer using convolutional neural networks. Cancers (Basel) (2021) 13(4):661–74. doi: 10.3390/cancers13040661
33. Tan M, Le QV. (2021). Effificientnetv2: Smaller models and faster training. International conference on machine learning. PMLR (2021) 2021:10096–106. doi: 10.48550/arXiv.2104.00298
34. Aldi J, Parded HF. Comparison of classification of birds using lightweight deep convolutional neural networks. J Elektronika dan Telekomunikasi (2022) 22(2):87–94. doi: 10.55981/jet.503
35. Heidari M, Mirniaharikandehei S, Liu W, Hollingsworth AB, Liu H, Zheng B. Development and assessment of a new global mammographic image feature analysis scheme to predict likelihood of malignant cases. IEEE Trans Med Imaging (2020) 39(4):1235–44. doi: 10.1109/TMI.2019.2946490
36. Murali S, Dinesh M. Model based approach for detection of architectural distortions and spiculated masses in mammograms. Int J Comput Sci Eng (2011) 3(11):3534–46.
37. Banik S, Rangayyan RM, Desautels JE. Detection of architectural distortion in prior mammograms. IEEE Trans Med Imaging (2011) 30(2):279–94. doi: 10.1109/TMI.2010.2076828
38. Jasionowska M, Przelaskowski A, Rutczynska A, Wroblewska A. A two-step method for detection of architectural distortions in mammograms. In: Information technologies in biomedicine. Berlin: Springer (2010).
39. Boca Bene I, Ciurea AI, Ciortea CA, Stefan PA, Lisencu LA, Dudea SM. Differentiating breast tumors from background parenchymal enhancement at contrast-enhanced mammography: The role of radiomics-a pilot reader study. Diagnostics (Basel) (2021) 11(7):1248–1259. doi: 10.3390/diagnostics1071248
Keywords: deep learning, convolutional neural network, artificial intelligence, malignant architectural distortion, full-field digital mammography
Citation: Liu Y, Tong Y, Wan Y, Xia Z, Yao G, Shang X, Huang Y, Chen L, Chen DQ and Liu B (2023) Identification and diagnosis of mammographic malignant architectural distortion using a deep learning based mask regional convolutional neural network. Front. Oncol. 13:1119743. doi: 10.3389/fonc.2023.1119743
Received: 19 December 2022; Accepted: 27 February 2023;
Published: 22 March 2023.
Edited by:
L. J. Muhammad, Federal University Kashere, NigeriaReviewed by:
Yahaya Zakariyau, Federal University Kashere, NigeriaLisha Qi, Tianjin Medical University Cancer Institute and Hospital, China
Copyright © 2023 Liu, Tong, Wan, Xia, Yao, Shang, Huang, Chen, Chen and Liu. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Bo Liu, liubogzcm@163.com; Daniel Q. Chen, dchen@bostonmeditech.com