- 1School of Computer Engineering and Science, Shanghai University, Shanghai, China
- 2Department of Radiation Oncology, Fudan University Shanghai Cancer Center, Shanghai, China
- 3School of Medicine, Shanghai University, Shanghai, China
- 4Department of Ultrasound, Ren Ji Hospital, Shanghai Jiao Tong University School of Medicine, Shanghai, China
The early prediction of a patient’s response to neoadjuvant chemotherapy (NAC) in breast cancer treatment is crucial for guiding therapy decisions. We aimed to develop a novel approach, named the dual-branch convolutional neural network (DBNN), based on deep learning that uses ultrasound (US) images for the early prediction of NAC response in patients with locally advanced breast cancer (LABC). This retrospective study included 114 women who were monitored with US during pretreatment (NAC pre) and after one cycle of NAC (NAC1). Pathologic complete response (pCR) was defined as no residual invasive carcinoma in the breast. For predicting pCR, the data were randomly split into a training set and test set (4:1). DBNN with US images was proposed to predict pCR early in breast cancer patients who received NAC. The connection between pretreatment data and data obtained after the first cycle of NAC was considered through the feature sharing of different branches. Moreover, the importance of data in various stages was emphasized by changing the weight of the two paths to classify those with pCR. The optimal model architecture of DBNN was determined by two ablation experiments. The diagnostic performance of DBNN for predicting pCR was compared with that of four methods from the latest research. To further validate the potential of DBNN in the early prediction of NAC response, the data from NAC pre and NAC1 were separately assessed. In the prediction of pCR, the highest diagnostic performance was obtained when combining the US image information of NAC pre and NAC1 (area under the receiver operating characteristic curve (AUC): 0.939; 95% confidence interval (CI): 0.907, 0.972; F1-score: 0.850; overall accuracy: 87.5%; sensitivity: 90.67%; and specificity: 85.67%), and the diagnostic performance with the combined data was superior to the performance when only NAC pre (AUC: 0.730; 95% CI: 0.657, 0.802; F1-score: 0.675; sensitivity: 76.00%; and specificity: 68.38%) or NAC1 (AUC: 0.739; 95% CI: 0.664, 0.813; F1-score: 0.611; sensitivity: 53.33%; and specificity: 86.32%) (p<0.01) was used. As a noninvasive prediction tool, DBNN can achieve outstanding results in the early prediction of NAC response in patients with LABC when combining the US data of NAC pre and NAC1.
Introduction
Breast cancer is the most common cause of cancer-related death among women worldwide (1). Neoadjuvant chemotherapy (NAC) has been used as a systematic preoperative treatment for patients with locally advanced breast cancer (LABC) (2). NAC has the advantage of downsizing breast cancers, thus allowing breast-conserving surgery and assessments of the response to chemotherapy during treatment. The achievement of pathologic complete response (pCR) may be a potential independent predictor of better disease-free survival (DFS) and overall survival (OS), especially in patients with triple-negative and human epidermal growth factor 2 (HER2)-enriched breast cancer (3). However, even with the continuous improvements in chemotherapy regimens, the number of patients who achieve pCR remains low (4). Due to the different molecular types and histopathology of breast cancer, the response to chemotherapy may be different. Therefore, identifying patients with superior responses to NAC early has naturally become one of the current hotspots of study.
The optimal method for monitoring the response to NAC has not been established (5). Imaging examination can be used as one of the primary assessment methods. Magnetic resonance imaging (MRI), US, and positron emission tomography (PET)/computed tomography (CT) have been used as evaluation tools (5–7). However, imaging examinations have limitations when used clinically because image interpretation is mainly based on a radiologist’s visual assessment and is not standardized. Furthermore, MRI and PEC/CT are expensive, and PEC/CT is radioactive, making them impractical for frequent scans of patients receiving NAC. Among those methods, ultrasound (US) may become the primary monitoring tool due to its reusability, versatility, sensitivity, and safety.
With the continuous development of deep learning, computer-aided diagnosis (CAD) has become an important research topic, especially in breast cancer research. CAD research has involved the classification (8), segmentation (9), and detection (10) of breast tumours. Especially for classification tasks, which mainly focus on the differentiation of benign and malignant breast tumours, CAD has attracted increasing attention from researchers (11). Deep convolutional neural networks (CNNs) have been widely applied to many healthcare and medical imaging works, leading to state-of-the-art results (12–16). The classification operation procedure of a CNN is that an input image is fed into the CNN to learn essential features and save these parameters as weights and biases to classify images (17). Recently, with the help of deep learning methods, there have been several published studies for predicting breast cancer treatment responses based on PET/CT and MRI images (18–20). El Adoui M et al. introduced a two-branch CNN for the early prediction of breast cancer response to chemotherapy using DCE-MRI volumes acquired before and after chemotherapy (18). Braman N et al. developed a CNN for predicting pCR to HER2-targeted NAC with pretreatment DCE-MRI (19). Choi J H et al. used a CNN algorithm based on Alexnet to predict responses to NAC for advanced breast cancer using PET and MRI images (20). Those studies have shown that deep learning has emerged as a promising tool for breast cancer response prediction.
High-resolution breast US images contain rich texture and echo features that, when combined with deep learning techniques, may potentially be used to achieve a highly accurate and noninvasive NAC response detection method. At present, there are some studies about the use of CAD with US images for predicting the response of breast cancer to NAC (21–23). However, most of these studies focus on feature engineering work based on semiautomatic intermediate steps, and the technique is labour intensive and time consuming. The accuracy of a deep network has far exceeded that of a traditional machine learning method based on handcrafted features (8). However, in the learning process of existing deep learning models, the correlation and importance of the data during different chemotherapy courses have been ignored, and the characteristics of the data have not been well grasped. The purpose of our study is to construct a novel deep learning-based approach named the dual-branch convolutional neural network (DBNN) based on US images at different stages of chemotherapy for the early prediction of NAC in patients with LABC.
Methods
Study Participants
This retrospective single-centre study was approved by the Ethics Committee of ShangHai RenJi Hospital (ShangHai P.R. China), and the requirement for written informed consent was waived. Between February 2015 and June 2019, we enrolled 132 women with LABC who were treated with NAC and surgical resection at our institution. The eligibility criteria were as follows: (a) patients with breast cancer aged 18 to 80 years; (b) patients with histologically confirmed breast cancer and no history of treatment for breast cancer; (c) patients for which US was performed during NAC; and (d) after NAC, the patients underwent surgery and a pathological evaluation was performed. Of the 132 patients, 18 were excluded for the following reasons: (a) US was performed at an outside hospital (n= 3); (b) no midtreatment US data were available (n= 12); and (c) the US images were of poor quality (n=3). A total of 114 patients (age range: 26-72 years; mean age: 49.92 years) comprised the study group. (Figure 1).
Figure 1 Flowchart for the study. LABC, Locally Advanced Breast Cancer; NAC, Neoadjuvant Chemotherapy; US, Ultrasound; pCR, Pathologic Complete Response; non-pCR, non-Pathologic Complete Response.
US Examination
The ultrasonography examinations were performed using MyLab Twice (Esaote, Genoa, Italy) with a 4–13-MHz LA523 linear transducer by an experienced radiologist at the Department of Ultrasound (C.F.W. with 10 years of experience in breast US). In this study, US images were collected before and after the first course of chemotherapy. The US images of pCR and non-pCR samples collected at different treatment stages are shown in Figure 2. The primary dataset called Renji NAC (RJNAC) contains 1936 (968×2 stages) US images (800×608 pixels) at different treatment stages, including 968 US images at each stage, with an average of 16 to 20 images per patient. For the prediction of pCR, the dataset was randomly split into training data (80%) and test data (20%) (a ratio of 4:1). That is, when dividing the dataset, the pCR and non-pCR ratios in the samples were kept close. In the training set and the test set, the pCR and non-pCR ratios were both approximately 0.63. Specifically, each stage of the training set contained 776 images, including 300 pCR images and 476 non-pCR images, while the test set contained 192 images, including 75 pCR images and 117 non-pCR images. (Figure 1).
Figure 2 Two sets of tumour US images corresponding to different stages of NAC. (A) a set of images of pCR. (B) a set of images of non-pCR. NACpre, US images before chemotherapy; NAC1, US images after the first stage of chemotherapy.
Data Preprocessing
The data collected in this study are ultrasonic video data. To input it into the neural network, we perform a video frame cutting operation on the video data (24–26). Four preprocessing steps are applied before starting the training process. As detailed in Figure 3, the first step is to cut the video with different time lengths according to the fixed frame interval to form an indefinite number of M ultrasonic images. The second step is to select N high-quality breast tissue images by removing some images containing artifacts, blur, and non-lesion tissue. Blind to the patients’ private information and pathological results, two professional radiologists (Q.D. and C.F.W. with five and ten years of experience in breast US, respectively) independently read the breast US images. They reach a consensus through discussion to ensure the correctness and repeatability of the dataset. The N of two stages of each patient must be the same but can vary for different patients, depending on how many clear and usable mass images were contained in the indefinite number of M images of different patients. The change of N among different patients does not affect the model learning. N images of two stages are paired sequentially to ensure that the image pairs of each pair are closest in the video time sequence. The third step is that, after removing the nonrelevant breast tissue information, such as the model number of the instruments, time of scanning or imaging, and patient information, we retain the remaining information as a region of interest (ROI). In addition, the resolution of ROI images obtained after video processing is consistent with the resolution of ROI images obtained by static single frame cropping, both of which are 445×445 pixels. Finally, we use the median filter (27) to denoise the US images and preserve edge information. All US images are represented as greyscale images with sizes of 128 × 128 before being fed into the deep neural network.
Dual-Branch Convolutional Neural Network
In the prediction of NAC response, the existing studies failed to take advantage of the correlation among multistage data and the importance of data at each chemotherapy stage (5, 28–30). To solve this problem, we developed a model named DBNN based on feature sharing and weight assignment to predict chemotherapy response by utilizing US images before and after the first stage of chemotherapy (NACpre and NAC1, respectively). Dual branches were designed to extract data features from NACpre and NAC1. There are feature-sharing modules between different branches so that the model could fully use the correlation of the data from each stage. In addition, the model has a weight assignment module, which considers the importance of different branch features and provides prior knowledge for accurate classification.
As shown in Figure 4, the DBNN architecture is composed of two branches that take a 128 × 128 breast tumour ROI cropped from NACpre and NAC1 images as input. Each path contains four convolution blocks, which contain nine convolutional layers in total. Batch normalization layers (31) follow each convolutional layer to speed up network convergence, and a rectified linear unit (ReLU) activation function (32) is used to increase the nonlinearity of the network. Then, these layers are followed by four max-pooling layers (33), where each max-pooling layer is used to perform image downsampling. Furthermore, DBNN has two fully connected layers for feature weighting, and features are shared between each branch by feature fusion.
The details of DBNN feature sharing are shown in the black dotted box in Figure 5. DBNN consists of four convolutional blocks, and the input of each block is the output of the previous block (except for Block 1, where the input is US images from NACpre and NAC1). Sixty-four kernels are used for each convolutional layer in Block 1, 128 for each layer in Block 2, 256 for each layer in Block 3 and 512 for each layer in Block 4, and each kernel has a size of 3 × 3. An US image is input into the respective branch at each stage. Then, the fusion feature map is trained through the convolutional layer, batch normalization layer, and ReLU function and finally downsampled and input into the other blocks until the convolution operation is completed.
First, the network starts from the input layer and is expressed as:
where X denotes the input of NACpre and Y denotes the input of NAC1. Then, C0 and are input to their respective convolution layers, and features are extracted through the convolution kernel. Finally, the feature maps C1 and are generated. The formula is expressed as:
where Ci and represent the feature maps of layer i, i ϵ{1,3,5,7,8}. σi and indicate the ReLU activation function, ωi and stand for the network weights of layer i of the two paths, bi and are network biases for the convolution layer, and * denotes the convolution operation. Ci-1 and are used as inputs of the next layers, Ci and , respectively.
where Cj and represent the feature maps of layer j, j ϵ{2,4,6,9}. Cj-1 and are used as inputs of the next layers, Cj and , respectively.
After each convolution block, we obtain Ck and and input them into the max-pooling layer to reduce the number of parameters of the feature map:
where Ck and represent the feature maps of layer k, k ϵ{2,4,7,9}.
In contrast to the fusion method in the fully connected layer, DBNN shares the features between each branch; that is, it uses fusion when extracting low-level features. As a result, the model could be trained effectively to screen out crucial features, including changes in lesion areas before and after NAC treatment, thus affecting the prediction results of chemotherapy response.
As shown in Figure 6, the weight fusion strategy of DBNN is uncomplicated, and the black dotted box shows the details of the red dotted box. First, the feature vector F(X) from the NACpre branch and the feature vector F(Y) from the NAC1 branch are input, and then the updated feature vectors F(X') and F(Y') are obtained by multiplying the two feature vectors by α(0.2) and β(0.8), respectively. Finally, the sum operation is performed on the updated features to obtain the feature vector F(Z) which is fused with the two branches. The process is expressed by the formula:
After the fully connected layer, we used a dropout strategy (34) (with a rate of 0.5), which helps to prevent the model from overfitting during training. Then, the two branches were summed after the fully connected layer with 1024 hidden units, and a softmax function was applied for pCR classification.
The performance of machine learning algorithms is primarily affected by their hyperparameters because their performance will be inferior without optimal hyperparameter values (35). In particular, the deep learning model relies on good hyperparameter values to accelerate the convergence of the model and achieve optimal performance. To compile and evaluate each model, we use cross entropy (36) as the loss function and a standard accuracy metric that calculates the mean accuracy rate across all predictions. Table 1 shows the hyperparameter setup. The loss curves show no overfitting or underfitting in our model (Figure 7).
All experiments were performed on a Dell T640 tower server deep learning workstation with two NVIDIA GeForce RTX 2080Ti independent graphics cards and two Intel Xeon Silver 4110 CPUs, with RAM extended to 64 GB. The experimental platform was in Python version 3.7. DBNN was implemented by PyTorch, which is a deep learning platform.
Histopathologic Assessment
A pathologist with more than 20 years of experience in breast pathology assessed the histologic results. All pathologic results from outside biopsies were reviewed at our institution. Tumour pathologic characteristics were obtained from histopathologic reports of US guided core biopsies performed before NAC. The histologic type, grade, and expressions of HER2, the oestrogen receptor (ER), the progesterone receptor (PR), and antigen Ki67 were assessed. Tumours with >1% nuclear staining were denoted as ER/PR positive. The cut-off point for Ki-67 high expression was 30%. In terms of HER2 expression, tumours were considered HER2 negative if they had a score of 0 or 1+ during the immunohistochemical (IHC) examination, and a score of 3+ indicated that the tumour was HER2-positive. If the HER2 status was equivocal (IHC score: 2+ or 1+ to 2+), further investigation using in situ hybridization (ISH) was required. In our study, pCR was defined as no residual invasive carcinoma in the breast at surgical resection. Molecular subtypes were classified according to the St. Gallen Consensus (38).
Statistical Analysis
Our statistical analysis was performed using IBM SPSS Statistics 22 (Armonk, NY, USA). Clinicopathological characteristics and US images before and after the first stage of chemotherapy, including maximum tumour diameter and tumour histologic type, were collected. The continuous variables were described as the range, mean and standard deviation, while the categorical variables were reported as counts with percentages. T-tests, chi-squared tests, or Fisher exact tests for independent samples were used to determine significant differences between the pCR and non-pCR groups. To evaluate the performance of the developed models, we calculated six performance metrics: accuracy, sensitivity, specificity, positive predictive value (PPV), negative predictive value (NPV), and F1-score. The predicted performance was assessed by using receiver operating characteristic (ROC) curves, and the area under the curve (AUC) scores were compared. Then, the results were analysed to select the best model to predict NAC response in patients with breast cancer utilizing breast US images. P <.05 was considered to indicate a significant difference. The performance results of the model and other methods were compared by using the Mann-Whitney U test. The 95% CIs for AUC were estimated by using the DeLong method (39–41). Statistical computing was implemented with the Scipy package, a Python-based open-source data processing tool. For the prediction of pCR, DBNN was trained on the training set and then validated on the test set.
F1-score conveys the balance between PPV and sensitivity. The closer the value is to 1, the better the performance of the method. The F1-score equation is defined as follows:
Results
Patient Characteristics
One hundred and fourteen women comprised the final study group (age range: 26-72 years; mean age: 49.92 years). The median maximum diameter of the tumours in the pretreatment US images was 3.82 cm (range: 1.35-8.2cm). The patient characteristics and the sizes of the tumours in the pCR and non-pCR groups are listed in Table 2. Of the 114 patients, 39 (34.2%) achieved pCR at the final pathologic evaluation. No significant differences were found in age, molecular subtype, or maximum tumour diameter between the pCR and non-pCR groups. For the 39 patients who achieved pCR, no residual invasive carcinoma in the breast or axillary lymph nodes was found in 37 (94.87%) patients. Thirty-seven (85.29%) patients showed no evidence of malignant cells in the breast, and 2 (8.82%) patients showed only ductal carcinoma in situ. Of the 75 patients with non-pCR, partial response was observed in 72 patients, and disease stability was observed in 3 patients. Disease progression was not observed for any patient in this study cohort. The pCR group showed a higher proportion of ER negativity (20 [51.3%], P <0.001), PR negativity (13 [33.3%], P=0.029), HER2 positivity (20 [51.3%], P=0.042) and Ki-67 high expression (36 [92.3%], P=0.044) than the non-pCR group. There were significant differences in molecular types between the pCR and non-pCR groups (P<0.001), although luminal B was the main molecular type. The patient characteristics of the tumours in the training and test cohorts are listed in Table 3. There was no significant difference in the expression of biomarkers (i.e., ER, PR, HER2, and Ki-67) between the training and test cohorts. Of the 39 patients who achieved pCR at the final pathologic evaluation, 30/91 (32.97%) patients and 9/23 (39.13%) patients achieved pCR in the training and test sets, respectively (Table 4).
Table 4 Clinical characteristics of the training and test sets containing pCR and non-pCR breast cancer patient data.
Performance Analysis of DBNN Feature Sharing
As mentioned above, it can be understood that the number of layers in a CNN has a specific impact on the prediction and classification performance of the model. Thus, CNNs with different numbers of layers were designed in this experiment. The experimental results were compared to determine the best layer number for the dual branch network. The performance of different convolution layer numbers is shown in the first five rows of Table 5. It can be seen that with the deepening of the network, the performance indices of the dual branch model increased first and then decreased in general. Here, X denotes the number of layers of each branch network in CNN-X. CNN-9 performs the best out of the models with different numbers of layers, and it has an accuracy of 81.77%. Moreover, it also ranks the highest in specificity, PPV, and F1-score. Therefore, in this study, the nine-layer CNN was selected as the backbone of the model. Next, the influences of feature sharing and the weight assignment strategy on the model are explored.
Table 5 Performance of the model with different convolution layer numbers and feature-sharing methods.
At present, there are many methods of feature sharing, including feature element sum and feature concatenation, which are the classic feature fusion methods (42–46). Thus, we also explored the influence of two different strategies on model performance. In the last two rows of Table 5, the performance comparison results of the model with different feature-sharing strategies are shown. CNN-9 FSS represents the CNN model that uses the feature element sum method, while CNN-9 FSC represents the CNN model that uses the feature concatenation method. Table 5 shows that the model achieves better performance when the feature element sum method is used. The accuracy, sensitivity, NPV, and F1-score values were higher than those obtained by the CNN with feature concatenation and CNN-9 without feature sharing. Therefore, DBNN adopts the feature element sum method as its feature-sharing method.
Weight Assignment of DBNN Feature Connection
DBNN is a dual-branch network with two inputs and one output, and the two inputs are NACpre and NAC1 chemotherapy data. The output is the probability of predicting pathological results. Therefore, a feature map from each branch network needs to connect the features and then maps from a high-dimensional vector to a low-dimensional vector to complete the classification task. We compared the experimental results of the feature element sum method, feature concatenation method, and feature weight assignment method of the dual-branch network to explore different feature connection methods (see Table 6). CNN-9 FSS_concat represents the CNN model with the feature concatenation method, and CNN-9 FSS_sum represents the CNN model with the feature element sum method. CNN-9 FSS (A, B) represents the CNN model with the weight connection method, where A is the weight of the NACpre branch and B is the weight of the NAC1 branch.
As shown in Table 6, when the feature weight of the NACpre branch is 0.2 and when that of the NAC1 branch is 0.8, the model’s performance is the best, with an accuracy of 87.50%. In addition, the F1-score is higher than that of the other models, which may be because NAC1 stage data contributed more to the prediction than NACpre stage data. It can be seen from the last nine rows of Table 6 that the average accuracy and F1-score values are superior when the NAC1 branch is heavier than the NACpre branch. Therefore, the method of weight connection is adopted in the model, and the experimental results show that this method can achieve the best results. In the following experiments, CNN-9 FSS (0.2, 0.8) is called DBNN.
Results of DBNN Data Augmentation
As stated earlier, there was a data imbalance problem in RJNAC. The amount of data with non-pCR pathological results was approximately twice that with pCR pathological results, affecting the model’s performance. Therefore, we explored the impact of different data augmentation strategies on the performance of DBNN. The experimental results were compared using nonaugmented data, geometrically transformed data (47), Mixup data (48), and small amounts of upsampled data. Geometric transformation techniques include rotations, flips, and zooming to generate new training samples to maintain realistic tumour shapes. Moreover, small amounts of data upsampling techniques apply geometric transformations to non-pCR examples to achieve a quantity balance between the two categories, solving the data imbalance problem manually.
As seen from Table 7, the performance of the model is better without data augmentation. First, it can be seen that the performance of the model on nonaugmented data was better than that of the model on geometrically transformed data. Augmenting both types of data aggravate the data imbalance, leading to degradation in the performance of the model; hence, Mixup data augmentation also degrades model performance. In addition, Mixup may not be suitable for the augmentation of medical datasets because it disturbs the relationship between a lesion and the surrounding area, making the model learn incorrect information. Finally, we enhance the sample size of the two types of data so that they are consistent by sampling small numbers of samples. The experimental results on the augmented data were not as good as the results on the nonaugmented data. Perhaps DBNN learns the redundant features of the data during the learning process, resulting in model performance degradation.
Comparison With the Single Branch Models
To further validate the potential of DBNN in predicting the efficacy of NAC, it was used to predict the pathological classification of patients early based on the different stage data of NAC treatment in the RJNAC dataset. Compared with the AUC value in the first two rows and the last row in Table 8, we know that the model’s prediction results when using a single branch network for single-stage data were not as good as those when using multistage data. In addition, the performance of the model trained on the NAC1 data was slightly superior to that trained on the NACpre data when using single-stage data, which indicates the necessity of DBNN weight assignment. From Table 8 and Figure 8, we can see that the areas under the ROC curve for NACpre (Azpre), NAC1 (Az1) and NACpre+NAC1 (Azpre+1) were 0.730, 0.739 and 0.939, respectively. The performance of the model trained on the NAC1 data shows higher specificity than that trained on the NACpre data. The sensitivity of the model trained on NACpre was superior to that trained on NAC1 data. The value of Azpre+1 was significantly higher than that of Azpre and Az1 (P <0.01). However, there was no significant difference between the values of Azpre and Az1 (P =0.3244).
Moreover, some more sophisticated deep learning models were tested for single branch classification due to CNN-9 was used to train single branch models. The experimental results are shown in Table 9. The AUC of CNN-9 on NACpre data was the highest (AUC=0.730), and the AUC value on NAC1 data was very close to the optimal value (0.739 vs. 0.756). Therefore, we believe that CNN-9 can be used as a representative of the classical single branch network.
Comparison With the Latest Studies
At present, there are few studies on the prediction of NAC response for breast cancer based on US images, and the datasets used in each study and each imaging protocol are different, so it is difficult to compare the results directly. However, to verify the research value of DBNN, this study referred to the four latest papers, reproduced the methods according to the technical details described in the articles, and applied them to the RJNAC dataset (5, 18, 19, 28). Two identical Inception-ResNet-V2 CNNs based Siamese models without fine-tuning were reimplemented to extract generic features. Then the difference between the feature vectors was used to train a logistic regression model for the prediction (5). We reimplemented a two-input CNN, in which each input branch consisted of four blocks of 2D convolution layers, each followed by a ReLU activation function and max-pooling layer. A dropout layer was applied after every two convolutional blocks. Then, the two branches were concatenated after a fully connected layer followed by ReLU, dropout (with a rate of 40%), and a Sigmoid function for the final classification (18), while two dense layers were processed to yield the final output (19). The developed multi-input deep learning architecture contained two parallel sub-architectures with similar layers to the single architecture, consisting of six blocks with multiple convolutional layers, each followed by a ReLU activation function and max-pooling layer. Then, a concatenation was applied between two single architectures, a dropout of 50%, and a fully connected layer was used at the end of the network to provide a classification result (28). In these studies, three of the approaches were based on MRI data (18, 19, 28), and one was based on US image data (5). In Figure 9, the area under the ROC curve for DBNN (AzDBNN) was significantly higher than that of AzByra (5) (P =0.004). However, there was no significant difference in the values of AzDBNN and the values of the area under the ROC curve for the other methods (Table 10). We also show the prediction results and pathology labels of the model on NACpre and NAC1 images and the probability of the model output prediction results in Figure 10.
Discussion
The early prediction of chemotherapy response in patients with breast cancer is crucial for improving and personalizing patient treatment. In this study, a novel deep learning method, DBNN, based on US images for the early prediction of NAC response in patients with breast cancer was proposed and validated. The experimental results showed that the best prediction performance was obtained with the DBNN model using feature sharing and weight assignment. It was worth noting that all the performances shown in Tables 5–10 were from the test set. The highest diagnostic performance was obtained when the US image information of NACpre and NAC1 was combined, in which the accuracy, sensitivity, specificity, F1-score, and AUC values were 87.50%, 90.67%, 85.67%, 0.850, and 0.939, respectively.
The DBNN approach for the early prediction of NAC proposed in this study has several advantages.
First, compared to the previous traditional machine learning methods, which mainly depend on feature engineering and require domain knowledge to build feature extractors, our deep learning approach is automatic and does not require feature engineering. Methods based on machine learning are limited in their function, as they are dependent on handcrafted features. Moreover, our model considers not only the tumoral region but also the tumour’s surrounding tissue by using entire breast tumour images. Supplementing the US features extracted from a tumour itself with features computed within the tumour’s surrounding tissue, such as the peritumoural region, may improve the prediction of pCR from US images (49, 50). Second, different from the existing deep learning algorithms, DBNN fuses features of each branch in the process of extracting low-level features, which may effectively screen out important features through the training network to achieve more accurate early prediction results. Third, in contrast to the existing methods for predicting NAC response using two-stage data, we assume that the importance of the data before and after chemotherapy is inconsistent. Therefore, DBNN introduces the weight assignment strategy to increase the weight of data features after chemotherapy by using prior knowledge to guide network training to affect the NAC response prediction results.
It is difficult to directly compare our results to those of other methods reported in other studies due to different data acquisition techniques, analysis protocols and subject groups. Moreover, there are few studies that use deep learning for NAC response early prediction in breast cancer based on US images. Nevertheless, we can compare our results with those of models trained on our datasets. The studies performed by El Adoui et al. (28), Braman et al. (19), and El Adoui et al. (18) were based on MRI data, and the study designed by Byra et al. (5) was based on US image data. All four methods are two-input CNN architectures for the prediction of breast tumour NAC response from follow-up images. Each branch was operated on by a series of convolution-based operations and summarized into a set of deep features, which were then combined and processed by the feature fusion of two branches to generate a final score representing the response probability. However, those methods only considered the late fusion of deep features. The models cannot effectively share data features at different stages in their respective branches and may even filter out crucial features, such as changes in lesion areas. Therefore, they cannot make full use of the relationship between different data for model training. In Table 10, comparisons of the performance of the state-of-the-art methods and our method were made based on seven indices: accuracy, sensitivity, specificity, PPV, NPV, F1-score and AUC. Our method obtained better results on most of the evaluation indices. The ROC curves based on the true positive rate (TPR) and false positive rate (FPR) for the existing methods and our proposed method are shown in Figure 9. The AUC values of all the algorithms were over 0.8, and the largest AUC value (0.939) was obtained by our model. The area under the ROC curve obtained by DBNN (AzDBNN) was significantly higher than that obtained by AzByra (5) (P =0.004). The model developed by Byra et al. (5) was based on a small dataset with images from 30 patients, while our dataset contained images from 114 patients. We can train our deep learning model from scratch because a model pretrained on natural images is often not the best model when applied to medical images. Moreover, we shared the data features of the two streams in the training process and assigned the weights of the different stages by using prior knowledge to obtain more accurate results.
Although the proposed method has improved the prediction accuracy of NAC response, there are still some limitations in this study. First, due to the small dataset of US images collected from a single centre, the model’s generalization ability needs to be further improved. Since there is currently no public dataset of ultrasound images before and after the first stage of chemotherapy for NAC, our next work will continue to collect data from multi-centres to further verify our model’s generalization ability. It is generally accepted that the larger a dataset is, the better the performance of the deep learning models (51, 52). Limited datasets are a prevalent challenge in medical image analysis. Second, due to the heterogeneous nature of the histopathologic and molecular subtypes of breast cancer included in our study, the pathologic response to NAC may be affected and may cause selection bias. Finally, we did not add breast cancer molecular subtype to our method, which may help to predict the response of breast cancer to NAC early. The application of DBNN is only in the primary stage. Therefore, how to extend our method to clinical decision-making is worthy of in-depth study.
In the future, there will be at least two aspects of NAC response prediction models based on different stages of data that can be further developed. On the one hand, DBNN should also consider more feature methods, such as combining low-level features and high-level features by utilizing residual cross-branch connections. Moreover, adaptive weight allocation can be regarded as the weight assignment strategy. On the other hand, the robustness and generalization ability of DBNN need further verification.
In conclusion, our study proposes a novel dual-branch DBNN model based on feature sharing and weight assignment to predict the efficacy of NAC treatment for breast cancer utilizing greyscale US images. DBNN has two remarkable advantages: feature sharing and weight assignment. Feature sharing can make the model consider the correlations between data in different stages of NAC during training. Moreover, weight assignment, which provided prior knowledge, emphasizes the importance of data at different NAC treatment stages. The results show that DBNN has the potential to enable the early prediction of pCR and achieved good prediction performance when applied on NACpre and NAC1 data. However, a further large-scale study with an independent external validation dataset is needed before this approach can be used for actual clinical decision-making, and it may become an important monitoring tool for the early prediction of the response to NAC in patients with breast cancer.
Data Availability Statement
The data analyzed in this study is subject to the following licenses/restrictions: Due to the privacy of patients, the related data cannot be available for public access. Requests to access these datasets should be directed to Caifeng Wan, d2FuY2FpZmVuZ2t5QHNpbmEuY29t.
Ethics Statement
The studies involving human participants were reviewed and approved by Shanghai Jiao Tong University School of Medicine, Ren Ji Hospital Ethics Committee. Written informed consent was waived in this study.
Author Contributions
Conception and design: JX, CD, and XS. Collection and assembly of data: CW, QD and CD. Verification of the underlying data: HS, XS, and JW. Development of methodology: JX, HS, XS, and JW. Data analysis and interpretation: JX, HS, CW, XS, QD, and CD. Writing original draft: JX, HS, CW, QD, CD, and JW. All authors contributed to the article and approved the submitted version.
Funding
This work was supported by National Natural Science Foundation of China (Grant No.61873156, Grant No. 81801697, Grant No. 81571678).
Conflict of Interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher’s Note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Siegel RL, Miller KD, Jemal A. Cancer Statistics, 2018. Ca-Cancer J Clin (2018) 68(1):7–30. doi: 10.3322/caac.21442
2. Das U, Lakshmaiah KC, Babu KG, Suresh TM, Lokanatha D, Jacob L, et al. The Actual Scenario of Neoadjuvant Chemotherapy of Breast Cancer in Developing Country: A Report of 80 Cases of Breast Cancer From a Tertiary Cancer Center in India. J Cancer Res Clin Oncol (2014) 40(10):1777–82. doi: 10.1007/s00432-014-1724-1
3. Spring LM, Fell G, Arfe A, Sharma C, Greenup R, Reynolds KL, et al. Pathologic Complete Response After Neoadjuvant Chemotherapy and Impact on Breast Cancer Recurrence and Survival: A Comprehensive Meta-Analysis. Clin Cancer Res (2020) 26(12):2838–48. doi: 10.1158/1078-0432.CCR-19-3492
4. Matthews CM, Nymberg K, Berger M, Vargo CA, Dempsey J, Li J, et al. Pathological Complete Response Rates With Pertuzumab-Based Neoadjuvant Chemotherapy in Breast Cancer: A Single-Center Experience. J Oncol Pharm Pract (2020) 26(3):572–9. doi: 10.1177/1078155219857800
5. Byra M, Dobruch-Sobczak K, Klimonda Z, Piotrzkowska-Wroblewska H, Litniewski J. Early Prediction of Response to Neoadjuvant Chemotherapy in Breast Cancer Sonography Using Siamese Convolutional Neural Networks. IEEE J BioMed Health (2020) 25(3):797–805. doi: 10.1109/JBHI.2020.3008040
6. Sutton EJ, Onishi N, Fehr DA, Dashevsky BZ, Sadinski M, Pinker K, et al. A Machine Learning Model That Classifies Breast Cancer Pathologic Complete Response on MRI Post-Neoadjuvant Chemotherapy. Breast Cancer Res (2020) 22:1–11. doi: 10.1186/s13058-020-01291-w
7. Fantini L, Belli ML, Azzali I, Loi E, Bettinelli A, Feliciani G, et al. Exploratory Analysis of 18F-3’-Deoxy-3’-Fluorothymidine (18F-FLT) PET/CT-Based Radiomics for the Early Evaluation of Response to Neoadjuvant Chemotherapy in Patients With Locally Advanced Breast Cancer. Front Oncol (2021) 11:2315. doi: 10.3389/fonc.2021.601053
8. Xie J, Song X, Zhang W, Dong Q, Wan C. A Novel Approach With Dual-Sampling Convolutional Neural Network for Ultrasound Image Classification of Breast Tumors. Phys Med Biol (2020) 65(24):245001. doi: 10.1088/1361-6560/abc5c7
9. Huang Q, Huang Y, Luo Y, Yuan F, Li X. Segmentation of Breast Ultrasound Image With Semantic Classification of Superpixels. Med Image Anal (2020) 61:101657. doi: 10.1016/j.media.2020.101657
10. Li Y, He Z, Lu Y, Ma X, Guo Y, Xie Z, et al. Deep Learning of Mammary Gland Distribution for Architectural Distortion Detection in Digital Breast Tomosynthesis. Phys Med Biol (2021) 66(3):035028. doi: 10.1088/1361-6560/ab98d0
11. Fujioka T, Kubota K, Mori M, Kikuchi Y, Katsuta L, Kasahara M, et al. Distinction Between Benign and Malignant Breast Masses at Breast Ultrasound Using Deep Learning Method With Convolutional Neural Network. Jpn J Radiol (2019) 37(6):466–72. doi: 10.1007/s11604-019-00831-5
12. Krizhevsky A, Sutskever I, Hinton GE. Imagenet Classification With Deep Convolutional Neural Networks. Adv Commun ACM (2017) 60(6):84–90. doi: 10.1145/3065386
13. Ronneberger O, Fischer P, Brox T. U-Net: Convolutional Networks for Biomedical Image Segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Munich, Germany: Springer (2015). p.234–41.
14. Esteva A, Kuprel B, Novoa RA, Ko J, Swetter SM, Blau HM, et al. Dermatologist-Level Classification of Skin Cancer With Deep Neural Networks. Nature (2017) 542(7639):115–8. doi: 10.1038/nature21056
15. Rajpurkar P, Irvin J, Zhu K, Yang B, Mehta H, Duan T, et al. Chexnet: Radiologist-Level Pneumonia Detection on Chest X-Rays With Deep Learning (2017). Available at: https://arxiv.org/abs/1711.05225v3.
16. Gulshan V, Peng L, Coram M, Stumpe MC, Wu D, Narayanaswamy A, et al. Development and Validation of a Deep Learning Algorithm for Detection of Diabetic Retinopathy in Retinal Fundus Photographs. JAMA (2016) 316(22):2402–10. doi: 10.1001/jama.2016.17216
17. Simonyan K, Zisserman A. Very Deep Convolutional Networks for Large-Scale Image Recognition (2014). Available at: https://arxiv.org/abs/1409.1556.
18. El Adoui M, Larhmam MA, Drisis S, Benjelloun M. Deep Learning Approach Predicting Breast Tumor Response to Neoadjuvant Treatment Using DCE-MRI Volumes Acquired Before and After Chemotherapy. Med Imaging 2019: Computer-Aided Diagnosis (2019) 10950:649–58. doi: 10.1117/12.2505887
19. Braman N, Adoui ME, Vulchi M, Turk P, Etesami M, Fu P, et al. Deep Learning-Based Prediction of Response to HER2-Targeted Neoadjuvant Chemotherapy From Pretreatment Dynamic Breast MRI: A Multi-Institutional Validation Study (2020). Available at: https://arxiv.org/abs/2001.08570v1.
20. Choi JH, Kim H-A, Kim W, Lim I, Lee I, Byun BH, et al. Early Prediction of Neoadjuvant Chemotherapy Response for Advanced Breast Cancer Using PET/MRI Image Deep Learning. Sci Rep-Uk (2020) 10(1):1–11. doi: 10.1038/s41598-020-77875-5
21. DiCenzo D, Quiaoit K, Fatima K, Bhardwaj D, Sannachi L, Gangeh M, et al. Quantitative Ultrasound Radiomics in Predicting Response to Neoadjuvant Chemotherapy in Patients With Locally Advanced Breast Cancer: Results From Multi-Institutional Study. Cancer Med (2020) 9(16):5798–806. doi: 10.1002/cam4.3255
22. Sannachi L, Gangeh M, Tadayyon H, Gandhi S, Wright FC, Slodkowska E, et al. Breast Cancer Treatment Response Monitoring Using Quantitative Ultrasound and Texture Analysis: Comparative Analysis of Analytical Models. Transl Oncol (2019) 12(10):1271–81. doi: 10.1016/j.tranon.2019.06.004
23. Sannachi L, Gangeh M, Tadayyon H, Sadeghi-Naini A, Gandhi S, Wright FC, et al. Response Monitoring of Breast Cancer Patients Receiving Neoadjuvant Chemotherapy Using Quantitative Ultrasound, Texture, and Molecular Features. PloS One (2018) 13(1):e0189634. doi: 10.1371/journal.pone.0189634
24. Sanz-Requena R, Moratal D, García-Sánchez DR, Bodí V, Rieta JJ, Sanchis JM. Automatic Segmentation and 3D Reconstruction of Intravascular Ultrasound Images for a Fast Preliminar Evaluation of Vessel Pathologies. Comput Med Imaging Graph (2007) 31(2):71–80. doi: 10.1016/j.compmedimag.2006.11.004
25. Gao Y, Maraci MA, Noble JA. Describing Ultrasound Video Content Using Deep Convolutional Neural Networks. 2016 IEEE 13th Int Symposium Biomed Imaging (ISBI) (2016), 787–90. doi: 10.1109/ISBI.2016.7493384
26. Youk JH, Jung I, Yoon JH, Kim SH, Kim YM, Lee EH, et al. Comparison of Inter-Observer Variability and Diagnostic Performance of the Fifth Edition of BI-RADS for Breast Ultrasound of Static Versus Video Images. Ultrasound Med Biol (2016) 42(9):2083–8. doi: 10.1016/j.ultrasmedbio.2016.05.006
27. Gallagher N, Wise G. A Theoretical Analysis of the Properties of Median Filters. IEEE Trans Acoustics Speech Signal Process (1981) 29(6):1136–41. doi: 10.1109/TASSP.1981.1163708
28. El Adoui M, Drisis S, Benjelloun M. Multi-Input Deep Learning Architecture for Predicting Breast Tumor Response to Chemotherapy Using Quantitative MR Images. Int J Comput Assist Radiol Surg (2020) 15(9):1491–500. doi: 10.1007/s11548-020-02209-9
29. Qu Y, Zhu H, Cao K, Li X, Ye M, Sun Y. Prediction of Pathological Complete Response to Neoadjuvant Chemotherapy in Breast Cancer Using a Deep Learning (DL) Method. Thorac Cancer (2020) 11(3):651–8. doi: 10.1111/1759-7714.13309
30. Ravichandran K, Braman N, Janowczyk A, Madabhushi A. A Deep Learning Classifier for Prediction of Pathological Complete Response to Neoadjuvant Chemotherapy From Baseline Breast DCE-MRI. Med Imaging 2018: Computer-Aided Diagnosis. Int Soc Optics Photonics (2018) 10575:105750C. doi: 10.1117/12.2294056
31. Ioffe S, Szegedy C. Batch Normalization: Accelerating Deep Network Training by Reducing Internal Covariate Shift. In: International Conference on Machine Learning. PMLR (2015) 37:448–56. doi: 10.5555/3045118.3045167
32. Nair V, Hinton GE. Rectified Linear Units Improve Restricted Boltzmann Machines. Proc 27th Int Conf Int Conf Mach Learn (2010), 807–14. doi: 10.5555/3104322.3104425. https://icml.cc/Conferences/2010/papers/432.pdf
33. Giusti A, Cireşan DC, Masci J, Gambardella LM, Schmidhuber J. Fast Image Scanning With Deep Max-Pooling Convolutional Neural Networks. 2013 IEEE Int Conf Image Processing. IEEE (2013), 4034–8. doi: 10.1109/ICIP.2013.6738831
34. Srivastava N, Hinton G, Krizhevsky A, Sutskever I, Salakhutdinov R. Dropout: A Simple Way to Prevent Neural Networks From Overfitting. J Mach Learn Res (2014) 15(1):1929–58.
35. Guo B, Hu J, Wu W, Peng Q, Wu F. The Tabu_Genetic Algorithm: A Novel Method for Hyper-Parameter Optimization of Learning Algorithms. Electronics (2019) 8(5):579. doi: 10.3390/electronics8050579
36. Lopez-Garcia P, Onieva E, Osaba E, Masegosa AD, Perallos A. A Hybrid Method for Short-Term Traffic Congestion Forecasting Using Genetic Algorithms and Cross Entropy. IEEE T Intell Transp (2015) 17(2):557–69. doi: 10.1109/TITS.2015.2491365
37. Kingma DP, Ba J. Adam: A Method for Stochastic Optimization. (2014). Available at: https://arxiv.org/abs/1412.6980
38. Curigliano G, Burstein HJ, Winer EP, Gnant M, Dubsky P, Loibl S, et al. De-Escalating and Escalating Treatments for Early-Stage Breast Cancer: The St. Gallen International Expert Consensus Conference on the Primary Therapy of Early Breast Cancer 2017. Ann Oncol (2017) 28(8):1700–12. doi: 10.1093/annonc/mdx308
39. Clopper CJ, Pearson ES. The Use of Confidence or Fiducial Limits Illustrated in the Case of the Binomial. Biometrika (1934) 26(4):404–13. doi: 10.1093/biomet/26.4.404
40. DeLong ER, DeLong DM, Clarke-Pearson DL. Comparing the Areas Under Two or More Correlated Receiver Operating Characteristic Curves: A Nonparametric Approach. Biometrics (1988) 44(3):837–45. doi: 10.2307/2531595
41. Sun X, Xu W. Fast Implementation of DeLong's Algorithm for Comparing the Areas Under Correlated Receiver Operating Characteristic Curves. IEEE Signal Process Lett (2014) 21(11):1389–93. doi: 10.1109/LSP.2014.2337313
42. Long J, Shelhamer E, Darrell T. Fully Convolutional Networks for Semantic Segmentation. Proc IEEE Conf Comput Vision Pattern recognit (2015) .p:3431–40. doi: 10.1109/cvpr.2015.7298965
43. Hariharan B, Arbeláez P, Girshick R, Malik J. Hypercolumns for Object Segmentation and Fine-Grained Localization. Proc IEEE Conf Comput Vision Pattern Recognit (2015) 447–56. doi: 10.1109/CVPR.2015.7298642
44. Bell S, Zitnick CL, Bala K, Girshick R. Inside-Outside Net: Detecting Objects in Context With Skip Pooling and Recurrent Neural Networks. Proc IEEE Conf Comput Vision Pattern Recognit (2016) p:2874–83. doi: 10.1109/CVPR.2016.314
45. Liu W, Rabinovich A, Berg AC. Parsenet: Looking Wider to See Better. (2015). Available at: https://arxiv.org/abs/1506.04579
46. Kong T, Yao A, Chen Y, Sun F. Hypernet: Towards Accurate Region Proposal Generation and Joint Object Detection. Proc IEEE Conf Comput Vision Pattern recognit (2016), 845–53. doi: 10.1109/CVPR.2016.98
47. Shorten C, Khoshgoftaar TM. A Survey on Image Data Augmentation for Deep Learning. J Big Data-Ger (2019) 6(1):1–48. doi: 10.1186/s40537-019-0197-0
48. Zhang H, Cisse M, Dauphin YN, Lopez-Paz D. Mixup: Beyond Empirical Risk Minimization. (2017). Available at: https://arxiv.org/abs/1710.09412
49. Braman NM, Etesami M, Prasanna P, Dubchuk C, Gilmore H, Tiwari P, et al. Intratumoral and Peritumoral Radiomics for the Pretreatment Prediction of Pathological Complete Response to Neoadjuvant Chemotherapy Based on Breast DCE-MRI. Breast Cancer Res (2017) 19(1):1–14. doi: 10.1186/s13058-017-0846-1
50. Eben JE, Braman N, Madabhushi A. Response Estimation Through Spatially Oriented Neural Network and Texture Ensemble (RESONATE). Int Conf Med Image Comput Computer-Assisted Intervention (2019) 11767:602–10. doi: 10.1007/978-3-030-32251-9_66
51. Halevy A, Norvig P, Pereira F. The Unreasonable Effectiveness of Data. IEEE Intell Syst (2009) 24(2):8–12. doi: 10.1109/MIS.2009.36
Keywords: deep learning, breast cancer, neoadjuvant chemotherapy, pathologic complete response, ultrasound imaging
Citation: Xie J, Shi H, Du C, Song X, Wei J, Dong Q and Wan C (2022) Dual-Branch Convolutional Neural Network Based on Ultrasound Imaging in the Early Prediction of Neoadjuvant Chemotherapy Response in Patients With Locally Advanced Breast Cancer. Front. Oncol. 12:812463. doi: 10.3389/fonc.2022.812463
Received: 10 November 2021; Accepted: 07 March 2022;
Published: 07 April 2022.
Edited by:
Xiang Zhang, Sun Yat-sen University, ChinaReviewed by:
Shuoyu Xu, Southern Medical University, ChinaGuo-Qing Du, Guangdong Academy of Medical Sciences, China
Copyright © 2022 Xie, Shi, Du, Song, Wei, Dong and Wan. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Qi Dong, cmpkb25ncWlAMTYzLmNvbQ==; Caifeng Wan, d2FuY2FpZmVuZ2t5QHNpbmEuY29t
†These authors have contributed equally to this work