Skip to main content

ORIGINAL RESEARCH article

Front. Oncol., 02 February 2024
Sec. Breast Cancer

Identification of ipsilateral supraclavicular lymph node metastasis in breast cancer based on LASSO regression with a high penalty factor

  • 1West China Hospital, Sichuan University, Chengdu, China
  • 2West China Biomedical Big Data Center, West China Hospital, Sichuan University, Chengdu, China
  • 3Med-X Center for Informatics, Sichuan University, Chengdu, China
  • 4Division of Breast Surgery, Department of General Surgery, West China Hospital, Sichuan University, Chengdu, China
  • 5Breast Center, West China Hospital, Sichuan University, Chengdu, China
  • 6Clinical Research Center for Breast Diseases, West China Hospital, Sichuan University, Chengdu, China
  • 7Department of Medical Oncology, Cancer Center, West China Hospital, Sichuan University, Chengdu, China

Aiming at the problems of small sample size and large feature dimension in the identification of ipsilateral supraclavicular lymph node metastasis status in breast cancer using ultrasound radiomics, an optimized feature combination search algorithm is proposed to construct linear classification models with high interpretability. The genetic algorithm (GA) is used to search for feature combinations within the feature subspace using least absolute shrinkage and selection operator (LASSO) regression. The search is optimized by applying a high penalty to the L1 norm of LASSO to retain excellent features in the crossover operation of the GA. The experimental results show that the linear model constructed using this method outperforms those using the conventional LASSO regression and standard GA. Therefore, this method can be used to build linear models with higher classification performance and more robustness.

1 Introduction

The incidence and mortality of female breast cancer in China have been continuously increasing in recent years, posing a large threat to women’s health (1, 2). The survival outcomes of patients with ipsilateral supraclavicular lymph node (ISLN) metastasis after surgery are often unsatisfactory. Identifying ISLN status before surgery allows for categorizing patients. ISLN patients may benefit from initial treatments such as neoadjuvant chemotherapy (3) rather than surgery. Furthermore, breast ultrasound is a routine diagnostic method for breast cancer. This approach is noninvasive, inexpensive, real-time, portable, and radiation-free and provides internal structure imaging that contains rich information on tumor heterogeneity. However, this information is usually difficult to observed visually (4, 5). Radiomics refers to the conversion of medical images into mineable data and the subsequent analysis of these data to provide decision support (4, 5). Therefore, preoperative ultrasound based radiomics technology is an effective noninvasive approach to identify ISLN status before treatment and assist doctors in making clinical decisions.

The relatively low occurrence of ISLN events (4%) results in a limited number of samples that can be collected, especially for single centers (6). Predicting ISLN using an ultrasound radiomics-based approach is evidently a classical data mining problem with a small sample size. Therefore, methods such as deep convolutional neural networks that require a large amount of data and annotations may not be suitable for this study (7, 8). Radiomic feature engineering combined with linear classification models may achieve exceptional performance in terms of model robustness and interpretability (7, 9, 10). To date, there have been no studies on ISLN identification based on preoperative ultrasound radiomics. However, relevant studies have been reported. In 2019, Liu et al. (11) used dynamic contrast enhanced magnetic resonance imaging (DCE-MRI)-based radiomic feature-trained linear logistic regression (LR), extreme gradient boosting (XGBoost), and support vector machine models to identify axillary lymph node metastasis in breast cancer patients. In 2020, Qiu et al. (12) built an LR model based on ultrasound radiomic features to identify axillary lymph node metastasis in breast cancer patients. In the same year, Yu et al. (13) constructed an LR model based on the radiomic features of MRI to identify axillary lymph node metastasis and predicted the disease-free survival of patients with early breast cancer. In the above studies, the clinical problems to be solved were all data mining problems with a small sample size and high-dimensional feature space, and after feature selection, machine learning models were constructed for classification. Further, these studies employed the least absolute shrinkage and selection operator (LASSO) method for feature selection. Then, linear models were constructed to evaluate the classification performance. Although nonlinear models may offer better the training performance, they often have limited generalizability and poor interpretability, making it challenging to apply them in clinical practice. A feasible way to alleviate this issue is to use linear combinations with interpretability to build and train models (9, 10). Currently, the vast majority of radiomics studies use t tests or U tests in combination with correlation coefficients for preliminary feature selection to obtain stable candidate features. Then, conventional (widely used) LASSO regression methods are used for further selection of the candidate features (9, 14).

However, the feature space of datasets in radiomics research is generally a hyperdimensional space composed of thousands of radiomic features, and finding a suitable combination of a finite number of features remains a challenging task (9, 10). LASSO (15) is a linear regression algorithm that can intuitively and explicitly express the representation ability of a feature set. LASSO adds a regularization term to the general linear regression to ensure the best fitting error and to keep the parameters as simple as possible (reducing the number of nonzero parameters) to reduce overfitting. Therefore, LASSO can enable a machine learning model to exhibit strong generalizability and outstanding robustness (16, 17). However, as a multivariate regression method that searches for a feature subset from the feature space, LASSO first searches the entire feature space. When the dimension of the feature space is very high, the choice of the optimal penalty factor lambda (i.e., the weight of the L1 norm or penalty value/penalty factor) can be challenging. A high lambda can keep the parameters as simple as possible, thereby improving the robustness of the constructed linear model. However, the model performance may decrease. On the other hand, a low lambda value may lead to overfitting the model constructed in the next step. Liu et al. (9) proposed preselecting features based on statistical testing, i.e., preselecting candidate features. In this process, first, the total dimensionality of the feature space is reduced by defining a feature subspace in advance. Then, LASSO is used to search for the optimized feature combination within this subspace. This approach is widely used in radiomics studies (9, 10). Nevertheless, on the one hand, the predefined feature subspace may still be a high-dimensional feature space. On the other hand, LASSO only searches for feature combinations within this feature subspace, which is still redundant to include more candidate features. A more optimized solution is to further extract subspaces in this feature subspace multiple times and use LASSO to search for feature combinations in each of these extracted subspaces. To ensure completeness, extracted subspaces must cross each other, a requirement similar that when sampling with replacement.

This study aims to address the issue of small sample size and high-dimensional feature space in the identification of the ISLN metastasis status of breast cancer patients by ultrasound radiomics. To this end, we combine the genetic algorithm (GA) with LASSO and propose GALambda, in which the crossover operation of individuals is carried out based on LASSO regression with a high penalty factor. In this approach, the GA is used to extract feature subspaces as well as evaluate and iterate over these extracted subspaces to determine if LASSO can find feature combinations that can construct high-performance linear prediction models within these subspaces. Furthermore, for the GA, a strategy is designed to dynamically set excellent genes based on the high penalty parameter lambda. Thus, the excellent genes are preserved for crossover to generate offspring in the hope that the search for feature combination can be optimized.

2 Materials and methods

2.1 Dataset

2.1.1 Study population

Female patients diagnosed with primary breast cancer with ISLN metastasis at West China Hospital between December 2010 and May 2020 were retrospectively included in this study. The diagnosis of breast cancer and ISLN metastasis was confirmed by preoperative biopsy or postoperative pathological examination of ISLN specimens. The inclusion criteria were as follows: (1) female patients with primary breast cancer; and (2) patients who underwent preoperative ISLN biopsy or postoperative pathological examination of ISLN specimens. The exclusion criteria were as follows: (1) lack of ultrasound image records; and (2) negative ISLN biopsy results without postoperative pathological records. A total of 181 patients were finally included in this study. These patients were randomly divided into a training set (n=109) and an independent test set (n=72) at a ratio of 6:4. The results of the statistical tests indicated that there were no significant differences in important baseline features (age, T-stage, invasive ductal carcinoma, progesterone receptor status, estrogen receptor status, and intraductal carcinoma in situ of the breast) between the training and test sets (p≥ 0.05). In the statistical test of significant differences, Fisher’s exact test was used for categorical variables and the Mann–Whitney U test (also known as the Mann–Whitney rank sum test) was used for continuous variables. The p values for the above six baseline features were 0.7256, 0.3979, 0.1322, 0.4391, 0.2079, and 0.4429, respectively.

2.1.2 Ultrasound images

All patients in this study underwent preoperative breast and lymph node ultrasound. The ultrasound equipment used were as follows: an Acuson S3000 with an 18-5 MHz linear transducer (Siemens, Munich, Germany) and an iU22 with a 12-5 MHz probe (Philips, Amsterdam, Netherlands). The stored ultrasound images should (1) be from the most recent ultrasound examination before ISLN status confirmation through biopsy or dissection, (2) contain the entire tumor or lymph node lesions and as many features as possible, (3) clearly show the tumor lesions, and (4) be obtained by an experienced sonographer. Two sonographers with more than 10 years of experience in breast ultrasound outlined primary lesions as the region of interest (ROI) on the ISLN ultrasound images.

2.2 Methods

The technical route of this study, which is also the conventional technical roadmap of radiomics, is shown in Figure 1. Both feature selection and training were performed on the training set. A five-fold cross-validation (CV) strategy was used during the training. The model evaluation was primarily based on the results from CV and independent testing. GALambda is an improved GA proposed in this study based on LASSO regression that obtains the features corresponding to nonzero coefficients by selecting a high penalty factor from the LASSO CV results. GALambda dynamically sets excellent genes for the crossover operation on individuals in the GA based on the minimum mean square error from CV. As a result, the features that are retained under the high penalty factor in each combination from multivariate LASSO regression can be carried over to the new individuals generated. This algorithm will be described in more detail in the subsequent section.

Figure 1
www.frontiersin.org

Figure 1 Technical route of identifying ISLN by ultrasound radiomics.

2.2.1 Radiomic feature extraction

Quantitative radiomic features were extracted from the ROIs and the subbands of filtered ROIs. The filtering methods included the Laplacian of Gaussian (LoG), wavelet transform, and contourlet transform methods. The quantitative feature extraction methods included the histogram, cooccurrence matrix, run-length matrix, difference matrix, size-zone matrix, neighborhood difference matrix, and neighborhood dependency matrix methods. Shape features were also extracted. The above methods were applied again to the subbands of the filtered ROIs, and there were a total of 25 feature extraction methods, resulting in a total of 5031 features extracted from a single ROI (1820).

2.2.2 GALambda: GA and LASSO regression

In feature selection, the GA can be used to search randomly and in parallel for a feature combination in the feature space so that the objective cost (i.e., fitness) of the feature combination is minimized. However, standard GAs randomly converge and fail to effectively control the generalizability of the feature combination. On the other hand, the LASSO regression is characterized by variable screening and complexity adjustment (regularization, L1 norm) while fitting a generalized linear model. The L1 norm of LASSO regression constrains, adjusts, or shrinks the coefficient estimates towards zero, so LASSO regression can be used for feature selection. Besides, when the penalty factor for the L1 norm is larger, the feature corresponding to the nonzero coefficient can play a more important role in multivariate regression. As the lambda value increases, the number of nonzero coefficients gradually decreases. However, LASSO regression searches for which features are more important (controlled by the size of the lambda value) for the regression result within a specific combination. Therefore, the regression performance depends not only on the choice of lambda but also obviously on the combination to be searched. In general, conventional LASSO regression selects lambda based on the minimum mean square error of k-fold CV. However, if LASSO regression is used to screen for variables from a large feature space, this approach can still result in too many nonzero coefficients, causing overfitting. If it is possible to select variables from an appropriate feature space, the number of nonzero coefficients can be reduced. However, a search in a specific feature space may limit the results obtained (which potentially leads to a local optimal result). Therefore, the plan in this study is to screen for variables in two steps based on the principles of the GA: the first step is the feature subspace search, and the second step is LASSO regression within the identified feature subspace. The specific framework is shown in Figure 2.

Figure 2
www.frontiersin.org

Figure 2 Basic framework of the GALambda method.

First, feature preselection is performed based on the Fisher correlation coefficient and statistical U test to obtain candidate features (9, 10). In the U test, a p-value less than 0.05 is commonly considered a significant difference statistically. The Fisher correlation coefficient approach is to calculate the pairwise correlation coefficients of features, and a threshold of 0.8 is empirically chosen. When the correlation coefficient is greater than or equal to 0.8, the two features are strongly correlated, and the feature with a higher average correlation is removed (9).

Then, LASSO and its lambda configuration are combined in the GA as follows:

1. The gene expression of an individual is represented as a binary string, with a length equal to the number of candidate features. Here, ‘0’ indicates that the corresponding feature is not selected, while ‘1’ indicates that the corresponding feature is selected. In this context, ‘selected’ refers to being included in a feature subspace, rather than being in the final feature result. LASSO regression is then applied to this feature subspace to obtain the final feature results.

2. After performing LASSO regression with k-fold CV on the feature subspace, the features with a lambda value less than the minimum mean standard error and nonzero coefficients are the finally selected features. The fitness function is located as the 1-area under the curve (AUC) value of the k-fold CV after performing a linear logistic regression on the final selected features.

3. Crossover operation. The crossover operation is applied to the genes of two individuals to generate new individuals. The main idea of the GALambda method for crossover is as follows. When searching for feature combinations in the feature subspace specified by an individual, the features corresponding to nonzero coefficients that are still retained under the penalty parameter lambda can be retained, while the remaining genes are randomly selected from their parents. This process enables the excellent genes that are retained even under high penalty parameters in LASSO regression to be directly inherited by the next generation as features in the feature subspace specified for the next generation, allowing feature combinations in the next generation to still include these features in the search. This also ensures diversity in the feature subspace, maximally mitigating the issue of the feature search becoming too localized in a single feature subspace. The specific pseudocode for the crossover operation is shown in Algorithm 1.

www.frontiersin.org

Algorithm 1. Pseudocode for the crossover operation in GALambda.

In Algorithm 1, lines 6-13 are the mechanism for dynamically setting excellent genes; lines 14-17 set excellent genes for new individuals; and lines 18-19 randomly assign other genes from parental individuals to new individuals.

Algorithm 2 is the pseudocode that first screen for excellent genes of the parental individuals in the crossover process, that is, the mechanism that dynamically sets the excellent genes according to the lambda parameter and fitness in the LASSO regression of the individual to be crossed over. The search starts from the maximum lambda value and stops when the number of nonzero coefficients meets the value specified by the input parameter. The input parameter n is determined by the predefined fitness interval in the crossover (refer to lines 2-5 of Algorithm 1).

www.frontiersin.org

Algorithm 2. Pseudocode for screening for excellent genes based on high penalty parameters.

For a given lambda value, the minimization problem to be solved by LASSO regression is defined in Equation 1, where N is the number of training samples, i and j are the sample numbers, and x and y represent the sample (feature vector) and response (i.e., the outcome variable, which in this study is either 1 for the presence of metastasis or 0 for the absence of metastasis), respectively, λ is the regularization parameter, that is, the penalty parameter lambda, which is nonnegative, β is the coefficient vector corresponding to the feature, and b is the intercept scalar. As can be observed from the last term of Equation 1, the lambda parameter is actually the weight of the L1 norm/regularization as a penalty factor/parameter. In regression, the contribution of a feature to the total score is noted with a nonzero coefficient value, but a larger penalty factor reduces the values of less important coefficients, possibly even down to 0, thus removing the feature corresponding to the coefficient from the regression.

min(i=1N(yibxiTβ)22N+λj=1M|βj|)(1)

In addition, considering that the GALambda method searches for linear combinations of features from the feature subspace, if the dimension of the feature subspace in the initial population is too low, GALambda may search in the sparse feature subspace. However, the method itself has the capability to sparsify the feature subspace through lambda tuning. Therefore, the GALambda method generates an initial population based on the univariate regression results. Specifically, the value of the regression evaluation metric R-squared in univariate regression is converted to a probabilistic expression—the gene being 1 (greater than or equal to 0.5) or 0 (less than 0.5). R-squared, as defined in Equation 2, is also known as coefficient of determination and is an important metric for assessing how well the coefficients fit the true value of y. In Equation 2, y represents the true value, the symbol ˆ denotes the predicted value, the symbol ˉ denotes the average value, and N is the number of samples to be fitted/trained.

R2=1i=1N(y^yi)2i=1N(yiy¯)2(2)

3 Experimental results and discussion

The features selected by GALambda are shown in Table 1. A one-level wavelet decomposition can produce an approximate component that represents low-frequency information and three detail components that represent horizontal, vertical, and diagonal high-frequency information, respectively. The approximate component can be decomposed again. A one-level contourlet transform consists of a pyramid decomposition and a directional filter decomposition, which can produce an approximate component that represents low-frequency information and 2n (decided by the directional filter bank) detail components. The approximate component can be decomposed again. Specifically, (1) the wavelet transform is carried out with three-level decomposition; the method names have the prefix CMS-W; feature extraction is performed on nine high-frequency components, where components 1, 2, and 3 are obtained from the first-level decomposition, components 4, 5, and 6 are obtained from the second-level decomposition, and components 7, 8, and 9 are obtained from the third-level decomposition; and (2) the contourlet transform is performed with the three-level decomposition; the method names have the prefix CMS-C; eight high-frequency components are obtained from the first-level decomposition, and four components are each obtained from the second-level and third-level decomposition. The components of contourlet decompositions are similarly numbered as the wavelet decompositions.

Table 1
www.frontiersin.org

Table 1 Features selected by GALambda.

GALambda first performed feature preselection based on thresholds of Fisher’s correlation coefficient and the Mann–Whitney U test. Then, GALambda performed the genetic algorithm as described in section 1.2.2 on the preselected features. As for an individual, LASSO regression was performed on the feature space corresponding to this individual. As described in Algorithms 1 and 2, high penalty factors were set to screen excellent genes and then a customized crossover operation was performed. After population iteration, 16 features were finally selected. Logistic linear regression was performed based on the selected features, and predictions were made on the independent test set. The AUC and the accuracy corresponding to the maximum Youden index were calculated. The results are shown in Table 2 (the values in parentheses represent the 95% confidence interval based on the sampling strategy with replacement—100 calculations), and the corresponding ROC curves are shown in Figure 3. Considering that previous relevant studies (1113) have used LASSO to select features, and most radiomics studies also usually applied LASSO for feature selection (9, 10), a conventional LASSO method was also implemented in the experiments of this study. Besides, GALambda searches for linear combinations of features in random subspaces delineated by a genetic algorithm. For comparison, the standard genetic algorithm is also implemented to select features. The selected features by conventional LASSO and standard GA are shown in the Supplementary Materials.

Table 2
www.frontiersin.org

Table 2 AUCs and accuracies for CV and the test results.

Figure 3
www.frontiersin.org

Figure 3 Classification ROC curves for the linear models constructed using feature combinations obtained by three methods. (A) ROC curves of conventional LASSO-based model. (B) ROC curves of standard GA-based model. (C) ROC curves of GALambda-based model. (D) Compare test ROC curves.

In the results of Table 2 and Figure 3, all three methods are based on the search for linear combinations of features based on the preselected candidate features. In the medical literature, 1/10 or 1/5 of the training sample size is widely used as the lower limit of feature number for developing prediction models that predict a binary outcome, which also generally been accepted as a methodological quality item in appraising published prediction modeling studies (14). To further reduce overfitting, the maximum number of nonzero coefficients was limited in LASSO regression. This value was empirically set to 1/10 or 1/5 of the training sample size. The conventional method optimizes the parameters for feature preselection during the experiment, that is, a grid search was conducted for the thresholds of the statistical significance tests and Fisher’s correlation coefficient. A p-value less than 0.05 in the U test is commonly considered a significant difference statistically, and Fisher’s correlation coefficient is greater than or equal to 0.8 is commonly considered that the corresopnidng two features are strongly correlated (9, 10). Therefore, we developed a small threshold grid centered at 0.05 and 0.8 for the U test and Fisher’s correlation coefficient respectively, and used the cross-validation AUC as the evaluation criterion to select the optimal thresholds. Table 2 shows that the Galambda-based model achieves higher test AUC and accuracy than the models of the other two methods. Correspondingly, as shown in Figure 3, the test ROC curve of the GALambda-based method approaches the upper left corner of the figure faster than the other two curves.

Although the standard GA did identify a linear combination of features with the highest CV AUC, its robustness is poor because the standard GA lacks an effective regularization parameter to control overfitting. The linear combination of features identified by the conventional method exhibited lower CV and test AUCs than the GALambda method. The same comparison results were also performed in terms of accuracy. This may be attributed to the fact that it only conducts a global search within a large feature subspace. In contrast, GALambda, because of its combination with the GA, initially searches for feature subspaces and can obtain a good feature subspace before performing LASSO regression, resulting in a more focused and richer search for linear combinations of feature and thereby ultimately leading to the discovery of a linear combination of features that can better predict lymph node metastasis. In addition, confidence intervals account for the variabilities in results. As for CV AUCs of standard GA and GALambda, the upper bounds of 95% confidence intervals are almost close to each other (0.903 vs. 0.896). As for CV accuracies of standard GA and GALambda, the lower bounds of 95% confidence intervals are almost close to each other (0.688 vs. 0.684). It may indicate that the model of GLambda would be no worse than the model of standard GA in the training.

Figure 4 shows the population iteration process of the GALambda algorithm, which gradually converges after approximately 75 iterations. The convergence is relatively fast and the complexity of the GA is not significantly increased. To further evaluate the clinical utility of the linear combination of features optimized by the GALambda algorithm in the construction of a linear model to predict lymph node metastasis, decision curves were used to assess the utility of the constructed linear model for guiding clinical intervention. For predictive models related to clinical problems, decision curves are also a crucial metric for assessing the model robustness. The decision curves in Figure 5 show that the linear model constructed based on the feature combination obtained by GALambda for the prediction of lymph node metastasis to guide the interventions can achieve a higher net benefit than those of the other two methods over a broader range of risk thresholds, suggesting that the linear model constructed by the feature combination obtained using the GALambda method has greater robustness and clinical utility. Besides, prediction models such as linear combinations such as regression are generally considered effective means of reducing overfitting. However, this has high criteria for the number of training samples and variables. Generally, the smaller the number of variables relative to the number of training samples, the lower the potential overfitting of a model (14). GALambda screens features in random subspaces divided by a genetic algorithm, thus screening out fewer features, which helps reduce overfitting. The experimental results also showed that the GALambda-based model performs the best on the independent test.

Figure 4
www.frontiersin.org

Figure 4 The population iterative process using GALambda.

Figure 5
www.frontiersin.org

Figure 5 Decision curves of linear models constructed by feature combinations obtained using three methods.

However, this study also has some limitations. First, it is a single-center study, and an external validation set usually preferably reflects the robustness and generalization ability of models, which can further verify the superiority of GALambda. Then, GALambda is a method based on random search. The final result is related to an initial state. Initialization population can affect the searching in subspaces and the subsequent evolution. Future studies can try various initialization methods to optimize GALambda.

4 Conclusion

In this study, to address the problems of small sample size and high-dimensional feature space of datasets to predict ISLN metastasis for breast cancer patients using ultrasound radiomics, we explored a GA, referred to as GALambda. GALambda performs a crossover operation dynamically set by LASSO regression based on a high penalty factor to optimize the search for feature combinations and construct linear prediction models with high interpretability. The experimental results showed that the GALambda method exploits the advantages of the GA by first searching for a feature subspace and then performing regularized LASSO regression to identify feature combinations within this feature subspace. In addition, based on the minimum mean square error through CV in LASSO regression, the GALambda method dynamically selects the features corresponding to nonzero coefficients under a high penalty factor for individual crossover in the GA, obtaining better feature combinations than those obtained using the conventional LASSO method and the standard GA after a finite number of population iterations. The evaluation of the AUCs of the ROC and decision curves showed that the linear prediction model constructed based on the feature combination obtained by the GALambda method could obtain better prediction results, which is helpful for the construction of prediction models of ISLN metastasis for breast cancer patients using ultrasound radiomics. In addition, this method also has value in quantitative data mining in other radiomics studies.

Data availability statement

The data analyzed in this study is subject to the following licenses/restrictions: Currently, researchers can obtain a fully anonymized dataset by emailing the corresponding authors if the researchers declare that the data are only used for research purposes.. Requests to access these datasets should be directed to Haohan Zhang, zhanghaohan@stu.scu.edu.cn.

Author contributions

HZ: Writing – original draft. JY: Writing – original draft. CZ: Writing – review & editing. JQ: Writing – review & editing. JW: Writing – review & editing. QL: Writing – review & editing. TL: Writing – review & editing.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported by the Science and Technology Department of Sichuan Providence under Grant no. 2022ZDZX0023.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2024.1349315/full#supplementary-material

References

1. Sung H, Ferlay J, Siegel RL, Laversanne M, Soerjomataram I, Jemal A, et al. Global cancer statistics 2020: globocan estimates of incidence and mortality worldwide for 36 cancers in 185 countries. CA: A Cancer J Clin (2021) 71(3):209–49. doi: 10.3322/caac.21660

CrossRef Full Text | Google Scholar

2. Xia C, Dong X, Li H, Cao M, Sun D, He S, et al. Cancer statistics in China and United States, 2022: profiles, trends, and determinants. Chin Med J (2022) 135(5):584–90. doi: 10.1097/CM9.0000000000002108

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Zhu J, Jiao D, Guo X, Qiao J, Ma Y, Zhang J, et al. Predictive factors and prognostic value of pathologic complete response of ipsilateral supraclavicular lymph nodes in breast cancer after neoadjuvant chemotherapy. Ann Trans Med (2019) 7(22):666. doi: 10.21037/atm.2019.10.22

CrossRef Full Text | Google Scholar

4. Aerts HJWL, Velazquez ER, Leijenaar RTH, Parmar C, Grossmann P, Carvalho S, et al. Decoding tumour phenotype by noninvasive imaging using a quantitative radiomics approach. Nat Commun (2014) 5:4006. doi: 10.1038/ncomms5006

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Gillies RJ, Kinahan PE, Hricak H. Radiomics: images are more than pictures, they are data. Radiology (2016) 278(2):563–77. doi: 10.1148/radiol.2015151169

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Nilupai A, Wang JY, Li Q, Zhang P, Ma F, Yuan P, et al. Analysis of the effect of ipsilateral supraclavicular lymph node metastasis on the prognosis of N3 breast cancer. Chin J Oncol (2021) 43(10):1069–75. doi: 10.3760/cma.j.cn112152-20191227-00845

CrossRef Full Text | Google Scholar

7. Bi WL, Hosny A, Schabath MB, Giger ML, Birkbak NJ, Mehrtash A, et al. Artificial intelligence in cancer imaging: Clinical challenges and applications. CA: A Cancer J Clin (2019) 69(2):127–57. doi: 10.3322/caac.21552

CrossRef Full Text | Google Scholar

8. Cho J, Lee K, Shin E, Choy G, Do S. How much data is needed to train a medical image deep learning system to achieve necessary high accuracy? arXiv (2015). doi: 10.48550/arXiv.1511.06348

CrossRef Full Text | Google Scholar

9. Liu Z, Wang S, Dong D, Wei J, Fang C, Zhou X, et al. The applications of radiomics in precision diagnosis and treatment of oncology: opportunities and challenges. Theranostics (2019) 9(5):1303–22. doi: 10.7150/thno.30309

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Afshar P, Mohammadi A, Plataniotis KN, Oikonomou A, Benali H. From handcrafted to deep-learning-based cancer radiomics: challenges and opportunities. IEEE Signal Process Magazine (2019) 36(4):132–60. doi: 10.1109/MSP.2019.2900993

CrossRef Full Text | Google Scholar

11. Liu J, Sun D, Chen L, Fang Z, Song W, Guo D, et al. Radiomics analysis of dynamic contrast-enhanced magnetic resonance imaging for the prediction of sentinel lymph node metastasis in breast cancer. Front Oncol (2019) 9:980. doi: 10.3389/fonc.2019.00980

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Qiu X, Jiang Y, Zhao Q, Yan C, Huang M, Jiang T. Could ultrasound-based radiomics noninvasively predict axillary lymph node metastasis in breast cancer? J Med Ultrasound (2020) 39(10):1897–905. doi: 10.1002/jum.15294

CrossRef Full Text | Google Scholar

13. Yu Y, Tan Y, Xie C, Hu Q, Ouyang J, Chen Y, et al. Development and validation of a preoperative magnetic resonance imaging radiomics-based signature to predict axillary lymph node metastasis and disease-free survival in patients with early-stage breast cancer. JAMA Network Open (2020) 3(12):e2028086. doi: 10.1001/jamanetworkopen.2020.28086

PubMed Abstract | CrossRef Full Text | Google Scholar

14. van Smeden M, Moons KG, de Groot JA, Collins GS, Altman DG, Eijkemans MJ, et al. Sample size for binary logistic prediction models: Beyond events per variable criteria. Stat Methods Med Res (2019) 28(8):2455–74. doi: 10.1177/0962280218784726

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Tibshirani R. Regression shrinkage and selection via the lasso. J R Stat Soc Ser B: Stat Method (1996) 58(1):267–88. doi: 10.1111/j.2517-6161.1996.tb02080.x

CrossRef Full Text | Google Scholar

16. Zou H, Hastie T. Regularization and variable selection via the elastic net. J R Stat Soc Ser B: Stat Method (2005) 67(2):301–20. doi: 10.1111/j.1467-9868.2005.00503.x

CrossRef Full Text | Google Scholar

17. Remeseiro B, Bolon-Canedo V. A review of feature selection methods in medical applications. Comput Biol Med (2019) 112:103375. doi: 10.1016/j.compbiomed.2019.103375

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Qiu J-J, Yin J, Ji L, Lu CY, Li K, Zhang YG, et al. Differential diagnosis of hepatocellular carcinoma and hepatic hemangioma based on maximum wavelet-coefficient statistics: Novel radiomics features from plain CT. Inf Process Manage (2022) 59(5):103046. doi: 10.1016/j.ipm.2022.103046

CrossRef Full Text | Google Scholar

19. Qiu J-J, Yin J, Qian W, Liu JH, Huang ZX, Yu HP, et al. A novel multiresolution-statistical texture analysis architecture radiomics-aided diagnosis of pdac based on plain CT images. IEEE Trans Med Imaging (2021) 40(1):12–25. doi: 10.1109/TMI.2020.3021254

PubMed Abstract | CrossRef Full Text | Google Scholar

20. van Griethuysen JJM, Fedorov A, Parmar C, Hosny A, Aucoin N, Narayan V, et al. Computational radiomics system to decode the radiographic phenotype. Cancer Res (2017) 77(21):e104–7. doi: 10.1158/0008-5472.CAN-17-0339

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: breast cancer, ipsilateral supraclavicular lymph node metastasis, radiomics, genetic algorithm, LASSO regression, feature combination

Citation: Zhang H, Yin J, Zhou C, Qiu J, Wang J, Lv Q and Luo T (2024) Identification of ipsilateral supraclavicular lymph node metastasis in breast cancer based on LASSO regression with a high penalty factor. Front. Oncol. 14:1349315. doi: 10.3389/fonc.2024.1349315

Received: 04 December 2023; Accepted: 19 January 2024;
Published: 02 February 2024.

Edited by:

Xiaowei Qi, Army Medical University, China

Reviewed by:

Nan Wang, First Affiliated Hospital of Zhengzhou University, China
Jia Wang, Second Affiliated Hospital of Dalian Medical University, China

Copyright © 2024 Zhang, Yin, Zhou, Qiu, Wang, Lv and Luo. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Qing Lv, lvqingwestchina@163.com

These authors have contributed equally to this work and share first authorship

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.