Skip to main content

ORIGINAL RESEARCH article

Front. Artif. Intell., 20 December 2024
Sec. Medicine and Public Health
This article is part of the Research Topic AI in Digital Oncology: Imaging and Wearable Technology for Cancer Detection and Management View all 6 articles

Prediction of PD-L1 tumor positive score in lung squamous cell carcinoma with H&E staining images and deep learning

Qiushi Wang&#x;Qiushi Wang1Xixiang Deng&#x;Xixiang Deng2Pan Huang&#x;Pan Huang2Qiang MaQiang Ma1Lianhua ZhaoLianhua Zhao1Yangyang FengYangyang Feng1Yiying WangYiying Wang1Yuan ZhaoYuan Zhao1Yan ChenYan Chen1Peng ZhongPeng Zhong1Peng HePeng He2Mingrui MaMingrui Ma3Peng Feng
Peng Feng2*Hualiang Xiao
Hualiang Xiao1*
  • 1Department of Pathology, Daping Hospital, Army Medical University, Chongqing, China
  • 2The Key Lab of Optoelectronic Technology and Systems, Ministry of Education, Chongqing University, Chongqing, China
  • 3Department of Information, Affiliated Tumor Hospital of Xinjiang Medical University, Urumchi, China

Background: Detecting programmed death ligand 1 (PD-L1) expression based on immunohistochemical (IHC) staining is an important guide for the treatment of lung cancer with immune checkpoint inhibitors. However, this method has problems such as high staining costs, tumor heterogeneity, and subjective differences among pathologists. Therefore, the application of deep learning models to segment and quantitatively predict PD-L1 expression in digital sections of Hematoxylin and eosin (H&E) stained lung squamous cell carcinoma is of great significance.

Methods: We constructed a dataset comprising H&E-stained digital sections of lung squamous cell carcinoma and used a Transformer Unet (TransUnet) deep learning network with an encoder-decoder design to segment PD-L1 negative and positive regions and quantitatively predict the tumor cell positive score (TPS).

Results: The results showed that the dice similarity coefficient (DSC) and intersection overunion (IoU) of deep learning for PD-L1 expression segmentation of H&E-stained digital slides of lung squamous cell carcinoma were 80 and 72%, respectively, which were better than the other seven cutting-edge segmentation models. The root mean square error (RMSE) of quantitative prediction TPS was 26.8, and the intra-group correlation coefficients with the gold standard was 0.92 (95% CI: 0.90–0.93), which was better than the consistency between the results of five pathologists and the gold standard.

Conclusion: The deep learning model is capable of segmenting and quantitatively predicting PD-L1 expression in H&E-stained digital sections of lung squamous cell carcinoma, which has significant implications for the application and guidance of immune checkpoint inhibitor treatments. And the link to the code is https://github.com/Baron-Huang/PD-L1-prediction-via-HE-image.

1 Background

Lung cancer is a malignant tumor with significant morbidity and mortality rates. Whereas non-small-cell lung cancer (NSCLC) comprises 80% of all lung cancers, the main types include adenocarcinoma (32–40%), squamous (25–30%), and large cell (8–16%) (Zarogoulidis et al., 2013). Immune checkpoint inhibitors (ICIs) have shown remarkable efficacy in the clinical treatment of NSCLC. Expression of programmed death ligand 1 (PD-L1) is significant for ICI efficacy. Therefore, immunohistochemical staining (IHC) for PD-L1 expression has been approved as a companion diagnostic marker in the clinical treatment of ICIs. However, there are disadvantages to using IHC to detect PD-L1 expression, including high detection and time costs, inconsistent interpretation standards, and the need for strong professional knowledge (Shamai et al., 2022). Furthermore, the interpretation criteria for PD-L1 vary significantly among tumor types. Meanwhile, factors including tumor heterogeneity, the complexity of the immune microenvironment, and the atypical expression of tumor cells presented by IHC staining easily cause subjective errors in the interpretation results of pathologists (Wang J. et al., 2019; Wang S. et al., 2019).

Hematoxylin–eosin (H&E), the most developed staining method in clinical pathology, enables the distinct visualization of cell morphology and tissue structure and is economical and easily operable. With the advancement of technology, digital high-resolution whole slide images (WSI) obtained from H&E-stained slides are used to provide a new direction for artificial intelligence (AI) assisted diagnosis using deep learning methods (Litjens et al., 2016). In recent years, researchers have found that the application of deep learning to pathological H&E-stained images can complete tasks that can be recognized by the human eye, such as tumor classification and tumor grading (Hu et al., 2021; Graham et al., 2019a,b). Moreover, it has a good predictive effect on higher-order tasks, such as gene mutation and survival period prediction (Zeng et al., 2019; Graham et al., 2019a,b) This method can effectively alleviate the problems of traditional PD-L1 detection technology, including its high cost, low efficiency, and subjective interpretation differences.

Wu et al. proposed a high-precision AI system to automatically evaluate the tumor positive cell score (TPS) for NSCLC PD-L1 expression (22C3 and SP263), and the calculated results showed a high degree of agreement between the AI and the pathologist (Wu et al., 2022). Mayer et al. proposed a convolutional neural network (CNN) classification model. Fusions of anaplastic lymphoma kinase (ALK), and ROS proto-oncogene 1 receptor tyrosine kinase (ROS1), were predicted directly from the H&E-stained WSI of postoperative tissues of patients with NSCLC. The sensitivity and specificity of the classifier for ALK and ROS1 were 100 and 98.6%, respectively (Mayer et al., 2022). Shamai et al. developed a WSI classification dataset for H&E staining of breast cancer and predicted the PD-L1 state of H&E-stained images using deep learning technology. The area under the curve (AUC) value of the predicted result was 0.91–0.93 2. (Sha et al., 2019) used deep learning to predict the PD-L1 status of H&E-stained WSIs in NSCLC tissues, with an AUC value of 0.8.

Considering the tissue structural complexity and heterogeneity of lung adenocarcinoma, lung squamous cell carcinoma was chosen as the object in this study. We constructed a WSI dataset using H&E-stained lung squamous cell carcinoma. To further quantitatively predict TPS, a deep learning model was employed for the first time to segment the PD-L1 expression region in H&E-stained WSIs of formalin-fixed and paraffin-embedded (FFPE) lung squamous cell carcinoma samples. The deep learning framework adopted the recently released deep segmentation network TransUnet (Chen et al., 2021), which introduced advanced transformer technology based on the classic Unet model to enhance the model’s ability to acquire contextual structural information in pathological images. This solved the problem wherein traditional convolutional networks exclusively focus on local knowledge, leading to segmented regions with limited structural similarity. Based on H&E-stained WSIs, the model can assist pathologists in predicting PD-L1 TPS and achieving end-to-end prediction.

2 Methods

2.1 Datasets

We enrolled surgical excision samples of lung squamous cell carcinoma. From January 2018 to December 2021, FFPE tissue samples from lung squamous cell carcinoma with PD-L1 test results from the Department of Pathology, Dapping Hospital, Army Military Medical University were enrolled in the study. Finally, 2,496 H&E-stained digital images were obtained after being labeled by senior pathologists, and each image size was 959 × 461 pixels. The image set was randomly divided as follows: a training set comprising 1,497 images was designated as the training data set for the deep learning model; a validation set comprising 499 images was used to assess the performance of PD-L1 state prediction and determine the optimal structure for the deep learning model; and a test set comprising 500 images was employed to evaluate the deep learning model’s segmentation and TPS prediction capabilities compared to the TPS as assessed by the pathologist.

2.2 Slice staining, scanning, and dataset labeling

Formalin fixed paraffin-embedded (FFPE) tissue samples of squamous cell carcinoma with known PD-L1 positive expression were successively sliced as 4 μm thick by technicians for H&E staining and PD-L1 IHC detection. H&E staining was performed using a fully automatic dyeing workstation Autostainer XL CV5030 (LEICA, Wetzlar, Germany). PD-L1 IHC staining was performed on the Dako Autostainer Link 48 platform (Dako, Copenhagen, Denmark) using a companion diagnostic kit (Antibody clone No. 22C3). Both H&E- and PD-L1-stained slices were generated digitally on a fully automated digital slice scanning system PRECISE 500B (UNIC-TECH, Beijing, China). The same areas on the H&E-stained digital images were synchronously labeled by trained pathologists using Adobe Photoshop CS6 software (Version 13.0), referring to the positive and negative tumor areas of PD-L1 expression on the IHC digital images. The labeled images were further confirmed by two senior pathologists as gold-standard.

The region of interest (ROI) labeled by pathologists for IHC and H&E images and the mask generated by the algorithm are illustrated in Figure 1A. The area marked by the green line in the ROI map and the gray area of the mask generated by the algorithm are tumor regions with negative PD-L1 expression, whereas the area marked by the red line and the white area of the mask are tumor regions with positive PD-L1 expression.

Figure 1
www.frontiersin.org

Figure 1. Annotated dataset generation and tumor segmentation framework construction. (A) Annotated pathological dataset of lung squamous carcinoma (the region of interest (ROI) labeled by pathologists based on IHC images, and synchronous generate ROI on H&E images and masks by the algorithm). Green line and the gray area: PD-L1 negative, red line and the white area: PD-L1 positive. (B) The entire framework of TransUnet. (C) Structure of Unet. (D) Structure of Transformer.

2.3 Deep learning model development

2.3.1 TransUnet framework

Focusing on the two problems of WSI resolution being too high for direct downsampling and patch-level classification lacking detailed representation, we proposed a more comprehensive solution for completing the recognition and segmentation of PD-L1 status in pathological images. The comprehensive detection framework is illustrated in Figure 1B, where black indicates background, normal tissue, white indicates PD-L1-positive areas of the tumor and gray indicates PD-L1-negative areas of the tumor.

(1) During the data collection process, H&E and PD-L1 IHC staining were performed on successive sections of FFPE tumor tissues. WSI images were generated using an automatic section scanner. According to the PD-L1 IHC digital images, the pathologists synchronically masked the positive and negative PD-L1 expression regions in the H&E images according to IHC images to obtain labeled WSI images. The data were further processed to satisfy the model. First, we used the flood fill algorithm to fill the ROI region in the WSI with the same pixel and then generated the mask label required by the training model. Using the sliding window algorithm, the original WSI images were divided into patches with no overlapping regions, and the size of the PD-L1 expression region in each patch was evaluated. Patches with no and few targets were discarded to avoid introducing many irrelevant targets, ensuring that the remaining patches had both background and target to better guide model learning.

(2) During model training, the segmented patch was first enhanced with data such as translation, rotation, and flip in random proportions. Subsequently, the depth segmentation network model TransUnet was trained based on the enhanced patch, and the combined loss of dice (Sudre et al., 2017) and Cross Entropy (CE) was adopted as the loss function. The former was responsible for measuring the overlap differences between the prediction results and label samples, while the latter calculated the pixel-level differences between the prediction results and label values. Combining the two losses can correctly guide model training and improve the segmentation model’s convergence speed. The model output was a pixel-by-pixel classification segmentation map, and more detailed WSI segmentation results were obtained.

(3) During model inference, the model input was the original patch images segmented after H&E staining, and the output was the segmentation results combined with the original patches. The segmentation results included two types of targets: the PD-L1 positive region and the PD-L1 negative region. The TPS of a patch can be quantitatively analyzed by calculating the pixel ratio of the positive area to the overall tumor area, and the TPS score of the entire WSI image can be obtained by combining the results of each patch.

We designed the framework by adopting the TransUnet network model with an encoder-decoder structure to predict and segment the negative/positive expression region of PD-L1, which was built based on Unet (Zunair and Hamza, 2021). The encoder included the following: The backbone network adopted the ResNet-V2 (He et al., 2016a,b) model. Compared with the traditional ResNet (He et al., 2016a,b) model, a pre-activation design was added to improve the performance of the residual module. Furthermore, transformer module was introduced after ResNet to enhance the modeling of context information and enrich the expression of long-range dependent information. The decoder included the following: The low-resolution feature map of the upper layer was continuously amplified by four layers of upsampling because the sampling process produced the problem of spatial structure information loss, and the shallow detailed spatial texture representation was transmitted to the upper layer with the same resolution through the shallow three skip connections. The scaling ratios of the three-hop connection were 1/2, 1/4, and 1/8. After the skip connection, the number of channels was doubled. In this case, 3 × 3 convolution refined information and compressed the number of channels to reduce the complexity of the model and improve the convergence speed. Multiscale features were captured by multiple sampling between the encoder and decoder. Then the complementary characteristics of deep and shallow, strong and weak were also taken into account. After fusion and refining these features, outputs feature maps with rich discriminant information.

2.3.2 Unet structure

Unet (Zunair and Hamza, 2021) is a symmetric segmentation network that fuzes shallow structural features and deep semantic features with skip connections and effectively uses different features for segmentation prediction. The network mainly includes an encoder, decoder, and hop connection. As demonstrated in Figure 1C, the network consists of four downsampling and four upsampling modules. The downsampling module extracted sophisticated image features, reduced image resolution, and expanded information channels. The upsampling module recovered feature information and simultaneously fused the features of the previous layer to supplement the details lost in the compression process (Zhou et al., 2020).

2.3.3 Transformer structure

Transformer (Vaswani et al., 2017) was first introduced in 2017 for natural language processing and have been widely used in computer vision in recent years. It adopts the structure of an encoder and a decoder, and the module structure is indicated in Figure 1D. The transformer mainly comprises position coding, layer regularization, a multihead attention mechanism, and a feedforward neural network. This uniquely designed transformer architecture can effectively capture long-distance dependencies and extract global features of images. Convolutional neural networks have significant advantages in extracting underlying features and obtaining local information. Therefore, a reasonable combination of convolutional operation and transformer modules can effectively compensate for their respective defects and fully exploit their advantages (Huang et al., 2022; Huang et al., 2023).

2.4 Segmentation effect comparison test and evaluation parameters

We selected several deep segmentation models to participate in the construction of the PD-L1 expression detection framework, including Unet (Zunair and Hamza, 2021), AttUnet (Oktay et al., 2018), TransUnet (Chen et al., 2021), FCN (Shelhamer et al., 2017), DeepLabv3 (Chen et al., 2017), DeepLabV3+ (Chen et al., 2018), DenseASPP (Yang et al., 2018), SETR (Zheng et al., 2020), Segmenter (Strudel et al., 2021), OCRNet (Wang J. et al., 2019; Wang S. et al., 2019), Segformer (Xie et al., 2021), BiseNetV2 (Changqian et al., 2021), and DDRNet (Pan et al., 2023). The Unet series models employed an encoder-decoder structure and integrated the recovered feature map with additional abundant features from the front layer. This approach facilitated the edge structure’s refinement. The DeepLab series models were characterized by the introduction of atrous spatial pyramid pooling (ASPP), which can improve the representation of multiscale information. To assess the efficacy of each segmentation model and select an optimal model for the construction of the PD-L1 prediction framework. We calculated the dice similarity coefficient (DSC), intersection overunion (IoU), accuracy pixel (AP), and Hausdorff distance (HD) to evaluate each model comprehensively. The calculation principle of each parameter index is as follows:

Dice = 2 | p r e d mask mask | p r e d mask | + | mask     (1)
I o U = p r e d mask mask p r e d mask mask     (2)

In Formulas 1 and 2, p r e d mask represents the segmentation result predicted by the model, and the mask represents the true expression state of PD-L1 in the sample.

Pixel Accuracy P A = k = 0 n P k i = 0 m j = 0 n P i j     (3)

In Formula 3, P k represents the pixels predicted correctly by the model and P i j represents all pixels in the output image of the model.

d H X Y = max d X Y d Y X = max max x X min y Y d x y , max y Y min x X d x y     (4)

In Formula 4, d X Y and d Y X , respectively, represents the distance between x set external and y set inside, and the distance between y set external and x set inside.

2.5 TPS consistency analysis

Five pathologists from Daping Hospital blindly evaluated TPS using IHC slides, and AI predicted TPS based on H&E slides in the test set. The root mean square error (RMSE) between the TPS evaluated by the five pathologists and the TPS predicted by AI with the gold standard of labeling labels was calculated. In the consistency analysis, intra-group correlation coefficients (ICC) were calculated between TPS predicted by AI and the gold standard and TPS evaluated by five pathologists (3,1). A total of 453 effective samples remained after preprocessing the digital slides and excluding 39 images with PD-L1 positive and negative proportions below 5% and 14 samples without sufficient tumor cells.

3 Results

3.1 Model performance and comparison test

The performance of each model on the test set is presented in Table 1. DeepLabv3+ (ResNet101) achieves poor segmentation accuracy because it uses single upsampling to integrate low-level details into high-level information, resulting in insufficient use of details. However, the Unet network continually combines bottom feature maps of varying resolutions with top layers to maximize the utilization of local details, thereby improving prediction accuracy. FCN (ResNet101), Attention Unet, and DeepLabv3 improve the effective feature extractor, introduce an attention mechanism to refine features, and expand the global receptive field, respectively, to further improve the segmentation effect of the network while retaining detailed information. Nevertheless, a common disadvantage of these models is that they fail to enhance the significance of global features in pathological image segmentation. To achieve the best segmentation effect, the TransUnet model employed in this study not only implemented the effective fusion of multiscale features but also enhanced the significance of global discriminant features in prediction. The DSC coefficient of overlap between PD-L1 expression regions predicted by the TransUnet model and real labels was 80%, the intersection ratio IoU was 72%, the edge accuracy was 89, and the average pixel accuracy was 88%. Compared with other models, TransUnet has apparent advantages regarding object recognition integrity and structural similarity (Table 1).

Table 1
www.frontiersin.org

Table 1. Comparative experimental results of the different models.

Figure 2 displays the pixel accuracy of the PD-L1 positive and negative regions for each model. The pixel recognition accuracy of TransUnet for negative and positive expression regions was 85 and 75%, respectively, and its accuracy was better than other models in both single and average categories. Accordingly, the expression status of PD-L1 can be identified more accurately.

Figure 2
www.frontiersin.org

Figure 2. Pixel accuracy of positive/negative tumor segmentation category.

To visually compare the prediction accuracies of various models, we opted to visualize the model’s segmentation effect. The model’s prediction results are displayed in gray for PD-L1 negative regions and white for PD-L1 positive regions. The ROI represented the initial mark of the doctor, and a mask was generated by the algorithm to identify the ROI. The inaccurate segmentation of a large target area resulted from a detailed information deficiency in the prediction layer of DeepLabv3+ (ResNet101). The Attention Unet and FCN (ResNet101) greatly emphasized the extraction of detailed features, which weakened the ability to locate lesions and compromised the integrity of the segmentation target. DenseASPP and DeepLabv3+ (HRNet) exhibit improved segmentation accuracy compared with other models with less difference between the target and background (Figure 3). TransUnet emphasized not only global information but also detailed information acquisition, resulting in accurate segmentation results and more detailed segmentation of PD-L1 expression regions. Regarding the extent of segmentation detail reduction, TransUnet exhibited a distinct advantage over the other two Unet-based models (Figure 4).

Figure 3
www.frontiersin.org

Figure 3. The comparative experiment of model segmentation effect. Green area: PD- L1 negative, Red area: PD-L1 positive.

Figure 4
www.frontiersin.org

Figure 4. The comparative experiment of the segmentation details of three Unet models. Gray area: PD-L1 negative, White area: PD-L1 positive.

3.2 TPS comparison test

The accuracy of TPS quantitatively predicted by AI using H&E-stained slides was compared with the TPS assessed by five senior pathologists. Five pathologists performed TPS evaluation blindly on IHC slides of 453 samples in the test set and recorded the results. Additionally, the TPS predicted quantitatively by AI was approximated by the ratio of the PD-L1 positive area to the total tumor area in the TransUnet inference segmentation results.

Five pathologists evaluated TPS (TPS-PATH N, N = 1–5); five pathologists evaluated mean TPS (TPS-mean PATHs); and the TPS evaluated by AI (TPS-AI) was compared to the gold standard to assess the diagnostic accuracy of pathologists and AI in TPS. The results showed that TPS-AI had the best agreement with the gold standard (TPS-AI vs. gold standard: R = 0.92), followed by TPS-PATH3 (TPS-PATH3 vs. gold standard: R = 0.9), and TPS-mean PATHs (TPS-mean PATHs vs. gold standard: R = 0.85) (Figure 5A). Where black indicates background, normal tissue. White indicates PD-L1-positive areas of the tumor. Gray indicates PD-L1-negative areas of the tumor.

Figure 5
www.frontiersin.org

Figure 5. TPS comparison test results. (A) The consistency of TPS assessed by five pathologists (TPS-PATH N, N = 1–5), the mean TPS of five pathologists (TPS-mean PATHs), and AI-TPS with the gold standard TPS. (B) Root mean square error (RMSE) between TPS-PATH N (N = 1–5), TPS-mean PATHs, and TPS-AI with the gold standard TPS.

Additionally, we calculated the RMSE between TPS-PATH N (N = 1–5) and the gold standard, between TPS-mean PATHs and the gold standard, and between TPS-AI and the gold standard to reflect the accuracy of TPS evaluated by TransUnet through H&E slides. As presented in Figure 5B, AI had the smallest RMSE (RMSE = 19.67) with the gold standard, which was lower than the best-performing pathologist (PATH3, RMSE = 21.08) and the TPS-mean PATHs (RMSE = 29.44).

3.3 Consistency analysis

For the 453 samples in the test set, we calculated the ICC (3,1) between TPS-PATH N (N = 1–5), TPS-mean PATHs, and TPS-AI, respectively, with the gold standard. The results showed that AI had the best consistency with the gold standard, ICC (3,1) value was 0.92 (95% CI: 0.90–0.93) than PATH N or mean PATHs (Figure 6). Where black indicates background, normal tissue. White indicates PD-L1-positive areas of the tumor. Gray indicates PD-L1-negative areas of the tumor.

Figure 6
www.frontiersin.org

Figure 6. Intra-group correlation coefficient (ICC) between TPS-PATH N (N = 1–5), TPS-mean PATHs and TPS-AI, respectively, with the gold standard.

3.4 Generalization experiment

To verify the generalizability and reliability of the TransUnet model, we performed experiments on the laryngeal tumor pathology image dataset. TransUnet also demonstrates significant advantages in object recognition completeness and structural similarity (Table 2).

Table 2
www.frontiersin.org

Table 2. Generalization experimental results of the different models.

4 Discussion

Under high magnification, the TPS evaluation of PD-L1 IHC-stained slides in lung squamous cell carcinoma is based on the accurate interpretation of tumor cells with positive cell membrane expression. However, the interpretation is hindered by necrotic and immune cells (e.g., lymphocytes and macrophages) that express PD-L1. Furthermore, determining the proportion of tumor cells exhibiting positive expression under low magnification is a challenging task, and pathologists’ diagnostic efficacy could be significantly compromised by frequent magnification changes on the microscope (Coudray et al., 2018; Fraz et al., 2020). Therefore, accurately identifying the TPS of PD-L1 expression at global and detailed levels using deep learning models is of significant clinical application value. Results of this research demonstrated that when segmenting PD-L1 status in pathological images, the convolutional neural network-based segmentation model emphasizes the overall integrity of the target instead of accurate learning of its details; thus, the structural similarity of the segmentation results has disadvantages. Particularly, the accuracy of segmentation was reduced when differences between the object and the background were minimal. By combining the benefits of Unet and Transformer technology, the TransUnet model is capable of not only accurately identifying regions of PD-L1 expression that are positive or negative (with the highest pixel recognition accuracy) but also presenting segmentation details in the segmentation results to ensure the target’s structural similarity is maintained.

Considering the effects on image segmentation and recognition and the degree of restoration, some models were inaccurate in the segmentation of large areas because of the lack of attention to details. Nevertheless, certain models have laid an excessive emphasis on detailed feature extraction, resulting in compromised segmentation targets due to weakened lesion localization abilities. Compared with these models, TransUnet segmented and identified PD-L1 expression regions with more accuracy. Compared to the Unet and AttentionUnet models, segmentation and recognition at the edge of the target region are more accurate, and the reduction with mask labels is significant.

Nevertheless, the clinical application of IHC staining is challenging due to its complicated process, long duration, lack of standardization of interpretation standards, and most importantly, its heavy reliance on subjective evaluation by experienced pathologists. Accurate prediction of PD-L1 status using deep learning models to analyze tissue and cell morphology in H&E-stained sections will undoubtedly facilitate the clinical application of PD-L1 expression scores. Although the TransUnet model applied to PD-L1 TPS prediction in this sdudy still needs to be trained with large number of cases, and confirmed in large-scale independent cohort studies. The current results showed that the model integrating the advantages of Unet and Transformer technology can potentially predict PD-L1 expression using H&E-stained digital slides. More importantly, compared with the subjective assessment of pathologists, the prediction results of PD-L1 TPS status using the TransUnet model are always objective and repeatable. Furthermore, this model is not only suitable for the prediction of PD-L1 in lung squamous cell carcinoma, but can also be applied to other cancers and different biomarkers based on deep learning, thus providing a new path for the description of histopathological characteristics of biomarkers and clinical treatment target screening.

This study has certain limitations, and we hope to solve these problems during continuous training and clinical cohort studies. First, due to the insufficient number of images in the training group, the TransUnet model’s ability to accurately predict PD-L1 expression in lung squamous cell carcinoma remains inadequately representative. Second, deep learning and testing of PD-L1 expression in lung adenocarcinoma are lacking. Compared with lung squamous cell carcinoma, lung adenocarcinoma is more complex and heterogeneous, which undoubtedly poses particular challenges for the deep learning and prediction of this model. Nevertheless, the complexity and heterogeneity of lung adenocarcinoma tissue structure are challenging and subject to interpretation differences for pathologists under light microscopy. Therefore, a head-to-head comparison of PD-L1 TPS between this model and the pathologists in the real working environment is of great clinical significance, and it is also helpful for correcting the inconsistent results of the pathologist.

5 Conclusion

We proposed TransUnet, a framework for quantitative PD-L1 TPS prediction using H&E-stained digital images. The backbone of the framework is the TransUnet semantic segmentation model. TransUnet enhanced the modeling of global context information in the encoder stage and combined the representation of local features using a CNN. In the decoder stage, the feature map was amplified by multiple upsampling, and the features of the shallow and high layers were fuzed to acquire more discriminative features. The DSC coefficient and IoU of this model for PD-L1 segmentation prediction based on H&E images of lung squamous cell carcinoma were 80 and 72%, respectively, which ensured the integrity of the segmentation target and improved its structural similarity. Furthermore, when comparing the deep learning network’s quantitative prediction of TPS to that of five pathologists assessment, it exhibited a greater degree of concurrence with the gold standard [ICC (3,1) = 0.92, 95% CI: 0.90–0.93] and a reduced error (RMSE = 19.67). AI not only has good accuracy in predicting TPS but also can eliminate the evaluation errors caused by subjective differences between different pathologists. In conclusion, the deep learning network proposed in this study can effectively assist pathologists in completing PD-L1 TPS prediction using H&E-stained digital images.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by the Ethics Committee of Daping Hospital, Army Medical University. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation was not required from the participants or the participants’ legal guardians/next of kin in accordance with the national legislation and institutional requirements.

Author contributions

QW: Conceptualization, Data curation, Validation, Writing – original draft. XD: Conceptualization, Methodology, Software, Writing – original draft. PHu: Conceptualization, Data curation, Methodology, Software, Writing – original draft. QM: Data curation, Validation, Writing – review & editing. LZ: Data curation, Validation, Writing – review & editing. YF: Data curation, Validation, Writing – review & editing. YW: Data curation, Validation, Writing – review & editing. YZ: Data curation, Validation, Writing – review & editing. YC: Data curation, Writing – review & editing. PZ: Data curation, Writing – review & editing. PHe: Methodology, Writing – review & editing. MM: Methodology, Writing – review & editing. PF: Funding acquisition, Methodology, Project administration, Writing – review & editing. HX: Conceptualization, Funding acquisition, Project administration, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was supported by the Science and Technology Innovation Project of Army Medical University (H Xiao), the Graduate Scientific Research and Innovation Foundation of Chongqing (CYB23047, P Feng), the Chongqing Technology Innovation and Application Development Project (cstc2021jscx-gksbX0056, P Feng), the Fundamental Research Funds for the Central Universities (2023CDJKYJH085, P Feng), and Autonomous Region Science and Technology Support Xinjiang Project Plan (2021E02078, M Ma).

Acknowledgments

Thanks to Shirong Wei and Ping Fu for assisting with PD-L1 IHC and H&E staining.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

Changqian, Y., Changxin, G., Jingbo, W., Gang, Y., Chunhua, S., and Nong, S. (2021). BiSeNet V2: bilateral network with guided aggregation for real-time semantic segmentation. arXiv:2004.02147. doi: 10.1007/s11263-021-01515-2

Crossref Full Text | Google Scholar

Chen, J., Lu, Y., Yu, Q., Luo, X., Adeli, E., Wang, Y., et al. (2021). TransUNet: transformers make strong encoders for medical image segmentation. ArXiv. doi: 10.48550/arXiv.2102.04306

Crossref Full Text | Google Scholar

Chen, L. C., Papandreou, G., Schroff, F., and Adam, H. (2017). Rethinking atrous convolution for semantic image segmentation. ArXiv. doi: 10.48550/arXiv.1706.05587

Crossref Full Text | Google Scholar

Chen, L. C., Zhu, Y., Papandreou, G., Schroff, F., and Adam, H. (2018). Encoder-decoder with Atrous separable convolution for semantic image segmentation. Eur. Conf. Comput. Vis. doi: 10.1007/978-3-030-01234-2_49

Crossref Full Text | Google Scholar

Coudray, N., Ocampo, P. S., Sakellaropoulos, T., Narula, N., Snuderl, M., Fenyö, D., et al. (2018). Classification and mutation prediction from non–small cell lung cancer histopathology images using deep learning. Nat. Med. 24, 1559–1567. doi: 10.1038/s41591-018-0177-5

Crossref Full Text | Google Scholar

Fraz, M. M., Khurram, S. A., Graham, S., Shaban, M., Hassan, M., Loya, A., et al. (2020). FABnet: feature attention-based network for simultaneous segmentation of microvessels and nerves in routine histology images of oral cancer. Neural Comput. Applic. 32, 9915–9928. doi: 10.1007/s00521-019-04516-y

Crossref Full Text | Google Scholar

Graham, S., Chen, H., Gamper, J., Dou, Q., Heng, P., Snead, D., et al. (2019a). MILD-net: minimal information loss dilated network for gland instance segmentation in colon histology images. Med. Image Anal. 52, 199–211. doi: 10.1016/j.media.2018.12.001

Crossref Full Text | Google Scholar

Graham, S., Vu, Q. D., Raza, S. E. A., Azam, A., Tsang, Y. W., Kwak, J. T., et al. (2019b). Hover-net: simultaneous segmentation and classification of nuclei in multi-tissue histology images. Med. Image Anal. 58:101563. doi: 10.1016/j.media.2019.101563

Crossref Full Text | Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016a). Identity mappings in deep residual networks. Cham: Springer.

Google Scholar

He, K., Zhang, X., Ren, S., and Sun, J. (2016b). Deep residual learning for image recognition. Proc. IEEE Conf. Comput. Vis. Pattern Recognit. doi: 10.1109/CVPR.2016.90

Crossref Full Text | Google Scholar

Hu, J., Cui, C., Yang, W., Huang, L., Yu, R., Liu, S., et al. (2021). Using deep learning to predict anti-PD-1 response in melanoma and lung cancer patients from histopathology images. Transl. Oncol. 14:100921. doi: 10.1016/j.tranon.2020.100921

Crossref Full Text | Google Scholar

Huang, P., He, P., Tian, S., Ma, M., Feng, P., Xiao, H., et al. (2023). A ViT-AMC network with adaptive model fusion and multiobjective optimization for interpretable laryngeal tumor grading from histopathological images. IEEE Trans. Med. Imaging 42, 15–28. doi: 10.1109/TMI.2022.3202248

Crossref Full Text | Google Scholar

Huang, P., Tan, X., Zhou, X., Liu, S., Mercaldo, F., and Santone, A. (2022). FABNet: fusion attention block and transfer learning for laryngeal cancer tumor grading in P63 IHC histopathology images. IEEE J. Biomed. Health Inform. 26, 1696–1707. doi: 10.1109/JBHI.2021.3108999

Crossref Full Text | Google Scholar

Litjens, G., Sánchez, C. I., Timofeeva, N., Hermsen, M., Nagtegaal, I., Kovacs, I., et al. (2016). Deep learning as a tool for increased accuracy and efficiency of histopathological diagnosis. Sci. Rep. 6:26286. doi: 10.1038/srep26286

Crossref Full Text | Google Scholar

Mayer, C., Ofek, E., Fridrich, D. E., Molchanov, Y., Yacobi, R., Gazy, T., et al. (2022). Direct identification of ALK and ROS1 fusions in non-small cell lung cancer from hematoxylin and eosin-stained slides using deep learning algorithms. Mod. Pathol. 35, 1882–1887. doi: 10.1038/s41379-022-01141-4

Crossref Full Text | Google Scholar

Oktay, O., Schlemper, J., Folgoc, L. L., Lee, M. J., Heinrich, M., Misawa, K., et al. (2018). Attention U-net: learning where to look for the pancreas. ArXiv. doi: 10.48550/arXiv.1804.03999

Crossref Full Text | Google Scholar

Pan, H., Hong, Y., and Jia, S Y. (2023). Deep dual-resolution networks for real-time and accurate semantic segmentation of traffic scenes. IEEE Transactions on Intelligent Transportation Systems.

Google Scholar

Sha, L., Osinski, B. L., Ho, I. Y., Tan, T. L., Willis, C., Weiss, H., et al. (2019). Multi-field-of-view deep learning model predicts nonsmall cell lung Cancer programmed death-ligand 1 status from whole-slide hematoxylin and eosin images. J. Pathol. Inform. 10:24. doi: 10.4103/jpi.jpi_24_19

Crossref Full Text | Google Scholar

Shamai, G., Livne, A., Polónia, A., Sabo, E., Cretu, A., Bar-Sela, G., et al. (2022). Deep learning-based image analysis predicts PD-L1 status from H&E-stained histopathology images in breast cancer. Nat. Commun. 13:6753. doi: 10.1038/s41467-022-34275-9

Crossref Full Text | Google Scholar

Shelhamer, E., Long, J., and Darrell, T. (2017). Fully convolutional networks for semantic segmentation. IEEE Trans. Pattern Anal. Mach. Intell. 39, 640–651. doi: 10.1109/TPAMI.2016.2572683

Crossref Full Text | Google Scholar

Strudel, R., Garcia, R., Laptev, I., and Schmid, C. (2021). Segmenter: transformer for semantic segmentation. arXiv:2105.05633. doi: 10.48550/arXiv.2105.05633

Crossref Full Text | Google Scholar

Sudre, C. H., Li, W., Vercauteren, T., Ourselin, S., and Cardoso, M. J. (2017). Generalised dice overlap as a deep learning loss function for highly unbalanced segmentations. Deep Learn Med Image Anal Multimodal Learn Clin Decis Support. doi: 10.1007/978-3-319-67558-9_28

Crossref Full Text | Google Scholar

Vaswani, A., Shazeer, N., Parmar, N., and Uszkoreit, J. (2017). Attention is all you need. Proceedings of the 31st International Conference on Neural Information Processing Systems. Long Beach, CA, USA.

Google Scholar

Wang, J., Chen, X., Chen, X., and Yuan, Y. (2019). Segmentation transformer: object-contextual representations for semantic segmentation. arXiv:1909.11065. doi: 10.48550/arXiv.1909.11065

Crossref Full Text | Google Scholar

Wang, S., Zhu, Y., Yu, L., Chen, H., Lin, H., Wan, X., et al. (2019). RMDL: recalibrated multi-instance deep learning for whole slide gastric image classification. Med. Image Anal. 58:101549. doi: 10.1016/j.media.2019.101549

Crossref Full Text | Google Scholar

Wu, J., Liu, C., Liu, X., Sun, W., Li, L., Gao, N., et al. (2022). Artificial intelligence-assisted system for precision diagnosis of PD-L1 expression in non-small cell lung cancer. Mod. Pathol. 35, 403–411. doi: 10.1038/s41379-021-00904-9

Crossref Full Text | Google Scholar

Xie, E., Wang, W., Yu, Z., et al. (2021). SegFormer: simple and efficient design for semantic segmentation with transformers. arXiv.2105.15203. doi: 10.48550/arXiv.2105.15203

Crossref Full Text | Google Scholar

Yang, M., Yu, K., Zhang, C., Li, Z., and Yang, K. (2018). DenseASPP for semantic segmentation in street scenes. IEEE Conf. Comp. Vis. Patt. Recogn. doi: 10.1109/CVPR.2018.00388

Crossref Full Text | Google Scholar

Zarogoulidis, K., Zarogoulidis, P., Darwiche, K., Boutsikou, E., Machairiotis, N., Tsakiridis, K., et al. (2013). Treatment of non-small cell lung cancer (NSCLC). J. Thorac. Dis. doi: 10.3978/j.issn.2072-1439.2013.07.10

Crossref Full Text | Google Scholar

Zeng, Z., Xie, W., Zhang, Y., and Lu, Y. (2019). RIC-Unet: an improved neural network based on Unet for nuclei segmentation in histology images. IEEE Access 7, 21420–21428. doi: 10.1109/ACCESS.2019.2896920

Crossref Full Text | Google Scholar

Zheng, S., Lu, J., Zhao, H., Zhu, X., Luo, Z., Wang, Y., et al. (2020). Rethinking semantic segmentation from a sequence-to-sequence perspective with transformers. arXiv.2012.15840. doi: 10.48550/arXiv.2012.15840

Crossref Full Text | Google Scholar

Zhou, Z., Siddiquee, M. M. R., Tajbakhsh, N., and Liang, J. (2020). UNet++: redesigning skip connections to exploit multiscale features in image segmentation. IEEE Trans. Med. Imaging 39, 1856–1867. doi: 10.1109/TMI.2019.2959609

Crossref Full Text | Google Scholar

Zunair, H., and Hamza, A. B. (2021). Sharp U-net: depthwise convolutional network for biomedical image segmentation. Comput. Biol. Med. 136:104699. doi: 10.1016/j.compbiomed.2021.104699

Crossref Full Text | Google Scholar

Keywords: lung squamous cell carcinoma, PD-L1, deep learning, TransUnet, H&E staining

Citation: Wang Q, Deng X, Huang P, Ma Q, Zhao L, Feng Y, Wang Y, Zhao Y, Chen Y, Zhong P, He P, Ma M, Feng P and Xiao H (2024) Prediction of PD-L1 tumor positive score in lung squamous cell carcinoma with H&E staining images and deep learning. Front. Artif. Intell. 7:1452563. doi: 10.3389/frai.2024.1452563

Received: 21 June 2024; Accepted: 10 December 2024;
Published: 20 December 2024.

Edited by:

Souptik Barua, New York University, United States

Reviewed by:

Ting Li, National Center for Toxicological Research (FDA), United States
Zhixiang Ren, Peng Cheng Laboratory, China
Abhimanyu Banerjee, Illumina, United States

Copyright © 2024 Wang, Deng, Huang, Ma, Zhao, Feng, Wang, Zhao, Chen, Zhong, He, Ma, Feng and Xiao. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Peng Feng, Y29lLWZwQGNxdS5lZHUuY24=; Hualiang Xiao, ZHBibF94aGxAdG1tdS5lZHUuY24=

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.