Skip to main content

ORIGINAL RESEARCH article

Front. Physiol., 08 August 2024
Sec. Computational Physiology and Medicine

Automatic cardiothoracic ratio calculation based on lung fields abstracted from chest X-ray images without heart segmentation

Yingjian Yang&#x;Yingjian Yang1Jie Zheng&#x;Jie Zheng1Peng Guo&#x;Peng Guo1Tianqi WuTianqi Wu1Qi GaoQi Gao2Yingwei Guo
Yingwei Guo3*Ziran ChenZiran Chen4Chengcheng LiuChengcheng Liu5Zhanglei OuyangZhanglei Ouyang1Huai Chen
Huai Chen6*Yan Kang,,,
Yan Kang4,7,8,9*
  • 1Department of Radiological Research and Development, Shenzhen Lanmage Medical Technology Co., Ltd., Shenzhen, Guangdong, China
  • 2Neusoft Medical System Co., Ltd., Shenyang, Liaoning, China
  • 3School of Electrical and Information Engineering, Northeast Petroleum University, Daqing, China
  • 4College of Medicine and Biological Information Engineering, Northeastern University, Shenyang, China
  • 5School of Life and Health Management, Shenyang City University, Shenyang, China
  • 6Department of Radiology, The Second Affiliated Hospital of Guangzhou Medical University, Guangzhou China
  • 7College of Health Science and Environmental Engineering, Shenzhen Technology University, Shenzhen China
  • 8School of Applied Technology, Shenzhen University, Shenzhen, China
  • 9Engineering Research Centre of Medical Imaging and Intelligent Analysis, Ministry of Education, Shenyang, China

Introduction: The cardiothoracic ratio (CTR) based on postero-anterior chest X-rays (P-A CXR) images is one of the most commonly used cardiac measurement methods and an indicator for initially evaluating cardiac diseases. However, the hearts are not readily observable on P-A CXR images compared to the lung fields. Therefore, radiologists often manually determine the CTR’s right and left heart border points of the adjacent left and right lung fields to the heart based on P-A CXR images. Meanwhile, manual CTR measurement based on the P-A CXR image requires experienced radiologists and is time-consuming and laborious.

Methods: Based on the above, this article proposes a novel, fully automatic CTR calculation method based on lung fields abstracted from the P-A CXR images using convolutional neural networks (CNNs), overcoming the limitations to heart segmentation and avoiding errors in heart segmentation. First, the lung field mask images are abstracted from the P-A CXR images based on the pre-trained CNNs. Second, a novel localization method of the heart’s right and left border points is proposed based on the two-dimensional projection morphology of the lung field mask images using graphics.

Results: The results show that the mean distance errors at the x-axis direction of the CTR’s four key points in the test sets T1 (21 × 512 × 512 static P-A CXR images) and T2 (13 × 512 × 512 dynamic P-A CXR images) based on various pre-trained CNNs are 4.1161 and 3.2116 pixels, respectively. In addition, the mean CTR errors on the test sets T1 and T2 based on four proposed models are 0.0208 and 0.0180, respectively.

Discussion: Our proposed model achieves the equivalent performance of CTR calculation as the previous CardioNet model, overcomes heart segmentation, and takes less time. Therefore, our proposed method is practical and feasible and may become an effective tool for initially evaluating cardiac diseases.

1 Introduction

X-ray is the most widely used primary imaging technique for routine chest and bone radiography as it is widely available, low cost, has fast imaging speed, and easy to acquire (Liu et al., 2022). Because of its fast imaging (seconds after exposure), X-ray has become the preferred imaging device to improve the work efficiency and facilitate the diagnosis of critically ill and/or emergency patients in clinical practice (Howell, 2016; Seah et al., 2021).

The relationship analysis between the heart and lungs has been a hot topic in clinical or scientific research (Yang et al., 2022). The cardiothoracic ratio (CTR) is one of the most commonly used cardiac measurement methods and a commonly used indicator for evaluating cardiac enlargement (Saul, 1919; Hada, 1995; Chang et al., 2022). Specifically, various etiologies, such as pathological changes in the heart itself and increased adaptability secondary to hemodynamic changes leading to left and right heart enlargement, will increase the CTR. Chest X-rays (CXRs), as an economical and convenient routine examination, can better display the situation of the chest, lung tissue, pulmonary blood vessels, heart, chest blood vessels, etc., providing a reliable basis for clinical diagnosis (Howell, 2016; Seah et al., 2021; Liu et al., 2022). Although cardiac enlargement should be diagnosed through echocardiography, follow-up and treatment can be based on the postero-anterior (P-A) CXR images. P-A CXR must be performed for the initial cardiac examination (Hada, 1995). In addition, the CTR is also a predictor of heart failure progression in asymptomatic patients with cardiac diseases (Nochioka et al., 2023). Therefore, the accurate CTR measurement of these vulnerable populations is crucial for precision healthcare.

Manual CTR measurement based on the P-A CXR image requires experienced radiologists and is time-consuming and laborious. Therefore, with the rapid development of artificial intelligence, such as convolutional neural networks (CNNs), automatic CTR calculation methods or models based on P-A CXR images have been successively proposed in recent years (Li et al., 2019; Saiviroonporn et al., 2021; Jafar et al., 2022). Li et al. proposed a computer-aided technique that is more reliable and time- and labor-saving than the manual method in CTR calculation. In addition, Pairash et al. verified that the AI (artificial intelligence)-only method could replace the manual CTR measurement. Meanwhile, Abbas et al. proposed a CardioNet model that achieved acceptable accuracy and competitive results across all datasets. However, the above methods or models in references 9–11 are still limited by heart segmentation. Although heart segmentation techniques based on P-A CXR images have made significant progress (Lyu and Tian, 2023), whether the CTR calculation requires the specific morphology and structure of the heart remains to be studied. Anatomically, the heart is located within the chest cavity between the left and right lungs. Specifically, about one-third of the heart is on the right side of the midline, about two-thirds is on the left side, and the apex is on the lower left front (Weinhaus and Roberts, 2005). The transverse diameter of the heart in the cardiothoracic ratio refers to the sum of the maximum distances from the left and right cardiac margins to the midline. However, the heart in the P-A CXR images is not prominent. Therefore, if the heart segmentation needs to be more precise, the error of this heart segmentation may fail to calculate the CTR automatically based on the P-A CXR images.

Based on the above, a novel, fully automatic CTR calculation method based on the lung field should be proposed to overcome the limitations to heart segmentation. Specifically, we train a robust and standard segmentation model of pathological lungs based on multi-center training datasets of the P-A CXR images and image enhancement techniques for extracting lung fields in P-A CXR images. Then, the CTR is automatically calculated based on the lung field based on graphics. Our contributions in this paper are briefly described as follows:

(1) We propose a fully automatic CTR calculation method based on lung fields abstracted from the P-A CXR images using CNNs, overcoming the limitations to heart segmentation, avoiding errors in heart segmentation, and taking less time.

(2) We propose a novel localization method of the heart’s right and left border points based on the two-dimensional projection morphology of the lung field mask images using graphics.

(3) The proposed automatic CTR calculation method based on lung fields abstracted from the P-A CXR images may become an effective tool for initially evaluating cardiac diseases.

2 Materials and methods

2.1 Materials

Here, 789 (635 + 54 + 72 + 15 + 13) sets of the P-A CXR images from public CXR datasets, the Google website, and a case of P-A CXR video are collected in the study for training the CNNs of lung field segmentation and automatic CTR calculation. Specifically, 21 P-A CXR images are used as a test set for evaluating lung field segmentation models. In addition, these 21 P-A CXR images and 13 dynamic P-A CXR images are used to calculate the CTR.

Figure 1 intuitively shows the detailed distribution of these 789 P-A CXR images in each dataset. Specifically, the dataset used in this study includes six sub-datasets (D1–D6). The public dataset D1 (The Shenzhen set — Chest X-ray database) includes 635 static P-A CXR images (324 normal cases and 311 cases with manifestations of tuberculosis). The public dataset D2 (The Shenzhen set — Chest X-ray database) includes 54 static P-A CXR images (47 normal cases and 7 cases with manifestation of tuberculosis). The public dataset D3 [NIAID TB portal program dataset (Online)] and D4 [kaggle. RSNA Pneumonia Detection Challenge (Online)] include 72 static P-A CXR images (7 cases with the manifestation of tuberculosis, 60 cases with the manifestation of pneumonia, and 5 normal cases). The public dataset D5 includes 15 static P-A CXR images collected from the Google website.

Figure 1
www.frontiersin.org

Figure 1. Distribution of the dataset in this study. (A) Distribution of case classification in each dataset and (B) distribution of abnormal cases in each dataset.

In addition, dataset D6 includes 13 dynamic P-A CXR images from the case of CXR video that were collected during free breathing. Specifically, the CXR video was collected from a female participant aged 53 using a digital X-ray imaging system (manufacturer: Lanmage, collection mode: sequence point slice, exposure parameters: 78 KV, 200 mA, 50 ms, and flat panel detector: IRAY) for chest photography. The ethics committee of the National Clinical Research Center for Respiratory Diseases in China approved this study. The subject has been provided written informed consent to the second affiliated hospital of Guangzhou Medical University before chest photography.

More specifically, the public CXR datasets D1 and D2 are collected from the website https://www.kaggle.com/datasets/kmader/pulmonary-chest-xray-abnormalities?select=ChinaSet_AllFiles. Meanwhile, the public CXR datasets D3 and D4 are collected from the websites https://data.tbportals.niaid.nih.gov/ and https://www.kaggle.com/c/rsna-pneumonia-detection-challenge/data, respectively.

2.2 Methods

Figure 2 intuitively shows the schematic diagram of the automatic cardiothoracic ratio algorithm. Specifically, the proposed fully automatic CTR calculation method based on lung field abstracted from the P-A CXR images includes lung field segmentation and cardiothoracic ratio calculation. Figure 2A shows that the lung field mask images are abstracted from the P-A CXR images based on the trained CNNs with the connected domain (CD) algorithm. Meanwhile, Figure 2B shows the automatic CTR calculation method based on the lung field mask images.

Figure 2
www.frontiersin.org

Figure 2. Automatic cardiothoracic ratio algorithm schematic diagram. (A) Lung field segmentation and (B) cardiothoracic ratio calculation.

2.2.1 Lung field segmentation

The organ segmentation model of medical images based on CNNs has become an indispensable technology for quantitative analysis (Conze et al., 2023; Jiang et al., 2023; Ma et al., 2024). CNNs have even been applied to the lung segmentation of rats for measuring lung parenchyma parameters (Yang et al., 2021). In addition, automatic lung field segmentation in routine imaging is a data diversity problem not a methodology problem (Hofmanninger et al., 2020).

The SegNet, U-Net, and its improved networks have been widely applied in organ segmentation of medical images. Based on the above, we train four traditional and basic CNNs to test whether different CNN lung field segmentation models have differences in CTR calculation, including SegNet (Badrinarayanan et al., 2017), U-Net (Ronneberger et al., 2015), and its two improved networks, ResU-Net++ (Jha et al., 2019) and AttU-Net (Wang et al., 2021).

The training process of four lung field segmentation models based on CNNs is detailed below. First, the 755 P-A CXR images’ lung field label images (ground truth) are labeled and examined by three experienced radiologists using the software Labelme (v5.1.0) and ITK-SNAP (v4.0.2). Second, each CNN is trained by 755 P-A CXR images (755 × 512 × 512 × 1) with their lung field label images (ground truth). Specifically, 755 CXR cases include 371 normal cases, 380 abnormal cases with the manifestation of tuberculosis (N = 320) and pneumonia (N = 60), and 15 unclear cases. In addition, data augmentation techniques were adopted to avoid overfitting, further improving the robustness and generalization ability of the lung field segmentation models in the training process (Chlap et al., 2021). The standard cross entropy loss function is selected to calculate the model’s loss and dynamically adjust each CNN’s parameters. Finally, the CD algorithm (Zhao et al., 2010) is applied to the lung field mask images generated by each CNN to eliminate non-lung field masks not connected to the lung field masks.

2.2.2 Automatic CTR calculation

Figure 3 intuitively shows the schematic diagram of the automatic cardiothoracic ratio algorithm. First, the right and left lungs are identified based on lung field mask images. Specifically, the largest and the second largest lung field areas in each lung field mask image are identified as the right and left lung mask images, respectively. Second, edge detection is performed separately on left and right lung images, obtaining the right and left lung mask edge images. Finally, the right and left thoracic inner edge points (D2 and D3) and right and left heart border points (E1 and E2) are located based on the images of the right and left lung mask edge for the CTR calculation. The source code is available on the website: https://github.com/YingjianYang/Automatic-Cardiothoracic-Ratio-Calculation.

Figure 3
www.frontiersin.org

Figure 3. Schematic diagram of the automatic cardiothoracic ratio calculation is based on the images of the lung edge.

Then, the right and left cardiophrenic angles C1 and C2 are configured as detection starting points for the right and left cardiac margins. Two straight lines are drawn perpendicular to the y-axis based on the y2 and y3 coordinates of C1 (x2, y2) and C2 (x3, y3). The intersection points C1 and C2 of these two straight lines with the right and left lung mask edge are configured as the termination points for the right and left cardiac margins. The heart’s right border E1 is located by calculating the maximum distance from all pixels on the right side along the right lung mask edge from the right cardiophrenic angles C1, and the intersection points C1 to the preset line O-O. Meanwhile, the heart’s left border E2 is located by calculating the maximum distance from all pixels on the left side along the left lung mask edge from the left cardiophrenic angles C2 and the intersection points C2 to the preset line O-O. Furthermore, the above algorithm’s Equations 16 are provided.

OO:a0x+b0y+c0=0,(1)
E1x,ymaxDC1C1=maxdr1pr1,dr2pr2,dr3pr3,...,drnprn=maxdr1xr1,yr1,dr2xr2,yr2,dr3xr3,yr3,...,drnxrn,yrn=maxa0xr1+b0yr1+c0a02+b02,a0xr2+b0yr2+c0a02+b02,a0xr3+b0yr3+c0a02+b02,...,a0xrn+b0yrn+c0a02+b02,(2)

where the vector DC1C1=dr1pr1,dr2pr2,dr3pr3,...,drnprn presents the Euclidean distances dri of these coordinates pr1,pr2,pr3,...,prn=xr1,yr1,xr2,yr2,xr3,yr3,...,xrn,yrn to the preset line O-O, and i = 1, 2, …, n. These coordinates xr1,yr1,xr2,yr2,xr3,yr3,...,xrn,yrn are extracted from all pixels in the left lung edge from the right cardiophrenic angles C1 and the intersection point C1 on the left side of the preset line O-O. The parameters a0,b0,c0 are the coefficients of the preset line O-O.

E2x,ymaxDC2C2=maxdl1pl1,dl2pl2,dl3pl3,...,dlnpln=maxdl1xl1,yl1,dl2xl2,yl2,dl3xl3,yl3,...,dlnxln,yln=max(a0xl1+b0yl1+c0a02+b02,a0xl2+b0yl2+c0a02+b02,a0xl3+b0yl3+c0a02+b02,...,a0xln+b0yln+c0a02+b02),(3)

where the vector DC2C2=dl1pl1,dl2pl2,dl3pl3,...,dlnpln presents the Euclidean distances dli of these coordinates pl1,pl2,pl3,...,pln=xl1,yl1,xl2,yl2,xl3,yl3,...,xln,yln to the preset line O-O, and i = 1, 2, …, n. These coordinates xl1,yl1,xl2,yl2,xl3,yl3,...,xln,yln are extracted from all pixels in the left lung edge from the left cardiophrenic angles C2 and the intersection point C2 on the right side of the preset line O-O.

In addition, the preset line O-O is perpendicular to the x-axis (b1 = 0), so the formula (Liu et al., 2022) is converted to Equation 4. Based on Equation 4, Equations 2, 3 are further simplified to Equations 5, 6.

OO:aox+co=0,(4)
E1x,ymaxDC1C1=maxdr1pr1,dr2pr2,dr3pr3,...,drnprn=maxdr1xr1,yr1,dr2xr2,yr2,dr3xr3,yr3,...,drnxrn,yrn=maxaoxr1+coao2,aoxr2+coao2,aoxr3+coao2,...,aoxrn+coao2=maxaoxr1+coao,aoxr2+coao,aoxr3+coao,...,aoxrn+coao,(5)
E2x,ymaxDC2C2=maxdl1pl1,dl2pl2,dl3pl3,...,dlnpln=maxdl1xl1,yl1,dl2xl2,yl2,dl3xl3,yl3,...,dlnxln,yln=maxaoxl1+coao2,aoxl2+coao2,aoxl3+coao2,...,aoxln+coao2=maxaoxl1+coao,aoxl2+coao,aoxl3+coao,...,aoxln+coao.(6)

Finally, the distance between the right and left heart border points is configured as the maximum transverse diameter of the chest Δx1. In addition, the distance between the right thoracic inner edge point and the left thoracic inner edge point D3 is configured as the maximum transverse diameter of the heart Δx2. Subsequently, the ratio of the heart’s maximum transverse diameter and the chest’s maximum transverse diameter Δx2 was configured as the CTR calculation. Furthermore, the above algorithm’s Equations 79 are rovided.

Δx1=xE1xE2=xmaxDC1C1xmaxDC2C2=xmaxdr1pr1,dr2pr2,dr3pr3,...,drnprnxmaxdl1pl1,dl2pl2,dl3pl3,...,dlnpln=xmaxdr1xr1,yr1,dr2xr2,yr2,dr3xr3,yr3,...,drnxrn,yrnxmaxdl1xl1,yl1,dl2xl2,yl2,dl3xl3,yl3,...,dlnxln,yln=xmaxa1xr1+c1a12,a1xr2+c1a12,a1xr3+c1a12,...,a1xrn+c1a12xmaxa1xl1+c1a12,a1xl2+c1a12,a1xl3+c1a12,...,a1xln+c1a12=xmaxa1xr1+c1a1,a1xr2+c1a1,a1xr3+c1a1,...,a1xrn+c1a1xmaxa1xl1+c1a1,a1xl2+c1a1,a1xl3+c1a1,...,a1xln+c1a1,(7)
Δx2=xD2xD3,(8)
CTR=Δx1Δx2,(9)

where xE1 and xE2 represent the horizontal coordinates x(E1) and x(E2) of the right and left heart border points E1 and E2, respectively. Similarly, xD2 and xD3 represent the horizontal coordinates x(D2) and x(D3) of the right and left thoracic inner edge points D2 and D3, respectively.

2.2.3 Experimental environment and evaluation metrics

These four traditional and basic CNNs (SegNet, U-Net, ResU-Net++, and AttU-Net) are trained on PyCharm 2017.3.3 (community edition) in Windows 10 Pro 64-bit with an NVIDIA GeForce GTX 1080 Ti GPU and 16 GB RAM. Then, the pth format of each CNN’s optimal lung field segmentation model is converted to the pth format based on PyCharm 2017.3.3. Finally, each CNN’s optimal lung field segmentation model with the pt format is called by C++ codes based on Visual Studio 2017 for lung field segmentation of 21 static P-A CXR images (Test set T1) and 13 dynamic P-A CXR images (Test set T2). Similarly, the CTR algorithm is automatically performed in Visual Studio 2017.

This study selects the six standard evaluation metrics of each lung field segmentation model, including accuracy, precision, recall, dice, intersection over union (IoU), and the median 95th Hausdorff distance (HD) (Yang et al., 2023; Zeng et al., 2023). Four x-axis direction distance errors are calculated to evaluate detected points between the right and left heart border points and thoracic inner edge points and their ground truths. Furthermore, the errors of calculated CTR and its ground truth were also calculated to evaluate the proposed method. Specifically, the evaluation metrics of the x-axis direction distance and CTR errors are calculated by Equations 10, 11.

xerror=xDPxGT,(10)
CTRerror=CTRCCTRGT,(11)

where xDP represents the horizontal coordinates x(E1), x(E2), x(D2), and x(D3) of the detection points of E1, E2, D2, and D3, respectively. In addition, xGT represents the ground truth of the horizontal coordinates x(E1), x(E2), x(D2), and x(D3). CTRC and CTRGT represent the calculated value and ground truth of the CTR.

3 Results

Figures 4, 5 show the visualized lung field segmentation results and evaluation metrics of the test set T1 based on various trained CNN models. In addition, Figure 6 shows that the visualized lung field segmentation results of the test set T2 are based on these trained CNN models. These results indicate that all CNNs can perform well in lung field segmentation for static and dynamic P-A CXR images.

Figure 4
www.frontiersin.org

Figure 4. Visualized lung field segmentation results of the test set T1 based on various trained CNN models. (A) Twenty-one static P-A CXR images; (B) twenty-one static P-A CXR images with their lung field label images (ground truth); (C) twenty-one static lung field mask images based on SegNet; (D) twenty-one static lung field mask images based on U-Net; (E) twenty-one static lung field mask images based on ResU-Net++; and (F) twenty-one static lung field mask images based on AttU-Net.

Figure 5
www.frontiersin.org

Figure 5. Visualized evaluation metrics of various trained CNN models on the test set T1. (A) Accuracy; (B) precision; (C) recall; (D) dice; (E) IoU; and (F) 95th HD.

Figure 6
www.frontiersin.org

Figure 6. Visualized lung field segmentation results of the test set T2 based on various trained CNN models. (A) Thirteen dynamic P-A CXR images; (B) thirteen dynamic lung field mask images based on SegNet; (C) thirteen dynamic lung field mask images based on U-Net; (D) thirteen dynamic lung field mask images based on ResU-Net++; and (E) thirteen dynamic lung field mask images based on AttU-Net.

Specifically, the mean accuracies (%) of these trained CNN models are 98.93 ± 0.63, 98.93 ± 0.85, 99.02 ± 0.60, and 99.05 ± 0.69, respectively. In addition, the mean precision (%) of these trained CNN models is 97.89 ± 1.49, 93.30 ± 1.40, 97.80 ± 1.96, and 98.36 ± 1.44, respectively. The mean recall (%) of these trained CNN models is 97.55 ± 1.93, 97.31 ± 2.70, 98.05 ± 1.69, and 97.67 ± 2.14, respectively. The mean dice (%) of these trained CNN models is 97.71 ± 1.56, 97.78 ± 1.63, 97.91 ± 1.46, and 97.99 ± 1.43, respectively. The mean IoU (%) of these trained CNN models is 95.57 ± 2.91, 95.71 ± 3.05, 95.95 ± 2.74, and 96.11 ± 2.69, respectively. Finally, the mean 95th HD of these trained CNN models is 5.41 ± 3.81, 5.72 ± 5.02, 5.46 ± 4.51, and 5.02 ± 4.15, respectively.

Meanwhile, Figures 710 show the typical key point detection results of a tuberculosis case, a normal case in test set T1, and a dynamic case in test set T2 based on various trained CNN models. Specifically, the typical key points include the right and left thoracic inner edge points (D2 and D3) and heart border points (E1 and E2).

Figure 7
www.frontiersin.org

Figure 7. Typical key point detection results of a normal case in the test set T1 based on various trained CNN models. (A) SegNet; (B) U-Net; (C) ResU-Net++; and (D) AttU-Net.

Figure 8
www.frontiersin.org

Figure 8. Typical key point detection results of a tuberculosis case in the test set T1 based on various trained CNN models. (A) SegNet; (B) U-Net; (C) ResU-Net++; and (D) AttU-Net.

Figure 9
www.frontiersin.org

Figure 9. Typical key point detection results of a dynamic case in the test set T1 based on various trained CNN models. (A) SegNet; (B) U-Net; (C) ResU-Net++; and (D) AttU-Net.

Figure 10
www.frontiersin.org

Figure 10. Visualized mean distance errors at the x-axis direction of key points on the test set T1. (A) Distribution map of key points’ distance errors based on SegNet; (B) mean with an SD plot of key points’ distance errors based on SegNet; (C) distribution map of key points’ distance errors based on U-Net; (D) mean with an SD plot of key points’ distance errors based on U-Net; (E) distribution map of key points’ distance errors based on ResU-Net++; (F) mean with an SD plot of key points’ distance errors based on ResU-Net++; (G) distribution map of key points’ distance errors based on AttU-Net; and (H) mean with an SD plot of key points’ distance errors based on AttU-Net.

To quantitatively evaluate the key points detection results of Figures 79, Table 1 reports the mean distance errors at the x-axis direction of key points in the test sets T1 and T2 based on various trained CNN models. In addition, Figures 10, 11 show the visualized mean distance errors at the x-axis direction of key points on the test set T1. The mean with SD plots of Figures 10, 11 is drawn based on the absolute values of these distance errors at the x-axis direction of key points.

Table 1
www.frontiersin.org

Table 1. Mean distance errors at the x-axis direction of key points in the sets T1 and T2 based on various trained CNN models.

Figure 11
www.frontiersin.org

Figure 11. Visualized mean distance errors at the x-axis direction of key points on the test set T2. (A) Distribution map of key points’ distance errors based on SegNet; (B) mean with an SD plot of key points’ distance errors based on SegNet; (C) distribution map of key points’ distance errors based on U-Net; (D) mean with an SD plot of key points’ distance errors based on U-Net; (E) distribution map of key points’ distance errors based on ResU-Net++; (F) mean with an SD plot of key points’ distance errors based on ResU-Net++; (G) distribution map of key points’ distance errors based on AttU-Net; and (H) mean with an SD plot of key points’ distance errors based on AttU-Net.

Specifically, mean distance errors at the x-axis direction of key points in the test set T1 based on the four trained CNN models are 3.8213 pixels, 4.1788 pixels, 4.3808 pixels, and 4.0833 pixels, respectively. Meanwhile, mean distance errors at the x-axis direction of key points in the test set T2 based on the four trained CNN models are 3.4423 pixels, 2.7115 pixels, 2.7310 pixels, and 3.9615 pixels, respectively. Mean distance errors at the x-axis direction of key points in the sets T1 and T2 of all trained CNN models are 4.1161 pixels and 3.2116 pixels, respectively. The mean distance error at the x-axis direction of key points in the test sets T1 and T2 based on various trained CNN models is approximately 3.6639 [≈(4.1161 + 3.2116)/2] pixels. Therefore, the deviation degree at the x-axis direction is about 0.72% (3.6639/512) on all 512 × 512 P-A CXR images.

Subsequently, Table 2 compares the mean CTR error in the test sets T1 and T2 based on the previous methods (CardioNet) (Jafar et al., 2022) and our proposed models. In addition, Figure 12 shows the visualized CTR of the test sets T1 and T2 based on various models.

Table 2
www.frontiersin.org

Table 2. Comparison of the mean CTR error in the test sets T1 and T2 based on previous and our proposed models.

Figure 12
www.frontiersin.org

Figure 12. Visualized CTR of the test sets T1 and T2 based on various models. (A) CTR distribution map of the test set T1; (B) mean with an SD plot of CTR on test set T1; (C) CTR distribution map of the test set T2; and (D) mean with an SD plot of CTR on test set T2.

Specifically, the mean CTR values in the test set T1 based on these models are 0.446, 0.454, 0.457, 0.452, and 0.458. In addition, the mean CTR values in the test set T2 based on these models are 0.615, 0.622, 0.644, 0.634, and 0.612. More specifically, the ground truths of CTR in the test sets T1 and T2 are determined by these three experienced radiologists. Mean CTR errors in the test set T1 based on these models are 0.030, 0.022, 0.019, 0.024, and 0.018. In addition, mean CTR errors in the test set T2 based on these models are 0.031, 0.024, 0.002, 0.012, and 0.034. Meanwhile, mean CTR errors in the test sets T1 and T2 based on these models are 0.0305 [(0.030 + 0.031/2)], 0.0230 [(0.022 + 0.024/2)], 0.0105 [(0.019 + 0.002/2)], 0.0180 [(0.024 + 0.012/2)], and 0.0260 [(0.018 + 0.034/2)]. Therefore, the experimental results show that our proposed model achieves the equivalent performance of CTR calculation as the previous CardioNet model.

Table 3 compares the mean segmentation time, mean CTR calculation time, and mean total time of the test sets T1 and T2 based on previous and our proposed models. Specifically, these trained CNN models run on the GPU for segmenting the lung field and/or heart, and then, the CTR calculation algorithm based on the lung field and/or heart mask images runs on the CPU. The previous CardioNet takes more mean segmentation time, mean CTR calculation time, and mean total time of the test sets T1 and T2 than our proposed model. However, when the GPU runs the segmentation task for the first CXR image, it requires an amount of time to configure and load the corresponding model. For example, it takes 4,897/4,885 ms (CardioNet),951/979 ms (SegNet), 2,249/2,350 ms (U-Net), 2,182/2,226 ms (ResU-Net++), and 2,144/2,158 ms (AttU-Net), when the GPU runs the segmentation task for the first CXR image of the test set T1/T2.

Table 3
www.frontiersin.org

Table 3. Comparison of the mean segmentation time, mean CTR calculation time, and mean total time of the test sets T1 and T2 based on previous and our proposed models.

4 Discussion

This section conducts the following discussion and points out this study’s limitations and the future direction based on the experimental results.

4.1 Dynamic lung field segmentation driven by CNNs with static P-A CXR images

The CNN lung field segmentation model trained by inspiratory chest computed tomography (CT) images has been applied to the lung field segmentation of inspiratory and expiratory chest CT images, achieving good performance (Deng et al., 2024; Wang et al., 2024). Similarly, the lung field segmentation model trained on CNNs based on static P-A CXR images also demonstrated good performance in lung field segmentation of dynamic P-A CXR images. Therefore, this provides a necessary foundation for quantitative analysis of dynamic P-A CXR images, such as the CTR calculation. Specifically, CNNs have also played a crucial role in semantic segmentation, where the goal is to assign a class label to each pixel in an image, enabling pixel-level understanding and overcoming the limitations of the traditional approaches (Robert, 2024). In addition, a robust and standard segmentation model of pathological lungs is crucial for quantitative analysis of the lungs based on P-A CXR images. However, generalizing lung field segmentation models based on P-A CXR images has always been a significant engineering problem in clinical applications (Rajaraman et al., 2024). The main reason for this engineering problem is the lack of cross-center P-A CXR images and their diversity. The data augmentation technology enriches the training set of the static P-A CXR images and relieves the engineering problem of generalization in lung field segmentation models (Hasan et al., 2024; Kiruthika and Khilar, 2024). Therefore, the static P-A CXR images form a single center, limiting the generalization of lung field segmentation models. Meanwhile, the diversity of pathological static P-A CXR images in the training set is also essential for improving the generalization of lung field segmentation models, enabling CNNs to learn more prior knowledge. The static P-A CXR images of the cross-center and the diversity and data augmentation techniques may fundamentally solve the generalization problem of lung field segmentation models.

4.2 The lung field morphology for automatic and precise CTR calculation

The positional relationship between the lung field and the heart on the P-A CXR images is the basis for calculating the CTR based on the lung field. Specifically, the right and left cardiophrenic angles C1 and C2 are two relatively prominent points on the P-A CXR images after careful observation and analysis of the P-A CXR images. This provides a certain possibility for automatically calculating the CTR based on the P-A CXR images. Significantly, the right cardiophrenic angle C1 is the farthest point from the line connecting the right apex pulmonis A1 and costophrenic angle B1, which helps determine the highest point D1 on the right hemi-diaphragm from the costophrenic angle B1 to the right cardiophrenic angles C1. Meanwhile, the right and left heart border points E1 and E2 are closely adjacent to the left and right lung fields and form inward indentations on the opposite sides of the left and right lung fields on the P-A CXR images. Therefore, this facilitates the location of the right and left heart border points E1 and E2 on lung mask edges from the right and left cardiophrenic angles C1 and C2 to the intersection points C1 and C2, respectively. Based on the above, the two-dimensional projection morphology of the lung field enables automatic and precise CTR calculation, overcoming the limitations to heart segmentation and avoiding errors in heart segmentation. In addition, our proposed models take less time to calculate the CTR, which benefits that the proposed models only segment the lung field compared with the previous CardioNet model. Our proposed model achieves the equivalent performance of CTR calculation as the previous CardioNet model (Jafar et al., 2022).

4.3 Providing the possibility for the analysis and evaluation of dynamic CTRs

Dynamic CTRs can directly reflect the relationship between the changes in the maximum transverse diameter of the chest during the respiratory process and the maximum transverse diameter of the heart at different cardiac cycles. Actually, the chest can be imaged by autonomously controlling breathing while performing a chest X-ray. However, the heartbeat process cannot be autonomously controlled while chest X-rays are performed. Therefore, the models developed above may provide evidence of the possibility of analyzing and evaluating dynamic CTRs. Specifically, the corresponding P-A CXR images of the different cardiac cycles can be obtained by controlling the breathing state at any moment during the breathing process, such as deep inhalation, deep exhalation, or breath holding. Subsequently, dynamic CRTs for different cardiac cycles can be calculated for clinical analysis and evaluation. Meanwhile, like inspiratory and expiratory chest CT images (Deng et al., 2024; Wang et al., 2024), P-A CXR images of the different cardiac cycles can be obtained by controlling the holding of breath to achieve deep inhalation and exhalation, to analyze the difference in CTR between deep inhalation and exhalation.

4.4 Limitations and future research directions

Although we propose an automated CTR calculation technique based on lung field models from an engineering perspective, our research still has certain limitations. First, the pathological lung image types used for training CNNs are insufficient. Second, we only explain the principle of automatically calculating the CTR based on the lung fields based on graphics and achieve the detection of dynamic CTRs. However, we do not have sufficient dynamic P-A chest X-ray images to further analyze the association between dynamic CTRs and specific lung or heart diseases from a clinical perspective, for example, differences in dynamic CTRs under different GOLD classifications in chronic pulmonary heart disease caused by chronic obstructive pulmonary disease. Based on the above, we encourage researchers to collect more dynamic P-A CXR images of different lung diseases to improve the lung field segmentation model further and hope to discover more clinically significant facts based on dynamic CTRs.

5 Conclusion

We propose an automatic CTR calculation model based on lung fields abstracted from P-A CXR images without heart segmentation. First, the lung field mask images are abstracted from the P-A CXR images based on the pre-trained CNNs. Second, a novel localization method of the heart’s right and left border points is proposed based on the two-dimensional projection morphology of the lung field mask images using graphics. The results show that the mean distance errors at the x-axis direction of the CTR’s four key points in the test sets T1 and T2 based on these pre-trained CNNs are 4.1161 and 3.2116 pixels, respectively. In addition, the mean CTR errors on the test set T1 and T2 based on four proposed models are 0.0208 and 0.0180, respectively. Our proposed model achieves the equivalent performance of CTR calculation as the previous CardioNet model and takes less time. Therefore, our proposed method is practical and feasible and may become an effective tool for initially evaluating cardiac diseases.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Ethics statement

The studies involving humans were approved by all patients who provided written informed consent, and this study was approved by the Guangzhou Medical University Ethics Committee (Grant Number: 2023-hg-ks-24, Approval Date: 28 August 2023, Tel: +86-20-34153599, and Fax: +86-20-34153066). The studies were conducted in accordance with the local legislation and institutional requirements. The participants provided their written informed consent to participate in this study. Written informed consent was obtained from the individual(s) for the publication of any potentially identifiable images or data included in this article.

Author contributions

YY: formal analysis, methodology, software, and writing–original draft. JZ: formal analysis, methodology, software, and writing–original draft. PG: formal analysis, methodology, software, and writing–original draft. TW: software, validation, visualization, and writing–original draft. QG: software, validation, visualization, and writing–original draft. YG: conceptualization, formal analysis, investigation, methodology, resources, supervision, validation, visualization, and writing–review and editing. ZC: validation, visualization, and writing–original draft. CL: investigation, validation, and writing–original draft. ZO: investigation, validation, and writing–original draft. HC: conceptualization, data curation, funding acquisition, project administration, resources, supervision, and writing–review and editing. YK: conceptualization, funding acquisition, project administration, supervision, and writing–review and editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This research was funded by the Zhongnanshan Medical Foundation of Guangdong Province, China (ZNSXS-20230001); the National Natural Science Foundation of China (62071311); and the special program for key fields of colleges and universities in Guangdong Province (biomedicine and health) of China, grant number (2021ZDZX 2008).

Acknowledgments

The authors thank the Department of Radiology, the Second Affiliated Hospital of Guangzhou Medical University, for providing the dataset.

Conflict of interest

Authors YY, JZ, PG, TW, and ZO were employed by Shenzhen Lanmage Medical Technology Co., Ltd. Author QG was employed by Neusoft Medical System Co., Ltd.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors, and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fphys.2024.1416912/full#supplementary-material

References

Badrinarayanan V., Kendall A., Cipolla R. (2017). Segnet: a deep convolutional encoder-decoder architecture for image segmentation. IEEE Trans. pattern analysis Mach. Intell. 39 (12), 2481–2495. doi:10.1109/TPAMI.2016.2644615

CrossRef Full Text | Google Scholar

Chang S. A., Yim J., Yeung D. F., Gin K., Jue J., Nair P., et al. (2022). Relationship between enlarged cardiac silhouette on chest X-ray and left ventricular size on transthoracic echocardiography. Int. J. Cardiovasc. Imaging 38 (4), 771–776. doi:10.1007/s10554-021-02465-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Chlap P., Min H., Vandenberg N., Dowling J., Holloway L., Haworth A. (2021). A review of medical image data augmentation techniques for deep learning applications. J. Med. Imaging Radiat. Oncol. 65 (5), 545–563. doi:10.1111/1754-9485.13261

PubMed Abstract | CrossRef Full Text | Google Scholar

Conze P.-H., Andrade-Miranda G., Singh V. K., Jaouen V., Visvikis D. (2023). Current and emerging trends in medical image segmentation with deep learning. IEEE Trans. Radiat. Plasma Med. Sci. 7, 545–569. doi:10.1109/trpms.2023.3265863

CrossRef Full Text | Google Scholar

Deng X., Li W., Yang Y., Wang S., Zeng N., Xu J., et al. (2024). COPD stage detection: leveraging the auto-metric graph neural network with inspiratory and expiratory chest CT images. Med. Biol. Eng. Comput. 62, 1733–1749. doi:10.1007/s11517-024-03016-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Hada Y. (1995). Cardiothoracic ratio. J. Cardiol. 26 (1), 51–54.

PubMed Abstract | Google Scholar

Hasan D., Abdulazeez A. M. (2024). Lung segmentation from chest X-ray images using deeplabv3plus-based CNN model. Indonesian J. Comput. Sci. 13 (1). doi:10.33022/ijcs.v13i1.3700

CrossRef Full Text | Google Scholar

Hofmanninger J., Prayer F., Pan J., Röhrich S., Prosch H., Langs G. (2020). Automatic lung segmentation in routine imaging is primarily a data diversity problem, not a methodology problem. Eur. Radiol. Exp. 4 (1), 50–13. doi:10.1186/s41747-020-00173-2

PubMed Abstract | CrossRef Full Text | Google Scholar

Howell J. D. (2016). Early clinical use of the X-ray. Trans. Am. Clin. Climatol. Assoc. 127, 341–349.

PubMed Abstract | Google Scholar

Jafar A., Hameed M. T., Akram N., Waqas U., Kim H. S., Naqvi R. A. (2022). CardioNet: automatic semantic segmentation to calculate the cardiothoracic ratio for cardiomegaly and other chest diseases. J. Personalized Med. 12 (6), 988. doi:10.3390/jpm12060988

CrossRef Full Text | Google Scholar

Jha D., Smedsrud P. H., Riegler M. A., et al. (2019). “Resunet++: an advanced architecture for medical image segmentation,” in 2019 IEEE international symposium on multimedia (ISM), USA, 9-11 Dec. 2019 (IEEE), 225–2255.

Google Scholar

Jiang H., Diao Z., Shi T., Zhou Y., Wang F., Hu W., et al. (2023). A review of deep learning-based multiple-lesion recognition from medical images: classification, detection and segmentation. Comput. Biol. Med. 157, 106726. doi:10.1016/j.compbiomed.2023.106726

PubMed Abstract | CrossRef Full Text | Google Scholar

Kiruthika K., Khilar R. (2024). Segmentation of lung on P-A CXR images based on CXR-auto encoder segmentation with MRF. Opt. Quantum Electron. 56 (4), 1–15.

CrossRef Full Text | Google Scholar

Li Z., Hou Z., Chen C., Hao Z., An Y., Liang S., et al. (2019). Automatic cardiothoracic ratio calculation with deep learning. IEEE Access 7, 37749–37756. doi:10.1109/access.2019.2900053

CrossRef Full Text | Google Scholar

Liu W., Luo J., Yang Y., Wang W., Deng J., Yu L. (2022). Automatic lung segmentation in chest X-ray images using improved U-Net. Sci. Rep. 12 (1), 8649. doi:10.1038/s41598-022-12743-y

PubMed Abstract | CrossRef Full Text | Google Scholar

Lyu Y., Tian X. (2023). MWG-UNet: hybrid deep learning framework for lung fields and heart segmentation in chest X-ray images. Bioengineering 10 (9), 1091. doi:10.3390/bioengineering10091091

PubMed Abstract | CrossRef Full Text | Google Scholar

Ma J., He Y., Li F., Han L., You C., Wang B. (2024). Segment anything in medical images. Nat. Commun. 15 (1), 654. doi:10.1038/s41467-024-44824-z

PubMed Abstract | CrossRef Full Text | Google Scholar

Nochioka K., Kato S., Pileggi G., Horii K., Shiroto T., Shimokawa H., et al. (2023). Serial assessment of cardiothoracic ratio as a predictor of progression from stage B to stage C heart failure in asymptomatic patients with cardiac diseases. IJC Heart & Vasc. 48, 101273. doi:10.1016/j.ijcha.2023.101273

CrossRef Full Text | Google Scholar

Rajaraman S., Zamzmi G., Yang F., Liang Z., Xue Z., Antani S. (2024). Uncovering the effects of model initialization on deep model generalization: a study with adult and pediatric chest X-ray images. PLOS Digit. Health 3 (1), e0000286. doi:10.1371/journal.pdig.0000286

PubMed Abstract | CrossRef Full Text | Google Scholar

Robert A. (2024). The role of deep learning in computer vision. China: EasyChair.

Google Scholar

Ronneberger O., Fischer P., Brox T. (2015). “U-net: convolutional networks for biomedical image segmentation[C]//Medical Image Computing,” in Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Munich, Germany, October 5-9, 2015 (Springer International Publishing), 234–241.

CrossRef Full Text | Google Scholar

Saiviroonporn P., Rodbangyang K., Tongdee T., Chaisangmongkon W., Yodprom P., Siriapisith T., et al. (2021). Cardiothoracic ratio measurement using artificial intelligence: observer and method validation studies. BMC Med. Imaging 21 (1), 95. doi:10.1186/s12880-021-00625-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Saul C. (1919). The cardiothoracic ratio: an index of cardiac enlargement. Am. J. Med. Sci. 157, 513–554. doi:10.1097/00000441-191904000-00007

CrossRef Full Text | Google Scholar

Seah J. C. Y., Tang C. H. M., Buchlak Q. D., Holt X. G., Wardman J. B., Aimoldin A., et al. (2021). Effect of a comprehensive deep-learning model on the accuracy of chest x-ray interpretation by radiologists: a retrospective, multireader multicase study. Lancet Digital Health 3 (8), e496–e506. doi:10.1016/S2589-7500(21)00106-0

PubMed Abstract | CrossRef Full Text | Google Scholar

Wang S., Li L., Zhuang X. (2021). AttU-Net: attention U-Net for brain tumor segmentation[C]//International MICCAI Brainlesion Workshop. Cham: Springer International Publishing, 302–311.

Google Scholar

Wang S., Li W., Zeng N., Xu J., Yang Y., Deng X., et al. (2024). Acute exacerbation prediction of COPD based on Auto-metric graph neural network with inspiratory and expiratory chest CT images. Heliyon 10, e28724. doi:10.1016/j.heliyon.2024.e28724

PubMed Abstract | CrossRef Full Text | Google Scholar

Weinhaus A. J., Roberts K. P. (2005). Anatomy of the human heart. Handb. cardiac Anat. physiology, devices, 51–79. doi:10.1007/978-1-59259-835-9_4

CrossRef Full Text | Google Scholar

Yang Y., Li Q., Guo Y., Liu Y., Li X., Guo J., et al. (2021). Lung parenchyma parameters measure of rats from pulmonary window computed tomography images based on ResU-Net model for medical respiratory researches. Math. Biosci. Eng. 18 (4), 4193–4211. doi:10.3934/mbe.2021210

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang Y., Li W., Kang Y., Guo Y., Yang K., Li Q., et al. (2022). A novel lung radiomics feature for characterizing resting heart rate and COPD stage evolution based on radiomics feature combination strategy. Math. Biosci. Eng. 19, 4145–4165. doi:10.3934/mbe.2022191

PubMed Abstract | CrossRef Full Text | Google Scholar

Yang Y., Zeng N., Chen Z., Li W., Guo Y., Wang S., et al. (2023). Multi-layer perceptron classifier with the proposed combined feature vector of 3D CNN features and lung radiomics features for COPD stage classification. J. Healthc. Eng. 2023, 3715603. doi:10.1155/2023/3715603

PubMed Abstract | CrossRef Full Text | Google Scholar

Zeng X., Guo Y., Zaman A., Hassan H., Lu J., Xu J., et al. (2023). Tubular structure segmentation via multi-scale reverse attention sparse convolution. Diagnostics 13 (13), 2161. doi:10.3390/diagnostics13132161

PubMed Abstract | CrossRef Full Text | Google Scholar

Zhao J., Yu H., Gu X., et al. (2010). The edge detection of river model based on self-adaptive Canny Algorithm and connected domain segmentation[C]//2010 8th World congress on intelligent control and automation. IEEE, 1333–1336. doi:10.1109/wcica.2010.5554869

CrossRef Full Text | Google Scholar

Keywords: cardiothoracic ratio, chest X-ray images, lung field segmentation, edge detection, convolutional neural network, graphics

Citation: Yang Y, Zheng J, Guo P, Wu T, Gao Q, Guo Y, Chen Z, Liu C, Ouyang Z, Chen H and Kang Y (2024) Automatic cardiothoracic ratio calculation based on lung fields abstracted from chest X-ray images without heart segmentation. Front. Physiol. 15:1416912. doi: 10.3389/fphys.2024.1416912

Received: 13 April 2024; Accepted: 22 July 2024;
Published: 08 August 2024.

Edited by:

Domenico L. Gatti, Wayne State University, United States

Reviewed by:

Sharon Ackerman, Wayne State University, United States
Seeya Awadhut Munj, Wayne State University, United States

Copyright © 2024 Yang, Zheng, Guo, Wu, Gao, Guo, Chen, Liu, Ouyang, Chen and Kang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Yingwei Guo, Z3VveWluZ3dlaTg4MDFAMTYzLmNvbQ==; Huai Chen, Y2hlbmh1YWkxOTc3QDE2My5jb20=; Yan Kang, a2FuZ3lhbkBzenR1LmVkdS5jbg==

These authors have contributed equally to this work

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.