Skip to main content

ORIGINAL RESEARCH article

Front. Med., 17 August 2022
Sec. Infectious Diseases – Surveillance, Prevention and Treatment
This article is part of the Research Topic COVID-19: Integrating Artificial Intelligence, Data Science, Mathematics, Medicine and Public Health, Epidemiology, Neuroscience, Neurorobotics, and Biomedical Science in Pandemic Management, volume II View all 92 articles

Deep learning framework for prediction of infection severity of COVID-19

  • 1Department of Physics, Shahid Beheshti University, Tehran, Iran
  • 2School of Computer Science, Institute for Research in Fundamental Sciences (IPM), Tehran, Iran
  • 3Research Center for Molecular and Cellular Imaging, Tehran University of Medical Sciences, Tehran, Iran
  • 4Department of Medical Physics, School of Medicine, Mashhad University of Medical Sciences, Mashhad, Iran
  • 5Department of Medical Physics and Biomedical Engineering, Tehran University of Medical Sciences (TUMS), Tehran, Iran
  • 6Department of Radiology, Health Research Center, Baqiyatallah University of Medical Sciences, Tehran, Iran

With the onset of the COVID-19 pandemic, quantifying the condition of positively diagnosed patients is of paramount importance. Chest CT scans can be used to measure the severity of a lung infection and the isolate involvement sites in order to increase awareness of a patient's disease progression. In this work, we developed a deep learning framework for lung infection severity prediction. To this end, we collected a dataset of 232 chest CT scans and involved two public datasets with an additional 59 scans for our model's training and used two external test sets with 21 scans for evaluation. On an input chest Computer Tomography (CT) scan, our framework, in parallel, performs a lung lobe segmentation utilizing a pre-trained model and infection segmentation using three distinct trained SE-ResNet18 based U-Net models, one for each of the axial, coronal, and sagittal views. By having the lobe and infection segmentation masks, we calculate the infection severity percentage in each lobe and classify that percentage into 6 categories of infection severity score using a k-nearest neighbors (k-NN) model. The lobe segmentation model achieved a Dice Similarity Score (DSC) in the range of [0.918, 0.981] for different lung lobes and our infection segmentation models gained DSC scores of 0.7254 and 0.7105 on our two test sets, respectfully. Similarly, two resident radiologists were assigned the same infection segmentation tasks, for which they obtained a DSC score of 0.7281 and 0.6693 on the two test sets. At last, performance on infection severity score over the entire test datasets was calculated, for which the framework's resulted in a Mean Absolute Error (MAE) of 0.505 ± 0.029, while the resident radiologists' was 0.571 ± 0.039.

1. Introduction

Coronavirus 2019, or COVID-19, is a pandemic infectious disease that was reported in December 2019 from Wuhan, China, following an outbreak of the acute respiratory syndrome virus SARS-CoV-2 (14). Common symptoms include fever, cough, shortness of breath, and lethargy. Muscle pain, sputum production, sore throat, nausea, and red eyes are some of the less common symptoms (5, 6).

Reverse Transcription Polymerase Chain Reaction (RT-PCR) test is considered the de facto standard for diagnosing COVID-19 (7). However, lack of resources and strict environmental requirements limit the rapid and effective testing and screening of suspects. Moreover, RT-PCR has been reported having a high false-negative rate and a low sensitivity (8, 9).

Artificial Intelligence (AI) has risen its position in helping solve healthcare challenges in recent years. By involving a large amount of data, using advanced deep learning algorithms, and utilizing modern GPUs, exemplary achievements have been made in the fields of image classification and image segmentation (1012). AI algorithms have demonstrated equal if not higher performance in medical image diagnosis in recent years while being fast and utilizable in pandemic situations. For example, Qin et al. (13), Ardila et al. (14), Nash et al. (15), Liao et al. (16), Mei et al. (17), Zhu et al. (18), and Xie et al. (19) have introduced algorithms with very successful and reproducible performance levels on pulmonary diseases such as tuberculosis and lung nodules and for lung cancer screening. With enough hardware resources, these algorithms could perform on a very large scale.

Machine learning and deep learning models have directly challenged the Coronavirus and have been successful in diagnosing the virus with high accuracy and reducing manpower efforts (2022). Several deep learning models have been developed to diagnose COVID-19 from chest X-Ray or CT scans (17, 2325). Furthermore, the fields of machine learning and data science have also been used effectively to diagnose and prognoses the virus and predict its outbreaks (2628). Finally, it has been shown that a model serving as an assistant to the radiologists is successful in diagnosing the virus and in addition to increasing the expert's sensitivity, is also effective in increasing their specificity (9).

One of the applications of deep learning in computer vision, and in particular medical imaging, is segmentation, in which trained models automatically separate parts of the image (2931). Several pieces of researches have been conducted in the field of infection segmentation and measuring the volume of infection for COVID-19 (3235).

Pulmonary lobe segmentation has proven to be an important task as knowing the location and distribution of pulmonary diseases such as emphysema and nodule can be integral in determining the most suitable treatment. To this end, models for lobe segmentation have also been developed (3639). As an example, Hofmanninger et al. (40) has introduced a well-performing lobe segmentation model trained on patients diagnosed with COVID-19.

Measuring the severity of a lung infection in patients with COVID-19 is a very challenging task and is an important prognosis for a patient's treatment process. Several diagnosis methods, some specifically designed to assess the severity of the disease, have been proposed based on the observation that the imaging biomarkers in patients with COVID-19 such as Ground-Glass Opacity (GGO) and infection-associated thickening of the interlobular septa are similar (4144). Additionally, several similar works have tried to predict the severity of lung infection with the help of neural networks and machine learning models (42, 4547).

In this work, we propose a framework for accurate lung lobe infection severity prediction. We do so by collecting and labeling 253 chest CT-scan from three hospitals, involving 59 labeled external scans from public datasets, and with the help of several deep learning and machine learning models. We perform lobe and infection segmentation to predict lung lobes' infection severity percentage and classify that percentage into an infection severity score as the framework's final output.

In short, the main advantages and novelties of our work are as follows:

• In contrast to previous works, we isolate and predict the infection severity within each lung lobe.

• Our research involves three datasets collected internally, as well as two additional public datasets for more verity and range.

• The data involved was labeled for lobe segmentation, infection segmentation, and infection severity. In addition, our test sets were labeled with more precision and accuracy.

• Separate distinct deep learning models were trained for the lobe and infection segmentation tasks and a machine learning model was trained for predicting the infection severity score.

• For a more thorough evaluation, our train and test data are from different hospitals.

• Our framework's performance was comprehensively compared with the performance of resident radiologists, for which our framework outperformed.

• In a post-COVID world, this framework is usable and extendable to other pneumonia infections.

The rest of this paper is structured as follows; In the second section, we explain the data involved and the used methods for this research in detail. In the third section, we outline our model's results for lobe segmentation and infection segmentation and present the framework's performance in infection severity prediction, in addition to comparing our performance result's with that of human experts. Finally, in the last section, we conclude with a discussion on the results, our work's limitations and challenges, and possible future directions.

2. Data and method

In this paper, we first adopt a deep learning model to detect the lobes. Next, by utilizing the subset of the data with an infection mask label to train an infection segmentation model, we finally combine the results with a k-Nearest Neighbors (k-NN) model to reach a final infection severity prediction.

In this section, we start by describing the different datasets involved and an overview of the distinct used pre-processes in Section 2.1. In Sections 2.3, 2.4, we discuss the adopted deep learning models and infection severity prediction methods in each lobe, respectively. Section 2.5 discloses our evaluation methods and criteria. Finally, we conclude the section with an explanation for our methods of evaluating radiologists and residents in Section 2.6.

2.1. Datasets

Three datasets for training and two datasets for evaluation were involved in this work. A 232-case cohort was collected from the Ghiassi Hospital database and the MosMedData (48, 49) datasets that are available publicly. Two smaller sets of External Test 1 and External Test 2 from Kasra Hospital and Imam Hossein Hospital, respectively, were involved in the model's final evaluation. All three aforementioned centers are located in Tehran, Iran. A complete taxonomy of the data can be found in Table 1.

TABLE 1
www.frontiersin.org

Table 1. Involved datasets with their respective label count.

Three distinct scan labels are included in this work. The first is for infection segmentation and was extracted from 81 subjects. The second is for lobe segmentation which is extracted from a chest CT scan (masked with 5 colors; outside of the lung masked with black) in order to isolate each lobe; and the third is for lobe infection severity which we explain next.

According to the systematic report standard found on the Radiologyassistant (50) website, each severity percentage is categorized from 0 to 5. This can be found in Table 2. Radiology experts, relying on their experience and expertise in interpreting chest CT scans and their knowledge of the cross-sectional anatomy of different areas of the lung, have performed the lung lobe divisions and visually estimated the infection-resulted severity, which were present in a myriad of forms notably GGO and consolidation, in each lobe lattice with percentage and standard systematic report.

TABLE 2
www.frontiersin.org

Table 2. Lobe infection severity percentage categorized by the number of points.

In this research, 2 radiology residents generated systematic reports for all the scans in the study by referring to Table 2. Moreover, the External Test sets 1 and 2 scans that were labeled by two radiology residents were rectified by two senior radiologists each with more than 10 years of experience.

Lobe segmentation was performed semi-supervised by inputting the DICOM of scans into the 3D Slicer software. Within the Interactive Lobe Segmentation section, the Chest Imaging Platform (CIP) module is used to generate the Label Map Volume, in complement with a Fisser Volume file, for the selected scans. A Gaussian filter is applied to enhance the lobe segmentation performance. Other parameters, such as dimensions are chosen by an expert.

By marking several points on the generated Fissers in different scan views, especially sagittal, all five lobes are segmented and distinctly colored. The resulting export is then evaluated for authenticity by two radiology experts and if needed, is rectified by a skilled technician (rectification follows the standard procedure).

The Ghiassi Hospital cohort scans were taken using a TOSHIBA 16 CT scan machine. Each scan was taken with a low-dose setting and has a slice thickness of 2 mm. The 232-case cohort overall includes 30,157 axial view slices, containing 60 scans with infection segmentation and 54 with lobe segmentation labels. The entire cohort is labeled with lobe infection severity labels. The age and sex distribution of cohort subjects can be found in Supplementary Figure S1.

As lobe segmentation and infection labeling is a time-consuming process, it was performed restrictively and to necessity by two resident radiologists. The resulting set was involved in the training and validation of the machine learning model. Two additional sets, External Test 1 and External Test 2 were collected for final evaluation. The External Test 1 includes 11 scans from patients from Kasra Hospital, Tehran, Iran, captured on a General Electric (GE) CT scan machine with 120 KV and 130 MA parameters and 7 mm slice thickness, totaling 481 2D slices. The External Test 2 contains 10 scans from patients from Imam Hossein Hospital, Tehran, Iran, captured with a low-dose setting and 2 mm slice thickness. The obtained scans were in an axial view, each unprocessed slice in all the datasets was 512 × 512, and each pixel had a dimension of 0.76 × 0.76mm. Labeling for the two test sets was performed manually by two resident radiologists and subsequently revised by two expert senior radiologists.

Detailed information on the two public datasets involved in this research, MosMedData and Medical_Seg, are brought in Morozov et al. (48) and Jenssen (49). From the overall combination of these two sets, 80% was used for training and 20% for the validation of the framework.

Charts on the number of normal and infected slices, distribution of infection severity in scan slices, and sample count in each 6 class for different degrees of infection separated by lobe and for the training, validation, and test sets are described in Supplementary Figures S1S7.

All collected data from hospitals in this research has been anonymized, official permissions have been obtained from the relevant department heads or supervisors, written consents were taken from all the participating patients, and the ethical license of IR.SBMU.NRITLD.REC.1399.024 was obtained from the Iranian National Committee for Ethics in Biomedical Research.

2.2. Data pre-processing

In this section, we describe the used pre-processes in detail; starting with the image pre-processing for infection segmentation models and following with the mask pre-processing for infection segmentation model outputs.

2.2.1. Image pre-processing for infection segmentation models

In this pre-processing, first, a 3D resizing on the coronal, axial, and sagittal views is applied using the zoom function (51) with regards to the 2D cross-sections. For the 2D slices, the 3D image is initially resized to the following dimension:

512×512×number of slices,

but for extracting 2D slices from coronal and sagittal views, the 3D image, regardless of its initial dimensions, is resized to 256 × 256 × 256, and 256 slices are selected subsequently.

The window-level and window-width parameters were set to –600 and 1,500, respectively. This resulted in the image pixel intensity distribution to position between –1,024 (the lowest pixel value in all images) and 150. Next, this pixel range value was moved to the range [0, 255] using a linear transformation. The final result yielded three exact images from each 2D slice. Lastly, as our models utilized weights obtained from training on the ImageNet [a dataset of 1.2 million images categorized to 1,000 classes (10)], all images were also additionally normalized to the same luminance of mean = [0.485, 0.456, 0.406] and SD = [0.229, 0.224, 0.225]. Data augmentation operations such as random white-noise addition and vertical-flip with probabilities 0.3 and 0.5, respectively, were also performed for the model's training.

2.2.2. Infection mask pre-processing for training infection segmentation models

This pre-processing was only applied to the manual segmentation masks on the training and validation sets just before the model learning and its goal was to rectify manual mask edges. To this end, the manually masked area is enlarged using the Dilation (52) method to the point that it covers all of the infection areas initially left out. As seen in Figure 1, the red contour depicts the manual mask edge, while the blue contour shows the enlarged manual mask.

FIGURE 1
www.frontiersin.org

Figure 1. Mask pre-processing in the infection segmentation dataset. Left is the raw CT scan slice, middle is the lung image in the same slice (with infection in the upper-left lobe), and right is the segmented infection area in the image. In the right, the red contour depicts the radiologist manual mask edge, the blue contour is the dilated red contour, and the green contour s the rectified mask edge.

For the final mask, pixels contained within the blue contour that have a value within the 95% distribution range of all the pixels within the red contour are kept (and regarded as “infection” pixels) while the other non-relevant pixels within the blue contour are removed. The resulting final mask edge colored in green is also depicted.

This process helps with mask-edge correction by removing “normal” (for air, etc.) and very bright (organ or bone) pixels from the initial area and adding any left-out small areas.

Finally, this pre-processing method was carefully and manually reviewed on a large sample set ad after evaluation from an expert radiologist, its resulting rectified masks replaced undesirable ones. The test set rectified masks went through a more thorough and manual evaluation process under two senior radiologists.

2.3. Deep learning models

The deep learning models adopted by this research perform two tasks; lobe segmentation and infection segmentation. In this section, we explain each in detail.

2.3.1. Lobe segmentation deep learning models

For this task, the deep learning models described in Hofmanninger et al. (40) were adopted. These models include U-Net (53), ResU-Net (54, 55), Dilated Residual Network-D-22 (56), and Deeplab v3+(57). From these models, the U-Net R231 (40) model performed the best on our test set data, for which the results can be found in Section 3, and was thereby selected for our framework. The output of the selected model would segment a lung image to its five lobes using 2D slices in the axial view and as its results were sufficiently desirable, we refrained from developing our custom model for the lobe segmentation task. The model performance on our dataset is discussed in Section 3.

2.3.2. Infection segmentation deep learning models

For this task, we adopted several FPN (58), PSPNet (59), LinkNet (60), U-Net (53), and U-Net++ (61) models based on their respective EfficientNet (62), SE-ResNet (63), etc. architectures. As the performance of these models was closely similar, a detailed discussion on the subject was omitted.

In the end, an SE-ResNet-18 based U-Net model was selected for our framework. Three separate instances of the model are used. The first is for infection segmentation on 2D slices of the axial view. The input for this model instance, with dimensions 512 × 512 × 3 (3 is for the RGB channels) is the largest it can be in order to extract as many features as possible. The second and third model instances are used similarly for the coronal and sagittal views, respectively, but with a 256 × 256 × 3 input dimension. Overall, as previously discussed in Table 1, 119 CT scans with an infection segmentation label were used for the training and validation of these models.

The encoder part of these model was initialized with pre-trained ImageNet weights and their last layer activation function is Sigmoid. Additionally, the prediction layer of the models includes an infection detection channel. The cost function used for all the models is:

Loss=(12×PTP+T)          +(1Ni=1N[Tilog(Pi)+(1Ti)log(1Pi)]),

where T and P are the pixel label and prediction, respectively, the first RHS term is the Dice loss, and the second RHS term is the binary cross-entropy error function.

The Keras package was used for everything deep-learning related and the NiBabel (51) and PyDicom (64) packages were used for working with medical images. All of the development and evaluation processes were executed with the Python programming language version 3.7.9.

2.4. Framework overview

In this section, we break down our framework into its principal machine learning and deep learning components, as showcased in Figure 2, and characterize the final output construction methodology.

FIGURE 2
www.frontiersin.org

Figure 2. Overview of the lung lobes infection severity prediction framework. An input image is simultaneously given the lobe segmentation and infection segmentation models. Then, by combining the output of these two models, the infection percentage of each lobe is predicted and given as the input of the k-NN model to predict the severity of infection in terms of the 6 classes of infection severity for all the 5 lobes.

On an input, we first utilize the lobe segmentation model to determine what lobe each pixel belongs to (or if it does not belong to any). Next, we employ the three distinct U-Net models mentioned in Section 2.3.2 to obtain infection segmentation across the three different views. The outputs are then combined with a weighted-averaging ensemble learning technique (a weight of 5 for the sagittal and coronal views and a weight of 1 for the axial view) to produce a final infection segmentation for the framework input.

As the axial view in our data has a higher image resolution and will probably result in higher model accuracy, the used ensemble learning technique gives it a larger weight compared to the similar coronal and sagittal view weights. This assumption is studied in Section 3.

Next, using the ensemble model output, each lobe pixel is evaluated in terms of containing infection to determine the overall percentage of each lobe's pixels involved with infection. Finally, the lobes infection severity percentage is fed to a k-NN model to learn according to the label given by the specialist in order to classify the percentage into the 6 classes of infection severity predicted by the experts.

The k-NN model is used since the expert did not necessarily calculate the infection severity by counting the pixels, but based on the experience in the field. The overall structure of the flow described in this section is showcased in Section 3.

2.5. Statistical inference

In order to more reliably evaluate our results, we chose different criteria for different parts of our research. Our results are also reinforced by performing the same evaluations on a team of experts.

By incorporating error propagation and Bayesian statistics, the marginalized confidence region is calculated at a 95% level for each output. The prediction result significance is determined by calculating the p-value statistics systematically. In order to achieve a conservative decision, the 3σ significance level is considered.

The lobe and infection segmentation models were evaluated with the following criteria on the scan:

Dicescore=2×PT+ϵP+T+ϵ

where ϵ is a small value, added to prevent the denominator from becoming zero.

At last, the Mean Absolute Error (MAE) was used to evaluate the infection severity prediction performance.

2.6. Experts evaluation

As the train and validation sets of our data were manually labeled by two resident radiologists and were therefore prone to error, and since the segmentation labeling and infection severity categorization tasks were time-consuming and the data for it was not present in radiology reports by default, we opted to involve the data with the least possible amount of error in our framework evaluation. To this end, the External Test sets 1 and 2 were collected which were also used to evaluate our experts.

For the evaluation of experts, which included two resident radiologists, we asked them to label each case with segmentation and infection severity labels. Labels from one expert were regarded as the ground truth while labels from the other as a prediction. The metric similarity between the two experts is reported as the minimum expert accuracy. It is important to note that the labels produced by the two resident radiologists were rectified by the two senior radiologists for the final test set. This difference between the initial labels and the rectified labels is why the expert evaluation metric is accompanied by a bias value.

3. Results

In this research, overall 291 CT scans were involved in training and validation and 21 CT scans for final evaluation. This section starts off and continues with the evaluation results for the lobe segmentation and infection segmentation models, respectively, and closes with the framework performance evaluation and its comparison with that of experts.

3.1. Lobe segmentation results

For this task, a pre-trained model from Hofmanninger et al. (40) was adopted as its performance was sufficiently desirable. This model was evaluated on 21 CT scans based on the Dice score for which the results are shown in Table 3. Furthermore, a sample output of this model for segmenting a lung's lobes is presented in Figure 3.

TABLE 3
www.frontiersin.org

Table 3. Average Dice score of the lobe segmentation model on 21 CT scans.

FIGURE 3
www.frontiersin.org

Figure 3. Lobe segmentation from three different views.

3.2. Lobe infection severity prediction

Sixty scans from our own dataset and two 50 and 9 scan sets from the involved public datasets, all with infection segmentation labels, were used to train three U-Netbased SE-ResNet18 for the axial, coronal, and sagittal views and their outputs were combined using a weighted-average ensemble learning technique to gain a final infection segmentation result.

These models were evaluated on the validation and External Test sets 1 and 2 using Dice score and the final ensemble performance results can be seen in Table 4.

TABLE 4
www.frontiersin.org

Table 4. Dice score of the infection segmentation models and their comparison with the performance of two resident radiologists.

As the results demonstrate, the infection segmentation Dice score in the coronal and sagittal views are lower than the axial view, for which the ensemble performance of the three views barely matches. Moreover, the two resident radiologists' Dice score is on average lower than our framework. As the validation set was not labeled by both radiologists, the experts' performance over this set is not reported.

By incorporating the obtained lung lobe segmentation in the CT scan and having the infection segmentation model output, our framework predicts the overall infection severity with a number in the range [0, 100], which we report as a percentage.

In Figure 4, the framework's output for lobe segmentation, infection segmentation, and per-lobe infection severity prediction is showcased for a slice from a COVID-19 diagnosed patient axial chest CT scan.

FIGURE 4
www.frontiersin.org

Figure 4. Framework output for a COVID-19 diagnosed patient scan on the 72nd slice of the axial view. On the left, red contours show the infection and on the right, different colored contours show distinct lung lobes. The reported infection severity percentages correspond to each lobe and are reported the same for each slice.

But as the experts' prediction of the infection severity is purely visual and not by infected pixel count, our framework categorizes the infection severity percentage into 6 distinct levels utilizing a simple k-NN model with k = 7 with the combined outputs of the ensemble lobe and infection segmentation as its input to learn over infection severity score manually labeled by the experts. This k-NN model was eventually evaluated on 21 scans from the External Test sets 1 and 2, for which the framework achieved an MAE error of 0.505 ± 0.029 on all lung lobes. A more detailed overview of the framework's performance can be seen in Table 5. For a more comprehensive evaluation, the MAE error was calculated for the two resident radiologists. The expert's error of 0.571 ± 0.039 was obtained, showing the better performance of our framework over expert human prediction.

TABLE 5
www.frontiersin.org

Table 5. Model and expert (two resident radiologists) MAE error for different lung lobes over the 21 scans of our test sets.

As seen in Table 5, the prediction error for the right middle lobe is larger than other lobes due to this lobe being generally more difficult to predict for the framework (with the lowest Dice score of 0.918) and the experts.

To gain a better insight into the relationship between the predicted infection severity percentage by the framework and its corresponding infection severity score label, the prediction distribution over all 6 lobe infection severity classes are showcased as a violin plot in Figure 5. As depicted in the figure, the largest errors belong to classes 0 (normal) and 1 (infection severity lower than 5%). Notably, the infection severity percentages of these two classes are marginally close.

FIGURE 5
www.frontiersin.org

Figure 5. Violin plot of infection severity percentage predicted by the framework for 6 different classes.

In a similar manner, the violin plot of labels produced by resident radiologists is shown in Figure 6, which clearly demonstrated the error produced by classes 0 and 1 while asserting the fact that diagnosing smaller infections is generally harder.

FIGURE 6
www.frontiersin.org

Figure 6. Violin plot of infection severity percentage predicted by experts for 6 different classes.

4. Conclusion and discussion

In this research, we collected chest CT scans of patients positively diagnosed with COVID-19 for training from one center and additional scans from two other hospitals with different imaging devices. Furthermore, two public datasets were also involved to include a wide range of data from different centers and several countries. These datasets were then labeled with their corresponding lobe segmentation, infection segmentation, and infection severity labels.

Next, we extracted lung lobes in each scan using our framework's lobe segmentation model. Thereafter, lung infection segmentation was performed utilizing several task-specific deep learning models in an ensemble manner. Finally, by measuring infection severity percentage and incorporating a k-NN model, each lung lobe was predicted for its infection severity score.

For lobe segmentation, we adopted the pre-trained U-Net model from Hofmanninger et al. (40). This model performed satisfactorily as it obtained a Dice score of 0.958 on all lung lobes except for the middle-right, for which the Dice score was 0.918. As diagnosing the middle-right lobe is generally more difficult for human experts, the model's lower Dice score in this lobe might have resulted from the labeling error.

For infection segmentation, the framework achieved a 0.725 Dice score, similar to that of the resident radiologists on the External Test 1 set. The framework's performance on the External Test 2 set was marginally better. Our model's performance and the fact that these two test sets were collected from two different centers with different imaging devices attest that our framework is robust to the imaging device configuration and parameters. Yet, as the results show, this claim did not hold for the human experts. Overall, our framework demonstrates performance at levels similar to that of at least a resident radiologist.

Finally, we compared the MAE error of infection severity prediction between our framework and a resident radiologist. On all the combined data from the two External Test sets 1 and 2 and over all the lung lobes, our framework achieves a lower error compared to the resident radiologist, which is shown in Table 5. Moreover, the correlation between infection severity percentage and infection severity score categorized in 6 levels was studied and showcased via violin plots for both our framework and the human experts. This study showed that the difference between classes 0 and 1 is marginally small and differentiating these two classes yields the largest error.

One of the previous works in the field, Ma et al. (32), developed an infection segmentation deep learning model with a Dice score on the right and left lungs equal to 97.7 and 97.3, respectively. Li et al. (33) managed to develop a deep learning model for the same task to gain a 0.74 Dice score, while the expert score in their study was close to their model at 0.76. Voulodimos et al. (35) developed an FCN and a U-Net model for infection segmentation with a Dice score peaking at around 0.65. Lastly, Abdel-Basset et al. (34) developed a novel model for learning on a small-sized labeled set denoted Few-Shot Segmentation with a Dice score of 0.80. An overview of the evaluation results in this section can be found in Table 6.

TABLE 6
www.frontiersin.org

Table 6. Infection segmentation performance comparison over several sample sets against several similar works.

The majority of similar research in the field aims to predict infection severity in the entire lung, while our work narrows down and isolates the prediction to each lobe. In a similar fashion (65), the author developed a Support Vector Machine (SVM) model utilizing the probability density function to classify lung lobe infection severity with an Area Under Curve (AUC) score in the range of [0.64, 0.87] on the validation set.

In our initial assessments for this research, we also experimented on 3D convolution models with a 3D-UNet model for the infection segmentation for which its lower performance results made us refrain from discussing it. However, the interested reader is encouraged to study and evaluate different models and methods for this task.

We used an Nvidia RTX 5000 graphics card for the computational portions of this work. For a CT scan of about 100–150 slices, our framework will perform lobe segmentation, infection segmentation, and infection severity prediction on all slices and finish up with an overall infection severity for each lung lobe in the scan in under 5 min. While doing the exact same task takes more than an hour for a human expert.

As for the limitations of this research, the biggest one involves data collecting. Labeling for lobe and infection segmentation is a rather time-consuming process. The fact that several licenses from multiple official bodies are required to collect the data in the first place adds to the time consumption of the data collecting process.

For this research, since the necessary grants were not provided at the time of data labeling, the labels were produced by the resident radiologists and only the External Test sets 1 and 2 were later labeled by senior radiologists. This resulted in us only being to evaluate the resident radiologists. Evaluating the senior radiologists with a thoroughness level matching the rest of the research required expert manpower beyond what we could manage to bring (at least 5 more senior radiologists).

Another limitation was the rather small size of level 4 and 5 infection severity samples in our datasets, which certainly hampered our framework's performance. In addition, a much larger dataset of normal scans is required to reduce the prediction error between classes 0 and 1.

A future improvement on this work might include involving more data from more centers with different devices and configurations in the models training. A study on the effect of different scan dosages is also beneficial. To address one of the other limitations of this work, data from patients with an age distribution that includes younger subjects will certainly improve the framework's comprehensiveness. Improving the error margin on the manual process of lobe and infection segmentation labeling and infection severity estimation would also help the framework's performance.

In addition, we observed instances of the model incorrectly recognizing pulmonary vessels around the umbilical cord area as infection (which would get worsen with noisy data). To address this, a more complex deep learning model trained on data that also has vessel segmentation is likely needed. The pièce de résistance would be a model that could predict the infection severity score of a chest CT scan, without requiring to perform lobe or infection segmentation.

To conclude, we developed a framework for infection severity prediction in lung lobes by involving several datasets, collected and public, and by utilizing multiple machine learning and deep learning models, in order to serve as a prognosis tool for the experts. Finally, we comprehensively evaluated our framework and compared its performance to experts to determine its benefit in helping the treatment process of patients.

Data availability statement

The data analyzed in this study is subject to the following licenses/restrictions: The CT dataset has been collected in Ghiassi Hospital, Tehran, Iran. It cannot be shared publicly because of ethical restrictions and sensitive human study participant data. The anonymized (non-personally identifiable) data is, however, available from the Institutional Data Access through the Ethics Committee of Ghiassi Hospital for the researchers who meet the criteria to access the confidential data (contact the corresponding author at bW5hemVtemFkZWgmI3gwMDA0MDt0dW1zLmFjLmly). Requests to access these datasets should be directed to M-RN-Z bW5hemVtemFkZWgmI3gwMDA0MDt0dW1zLmFjLmly.

Ethics statement

The studies involving human participants were reviewed and approved by Tehran University of Medical Sciences, IR.TUMS.IKHC.REC.1399.447. The patients/participants provided their written informed consent to participate in this study.

Author contributions

MY: main contributor, design and implementation of machine learning algorithms, and writing the manuscript. MH: design and implementation of machine learning algorithms and revising the manuscript. MZ: data acquisition and manuscript preparation. FS, AY, and AM: data labeling and processing. RJ: radiologist, imaging reporting, and supervision of data labeling. PE: implementation of machine learning algorithms and revising the manuscript. M-RN-Z: definition of project, providing the fund, design of machine learning algorithms, and revising the manuscript. All authors contributed to the article and approved the submitted version.

Funding

This project was partially funded by the Tehran University of Medical Sciences (TUMS), the International Campus (IC) Research Affairs with Research Code: 99-2-163-49030. The funding was merely for the data processing and modeling and did not include any open access publication fees.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

The reviewer SK declared a shared affiliation with the author MY to the handling editor at the time of review.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Supplementary material

The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fmed.2022.940960/full#supplementary-material

Supplementary Figure 1. Left: Age distribution of study subjects. Right: Sex distribution of study subjects.

Supplementary Figure 2. Number of normal and infected slices of the data used in the train and validation sets.

Supplementary Figure 3. Slice infection percentage in the train and validation sets.

Supplementary Figure 4. Sample count of different classes of infection severity for the 5 lung lobes and overall.

Supplementary Figure 5. Number of normal and infected slices of the data used in the test set.

Supplementary Figure 6. Slice infection percentage in the test sets.

Supplementary Figure 7. Sample count of different classes of infection severity in External Test sets 1 and 2.

References

1. Roosa K, Lee Y, Luo R, Kirpich A, Rothenberg R, Hyman JM, et al. Real-time forecasts of the COVID-19 epidemic in China from February 5th to February 24th, 2020. Infect Dis Model. (2020) 5:256–63. doi: 10.1016/j.idm.2020.02.002

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Yan L, Zhang HT, Xiao Y, Wang M, Sun C, Liang J, et al. Prediction of criticality in patients with severe COVID-19 infection using three clinical features: a machine learning-based prognostic model with clinical data in Wuhan. MedRxiv. (2020) 27:2020. doi: 10.1101/2020.02.27.20028027

CrossRef Full Text | Google Scholar

3. Organization WH. Novel Coronavirus (2019-nCoV). Situation Report-22, Data as reported by 11 February 2020 (2020).

Google Scholar

4. Gorbalenya AE, Baker SC, Baric R, Groot RJd, Drosten C, Gulyaeva AA, et al. Severe acute respiratory syndrome-related coronavirus: The species and its viruses-a statement of the Coronavirus Study Group. MedRxiv. (2020). doi: 10.1101/2020.02.07.937862

CrossRef Full Text | Google Scholar

5. cdc.gov. Symptoms of Coronavirus. Centers for Disease Control and Prevention (2020). Available online at: https://www.cdc.gov/coronavirus/2019-ncov/symptoms-testing/symptoms

Google Scholar

6. WHO. Q&A on coronaviruses (COVID-19). Geneva: World Health Organization (2020). Available online at: https://www.who.int/covid-19 (accessed June 3, 2021).

Google Scholar

7. Ai T, Yang Z, Hou H, Zhan C, Chen C, Lv W, et al. Correlation of chest CT and RT-PCR testing for coronavirus disease 2019 (COVID-19) in China: a report of 1014 cases. Radiology. (2020) 296:E32–E40. doi: 10.1148/radiol.2020200642

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Xie X, Zhong Z, Zhao W, Zheng C, Wang F, Liu J. Chest CT for typical coronavirus disease 2019 (COVID-19) pneumonia: relationship to negative RT-PCR testing. Radiology. (2020) 296:E41–5. doi: 10.1148/radiol.2020200343

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Yousefzadeh M, Esfahanian P, Movahed SMS, Gorgin S, Rahmati D, Abedini A, et al. ai-corona: radiologist-assistant deep learning framework for COVID-19 diagnosis in chest CT scans. PLoS ONE. (2021) 16:e0250952. doi: 10.1371/journal.pone.0250952

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Deng J, Dong W, Socher R, Li LJ, Li K, Fei-Fei L. Imagenet: a large-scale hierarchical image database. In: 2009 IEEE Conference on Computer Vision and Pattern Recognition. Miami, FL: IEEE (2009). p. 248–55.

Google Scholar

11. Krizhevsky A, Sutskever I, Hinton GE. Imagenet classification with deep convolutional neural networks. In: Advance Neural Information Processing Systems. Lake Tahoe, CA (2012).

Google Scholar

12. LeCun Y, Bengio Y, Hinton G. Deep learning. Nature. (2015) 521:436–44. doi: 10.1038/nature14539

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Qin ZZ, Sander MS, Rai B, Titahong CN, Sudrungrot S, Laah SN, et al. Using artificial intelligence to read chest radiographs for tuberculosis detection: a multi-site evaluation of the diagnostic accuracy of three deep learning systems. Sci Rep. (2019) 9:1–10. doi: 10.1038/s41598-019-51503-3

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Ardila D, Kiraly AP, Bharadwaj S, Choi B, Reicher JJ, Peng L, et al. End-to-end lung cancer screening with three-dimensional deep learning on low-dose chest computed tomography. Nat Med. (2019) 25:954–61. doi: 10.1038/s41591-019-0447-x

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Nash M, Kadavigere R, Andrade J, Sukumar CA, Chawla K, Shenoy VP, et al. Deep learning, computer-aided radiography reading for tuberculosis: a diagnostic accuracy study from a tertiary hospital in India. Sci Rep. (2020) 10:1–10. doi: 10.1038/s41598-019-56589-3

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Liao F, Liang M, Li Z, Hu X, Song S. Evaluate the malignancy of pulmonary nodules using the 3-d deep leaky noisy-or network. IEEE Trans Neural Netw Learn Syst. (2019) 30:3484–95. doi: 10.1109/TNNLS.2019.2892409

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Mei X, Lee HC, Diao Ky, Huang M, Lin B, Liu C, et al. Artificial intelligence-enabled rapid diagnosis of patients with COVID-19. Nat Med. (2020) 26:1224–8. doi: 10.1038/s41591-020-0931-3

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Zhu W, Liu C, Fan W, Xie X. Deeplung: deep 3d dual path nets for automated pulmonary nodule detection and classification. In: 2018 IEEE Winter Conference on Applications of Computer Vision (WACV). Lake Tahoe, NV: IEEE (2018). p. 673–81.

Google Scholar

19. Xie H, Yang D, Sun N, Chen Z, Zhang Y. Automated pulmonary nodule detection in CT images using deep convolutional neural networks. Pattern Recognit. (2019) 85:109–19. doi: 10.1016/j.patcog.2018.07.031

CrossRef Full Text | Google Scholar

20. Vaishya R, Javaid M, Khan IH, Haleem A. Artificial Intelligence (AI) applications for COVID-19 pandemic. Diabet Metabol Syndrome. (2020) 14:337–9. doi: 10.1016/j.dsx.2020.04.012

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Shi F, Wang J, Shi J, Wu Z, Wang Q, Tang Z, et al. Review of artificial intelligence techniques in imaging data acquisition, segmentation, and diagnosis for COVID-19. IEEE Rev Biomed Eng. (2020) 14:4–15. doi: 10.1109/RBME.2020.2987975

PubMed Abstract | CrossRef Full Text | Google Scholar

22. McCall B. COVID-19 and artificial intelligence: protecting health-care workers and curbing the spread. Lancet Digit Health. (2020) 2:e166–7. doi: 10.1016/S2589-7500(20)30054-6

PubMed Abstract | CrossRef Full Text | Google Scholar

23. Huang L, Han R, Ai T, Yu P, Kang H, Tao Q, et al. Serial quantitative chest CT assessment of COVID-19: a deep learning approach. Radiology. (2020) 2:e200075. doi: 10.1148/ryct.2020200075

PubMed Abstract | CrossRef Full Text | Google Scholar

24. Wynants L, Van Calster B, Bonten MM, Collins GS, Debray TP, De Vos M, et al. Systematic review and critical appraisal of prediction models for diagnosis and prognosis of COVID-19 infection. MedRxiv. (2020) doi: 10.1101/2020.03.24.20041020

CrossRef Full Text | Google Scholar

25. Wynants L, Van Calster B, Collins GS, Riley RD, Heinze G, Schuit E, et al. Prediction models for diagnosis and prognosis of covid-19: systematic review and critical appraisal. BMJ. (2020) 369:m1328. doi: 10.1136/bmj.m1328

PubMed Abstract | CrossRef Full Text | Google Scholar

26. Muhammad L, Islam M, Usman SS, Ayon SI, et al. Predictive data mining models for novel coronavirus (COVID-19) infected patients' recovery. SN Comput. Sci. (2020) 1:1–7. doi: 10.1007/s42979-020-00216-w

PubMed Abstract | CrossRef Full Text | Google Scholar

27. Albahri AS, Hamid RA, Al-qays Z, Zaidan A, Zaidan B, Albahri AO, et al. Role of biological data mining and machine learning techniques in detecting and diagnosing the novel coronavirus (COVID-19): a systematic review. J Med Syst. (2020) 44:1–11. doi: 10.1007/s10916-020-01582-x

PubMed Abstract | CrossRef Full Text | Google Scholar

28. Latif S, Usman M, Manzoor S, Iqbal W, Qadir J, Tyson G, et al. Leveraging data science to combat COVID-19: a comprehensive review. IEEE Trans Artif Intell. (2020) 1:85–103. doi: 10.1109/TAI.2020.3020521

CrossRef Full Text | Google Scholar

29. Hesamian MH, Jia W, He X, Kennedy P. Deep learning techniques for medical image segmentation: achievements and challenges. J Digit Imaging. (2019) 32:582–96. doi: 10.1007/s10278-019-00227-x

PubMed Abstract | CrossRef Full Text | Google Scholar

30. Ciompi F, Chung K, Van Riel SJ, Setio AAA, Gerke PK, Jacobs C, et al. Towards automatic pulmonary nodule management in lung cancer screening with deep learning. Sci Rep. (2017) 7:1–11. doi: 10.1038/srep46479

PubMed Abstract | CrossRef Full Text | Google Scholar

31. Jiang H, Ma H, Qian W, Gao M, Li Y. An automatic detection system of lung nodule based on multigroup patch-based deep learning network. IEEE J Biomed Health Inform. (2017) 22:1227–37. doi: 10.1109/JBHI.2017.2725903

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Ma J, Wang Y, An X, Ge C, Yu Z, Chen J, et al. Toward data-efficient learning: a benchmark for COVID-19 CT lung and infection segmentation. Med Phys. (2021) 48:1197–210. doi: 10.1002/mp.14676

PubMed Abstract | CrossRef Full Text | Google Scholar

33. Li Z, Zhong Z, Li Y, Zhang T, Gao L, Jin D, et al. From community-acquired pneumonia to COVID-19: a deep learning-based method for quantitative analysis of COVID-19 on thick-section CT scans. Eur Radiol. (2020) 30:6828–37. doi: 10.1007/s00330-020-07042-x

PubMed Abstract | CrossRef Full Text | Google Scholar

34. Abdel-Basset M, Chang V, Hawash H, Chakrabortty RK, Ryan M. FSS-2019-nCov: a deep learning architecture for semi-supervised few-shot segmentation of COVID-19 infection. Knowl Based Syst. (2021) 212:106647. doi: 10.1016/j.knosys.2020.106647

PubMed Abstract | CrossRef Full Text | Google Scholar

35. Voulodimos A, Protopapadakis E, Katsamenis I, Doulamis A, Doulamis N. Deep learning models for COVID-19 infected area segmentation in CT images. In: The 14th PErvasive Technologies Related to Assistive Environments Conference. New York, NY (2021). p. 404–11.

Google Scholar

36. Xie W, Jacobs C, Charbonnier JP, Van Ginneken B. Relational modeling for robust and efficient pulmonary lobe segmentation in CT scans. IEEE Trans Med Imaging. (2020) 39:2664–75. doi: 10.1109/TMI.2020.2995108

PubMed Abstract | CrossRef Full Text | Google Scholar

37. Tang H, Zhang C, Xie X. Automatic pulmonary lobe segmentation using deep learning. In: 2019 IEEE 16th International Symposium on Biomedical Imaging (ISBI 2019). Venice: IEEE (2019). p. 1225–8. doi: 10.1109/ISBI.2019.8759468

CrossRef Full Text | Google Scholar

38. Ferreira FT, Sousa P, Galdran A, Sousa MR, Campilho A. End-to-end supervised lung lobe segmentation. In: 2018 International Joint Conference on Neural Networks (IJCNN). Rio de Janeiro: IEEE (2018). p. 1–8.

Google Scholar

39. Park J, Yun J, Kim N, Park B, Cho Y, Park HJ, et al. Fully automated lung lobe segmentation in volumetric chest CT with 3D U-Net: validation with intra-and extra-datasets. J Digit Imaging. (2020) 33:221–30. doi: 10.1007/s10278-019-00223-1

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Hofmanninger J, Prayer F, Pan J, Röhrich S, Prosch H, Langs G. Automatic lung segmentation in routine imaging is primarily a data diversity problem, not a methodology problem. Eur Radiol Exp. (2020) 4:1–13. doi: 10.1186/s41747-020-00173-2

PubMed Abstract | CrossRef Full Text | Google Scholar

41. Chen J, Wu L, Zhang J, Zhang L, Gong D, Zhao Y, et al. Deep learning-based model for detecting 2019 novel coronavirus pneumonia on high-resolution computed tomography. Sci Rep. (2020) 10:1–11. doi: 10.1038/s41598-020-76282-0

PubMed Abstract | CrossRef Full Text | Google Scholar

42. Tang Z, Zhao W, Xie X, Zhong Z, Shi F, Liu J, et al. Severity assessment of coronavirus disease 2019 (COVID-19) using quantitative features from chest CT images. arXiv preprint arXiv:200311988. (2020). doi: 10.48550/arXiv.2003.11988

PubMed Abstract | CrossRef Full Text | Google Scholar

43. Abbas A, Abdelsamea MM, Gaber MM. Classification of COVID-19 in chest X-ray images using DeTraC deep convolutional neural network. Appl Intell. (2021) 51:854–64. doi: 10.1007/s10489-020-01829-7

PubMed Abstract | CrossRef Full Text | Google Scholar

44. Oulefki A, Agaian S, Trongtirakul T, Laouar AK. Automatic COVID-19 lung infected region segmentation and measurement using CT-scans images. Pattern Recognit. (2021) 114:107747. doi: 10.1016/j.patcog.2020.107747

PubMed Abstract | CrossRef Full Text | Google Scholar

45. Iwendi C, Bashir AK, Peshkar A, Sujatha R, Chatterjee JM, Pasupuleti S, et al. COVID-19 patient health prediction using boosted random forest algorithm. Front Public Health. (2020) 8:357. doi: 10.3389/fpubh.2020.00357

PubMed Abstract | CrossRef Full Text | Google Scholar

46. Wu W, Shi Y, Li X, Zhou Y, Du P, Lv S, et al. Deep learning to estimate the physical proportion of infected region of lung for COVID-19 pneumonia with ct image set. arXiv preprint arXiv:200605018. (2020). doi: 10.48550/arXiv.2006.05018

CrossRef Full Text | Google Scholar

47. Li M, Arun N, Gidwani M, Chang K, Deng F, Little B, et al. Automated assessment of COVID-19 pulmonary disease severity on chest radiographs using convolutional Siamese neural networks. medRxiv. (2020) doi: 10.1101/2020.05.20.20108159

PubMed Abstract | CrossRef Full Text | Google Scholar

48. Morozov S, Andreychenko A, Pavlov N, Vladzymyrskyy A, Ledikhova N, Gombolevskiy V, et al. Mosmeddata: chest ct scans with covid-19 related findings dataset. arXiv preprint arXiv:200506465. (2020) doi: 10.1101/2020.05.20.20100362

CrossRef Full Text | Google Scholar

49. Jenssen HB. (2021). Available online at: https://www.eibir.org/covid-19-imaging-datasets/#segment2 (accessed June 3, 2021).

Google Scholar

50. Radiology-Assistant. (2021). Available online at: https://radiologyassistant.nl/chest/covid-19/covid19-imaging-findings (accessed June 3, 2021).

Google Scholar

51. Zoom. (2022). Available online at: https://docs.scipy.org/doc/scipy/reference/generated/scipy.ndimage.zoom.html (accessed June 3, 2021).

Google Scholar

52. Serra J. Image analysis and mathematical morphol-ogy. (1982).

Google Scholar

53. Ronneberger O, Fischer P, Brox T. U-net: convolutional networks for biomedical image segmentation. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Munich: Springer (2015). p. 234–41. doi: 10.1007/978-3-319-24574-4_28

CrossRef Full Text | Google Scholar

54. Srivastava RK, Greff K, Schmidhuber J. Training very deep networks. In: Advances in Neural Information Processing Systems 28, Vol. 2. Palais des Congres de Montreal (2015). p. 2377–2385.

Google Scholar

55. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Las Vegas, NV: IEEE (2016). p. 770–8.

PubMed Abstract | Google Scholar

56. Yu F, Koltun V. Multi-scale context aggregation by dilated convolutions. arXiv preprint arXiv:151107122. (2015). doi: 10.48550/arXiv.1511.07122

CrossRef Full Text | Google Scholar

57. Chen LC, Papandreou G, Kokkinos I, Murphy K, Yuille AL. Deeplab: semantic image segmentation with deep convolutional nets, atrous convolution, and fully connected crfs. IEEE Trans Pattern Anal Mach Intell. (2017) 40:834–48. doi: 10.1109/TPAMI.2017.2699184

PubMed Abstract | CrossRef Full Text | Google Scholar

58. Lin TY, Dollár P, Girshick R, He K, Hariharan B, Belongie S. Feature pyramid networks for object detection. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Honolulu, HI: IEEE (2017). p. 2117–25.

Google Scholar

59. Zhao H, Shi J, Qi X, Wang X, Jia J. Pyramid scene parsing network. In: Proceedings of the IEEE conference on computer vision and pattern recognition. Honolulu, HI: IEEE (2017). p. 2881–90.

Google Scholar

60. Chaurasia A, Culurciello E. Linknet: exploiting encoder representations for efficient semantic segmentation. In: 2017 IEEE Visual Communications and Image Processing (VCIP). St. Petersburg, FL: IEEE (2017). p. 1–4.

Google Scholar

61. Zhou Z, Rahman Siddiquee MM, Tajbakhsh N, Liang J. Unet++: a nested u-net architecture for medical image segmentation. In: Deep learning in medical image analysis and multimodal learning for clinical decision support. Granada: Springer (2018). p. 3–11.

PubMed Abstract | Google Scholar

62. Tan M, Le Q. Efficientnet: rethinking model scaling for convolutional neural networks. In: International Conference on Machine Learning. Long Beach, CA: PMLR (2019). p. 6105–14.

Google Scholar

63. Hu J, Shen L, Sun G. Squeeze-and-excitation networks. In: Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition. Salt Lake City, UT: IEEE (2018). p. 7132–41.

Google Scholar

64. Mason D. SU-E-T-33: pydicom: an open source DICOM library. Med Phys. (2011) 38:3493–3. doi: 10.1118/1.3611983

CrossRef Full Text | Google Scholar

65. Yousefzadeh M, Zolghadri M, Hasanpour M, Salimi F, Jafari R, Bozorg MV, et al. Statistical analysis of COVID-19 infection severity in lung lobes from chest CT. Inform Med Unlocked. (2022) 30:100935. doi: 10.1016/j.imu.2022.100935

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: COVID-19, deep learning, infection segmentation, lobe segmentation, CT scan, severity score

Citation: Yousefzadeh M, Hasanpour M, Zolghadri M, Salimi F, Yektaeian Vaziri A, Mahmoudi Aqeel Abadi A, Jafari R, Esfahanian P and Nazem-Zadeh M-R (2022) Deep learning framework for prediction of infection severity of COVID-19. Front. Med. 9:940960. doi: 10.3389/fmed.2022.940960

Received: 10 May 2022; Accepted: 15 July 2022;
Published: 17 August 2022.

Edited by:

Monica Catarina Botelho, Instituto Nacional de Saúde Doutor Ricardo Jorge (INSA), Portugal

Reviewed by:

Saeed Reza Kheradpisheh, Shahid Beheshti University, Iran
Oscar J. Pellicer-Valero, University of Valencia, Spain
Mohanad Alkhodari, Khalifa University, United Arab Emirates

Copyright © 2022 Yousefzadeh, Hasanpour, Zolghadri, Salimi, Yektaeian Vaziri, Mahmoudi Aqeel Abadi, Jafari, Esfahanian and Nazem-Zadeh. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Mohammad-Reza Nazem-Zadeh, bW5hemVtemFkZWgmI3gwMDA0MDt0dW1zLmFjLmly

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.