Skip to main content

ORIGINAL RESEARCH article

Front. Med., 10 March 2022
Sec. Translational Medicine

COVID-Net CT-2: Enhanced Deep Neural Networks for Detection of COVID-19 From Chest CT Images Through Bigger, More Diverse Learning

\nHayden Gunraj
Hayden Gunraj1*Ali Sabri,Ali Sabri2,3David Koff,David Koff2,4Alexander Wong,,Alexander Wong1,5,6
  • 1Vision and Image Processing Lab, University of Waterloo, Waterloo, ON, Canada
  • 2Department of Radiology, McMaster University, Hamilton, ON, Canada
  • 3Niagara Health System, St. Catharines, ON, Canada
  • 4Hamilton Health Sciences, Hamilton, ON, Canada
  • 5Waterloo Artificial Intelligence Institute, University of Waterloo, Waterloo, ON, Canada
  • 6DarwinAI Corp., Waterloo, ON, Canada

The COVID-19 pandemic continues to rage on, with multiple waves causing substantial harm to health and economies around the world. Motivated by the use of computed tomography (CT) imaging at clinical institutes around the world as an effective complementary screening method to RT-PCR testing, we introduced COVID-Net CT, a deep neural network tailored for detection of COVID-19 cases from chest CT images, along with a large curated benchmark dataset comprising 1,489 patient cases as part of the open-source COVID-Net initiative. However, one potential limiting factor is restricted data quantity and diversity given the single nation patient cohort used in the study. To address this limitation, in this study we introduce enhanced deep neural networks for COVID-19 detection from chest CT images which are trained using a large, diverse, multinational patient cohort. We accomplish this through the introduction of two new CT benchmark datasets, the largest of which comprises a multinational cohort of 4,501 patients from at least 16 countries. To the best of our knowledge, this represents the largest, most diverse multinational cohort for COVID-19 CT images in open-access form. Additionally, we introduce a novel lightweight neural network architecture called COVID-Net CT S, which is significantly smaller and faster than the previously introduced COVID-Net CT architecture. We leverage explainability to investigate the decision-making behavior of the trained models and ensure that decisions are based on relevant indicators, with the results for select cases reviewed and reported on by two board-certified radiologists with over 10 and 30 years of experience, respectively. The best-performing deep neural network in this study achieved accuracy, COVID-19 sensitivity, positive predictive value, specificity, and negative predictive value of 99.0%/99.1%/98.0%/99.4%/99.7%, respectively. Moreover, explainability-driven performance validation shows consistency with radiologist interpretation by leveraging correct, clinically relevant critical factors. The results are promising and suggest the strong potential of deep neural networks as an effective tool for computer-aided COVID-19 assessment. While not a production-ready solution, we hope the open-source, open-access release of COVID-Net CT-2 and the associated benchmark datasets will continue to enable researchers, clinicians, and citizen data scientists alike to build upon them.

1. Introduction

The coronavirus disease 2019 (COVID-19) pandemic, caused by severe acute respiratory syndrome coronavirus 2 (SARS-CoV-2), continues to rage on, with multiple waves causing substantial harm to health and economies around the world. Real-time reverse transcription polymerase chain reaction (RT-PCR) testing remains the primary screening tool for COVID-19, where SARS-CoV-2 ribonucleic acid (RNA) is detected within an upper respiratory tract sputum sample (1). However, despite being highly specific, the sensitivity of RT-PCR can be relatively low (2, 3) and can vary greatly depending on the time since symptom onset as well as sampling method (35).

Clinical institutes around the world have explored the use of computed tomography (CT) imaging as an effective, complementary screening tool alongside RT-PCR (2, 5, 6). In particular, studies have shown CT to have great utility in detecting COVID-19 infections during routine CT examinations for non-COVID-19 related reasons such as elective surgical procedure monitoring and neurological examinations (7, 8). Other scenarios where CT imaging has been leveraged include cases where patients have worsening respiratory complications, as well as cases where patients with negative RT-PCR test results are suspected to be COVID-19 positive due to other factors. Early studies have shown that a number of potential indicators for COVID-19 infections may be present in chest CT images (2, 5, 6, 912), but may also be present in non-COVID-19 infections. This can lead to challenges for radiologists in distinguishing COVID-19 infections from non-COVID-19 infections using chest CT (13, 14).

Inspired by the potential of CT imaging as a complementary screening method and the challenges of CT interpretation for COVID-19 screening, we previously introduced COVID-Net CT (15), a convolutional neural network (CNN) tailored for detection of COVID-19 cases from chest CT images. We further introduced COVIDx CT, a large curated benchmark dataset comprising chest CT scans from a cohort of 1,489 patients derived from a collection by the China National Center for Bioinformation (CNCB) (16). Both COVID-Net CT and COVIDx CT were made publicly available as part of the COVID-Net (17, 18) initiative, an open-source initiative1 aimed at accelerating advancement and adoption of deep learning in the fight against the COVID-19 pandemic. While COVID-Net CT was able to achieve state-of-the-art COVID-19 detection performance, one potential limiting factor is the restricted quantity and diversity of CT imaging data used to learn the deep neural network given the entirely Chinese patient cohort used in the study. As such, a greater quantity and diversity in the patient cohort has the potential to improve generalization, particularly when COVID-Net CT is leveraged in different clinical settings around the world.

Motivated by the success and widespread adoption of COVID-Net CT and COVIDx CT, as well as their potential data quantity and diversity limitations, in this study we introduce COVID-Net CT-2, enhanced CNNs for COVID-19 detection from chest CT images which are trained using a large, diverse, multinational patient cohort. More specifically, we accomplish this through the introduction of two new CT benchmark datasets (COVIDx CT-2A and COVIDx CT-2B), the largest of which comprises a multinational cohort of 4,501 patients from at least 16 countries. To the best of the authors' knowledge, these benchmark datasets represent the largest, most diverse multinational cohorts for COVID-19 CT images available in open access form. Additionally, we introduce a novel lightweight neural network architecture called COVID-Net CT S, which is significantly smaller and faster than the previously introduced COVID-Net CT architecture and achieves an improved trade-off between performance and efficiency. Finally, we leverage explainability to investigate the decision-making behavior of COVID-Net CT-2 models to ensure decisions are based on relevant visual indicators in CT images, with the results for select patient cases being reviewed and reported on by two board-certified radiologists with 10 and 30 years of experience, respectively. The COVID-Net CT-2 networks and corresponding COVIDx CT-2 datasets are publicly available as part of the COVID-Net initiative (17, 18). While not a production-ready solution, we hope the open-source, open-access release of the COVID-Net CT-2 networks and the corresponding COVIDx CT-2 benchmark datasets will enable researchers, clinicians, and citizen data scientists alike to build upon them.

2. Materials and Methods

2.1. COVIDx CT-2 Benchmark Dataset

The original COVIDx CT benchmark dataset consists of chest CT scans collected by the China National Center for Bioinformation (CNCB) (16) which were carefully processed and selected to form a cohort of 1,489 patient cases. While COVIDx CT is significantly larger than many CT datasets for COVID-19 detection in literature, a potential limitation with leveraging COVIDx CT for training neural networks is the limited diversity in terms of patient demographics. More specifically, the cohort of patients used in COVIDx CT are collected in different provinces of China, and as such the characteristics of COVID-19 infection as observed in the chest CT images may not generalize to patients around the world outside of China. Therefore, increasing the quantity and diversity of the patient cohort in constructing new benchmark datasets could result in more diverse, well-rounded training of neural networks. In doing so, improved generalization and applicability for use in different clinical environments around the world can be achieved.

In this study, we carefully processed and curated CT images from several patient cohorts from around the world which were collected using a variety of CT equipment types, protocols, and levels of validation. By unifying CT imaging data from several cohorts from around the world, we created two diverse, large-scale benchmark datasets:

COVIDx CT-2A: This benchmark dataset comprises 194,922 CT images from a multinational cohort of 3,745 patients between 0 and 93 years old (median age of 51) with strongly clinically-verified findings. The multinational cohort consists of patient cases collected by the following organizations and initiatives from around the world: (1) China National Center for Bioinformation (CNCB) (16) (China), (2) National Institutes of Health Intramural Targeted Anti-COVID-19 (ITAC) Program (hosted by TCIA (19), countries unknown), (3) Negin Radiology Medical Center (20) (Iran), (4) Union Hospital and Liyuan Hospital of Huazhong University of Science and Technology (21) (China), (5) COVID-19 CT Lung and Infection Segmentation initiative, annotated and verified by Nanjing Drum Tower Hospital (22) (Iran, Italy, Turkey, Ukraine, Belgium, some countries unknown), (6) Lung Image Database Consortium (LIDC) and Image Database Resource Initiative (IDRI) (23) (USA), and (7) Radiopaedia collection (24) (Iran, Italy, Australia, Afghanistan, Scotland, Lebanon, England, Algeria, Peru, Azerbaijan, some countries unknown).

COVIDx CT-2B: This benchmark dataset comprises 201,103 CT images from a multinational cohort of 4,501 patients between 0 and 93 years old (median age of 51) with a mix of strongly verified findings and weakly verified findings. The patient cohort in COVIDx CT-2B consists of the multinational patient cohort we leveraged to construct COVIDx CT-2A, which have strongly clinically-verified findings, with additional patient cases with weakly verified findings collected by the Research and Practical Clinical Center of Diagnostics and Telemedicine Technologies, Department of Health Care of Moscow (MosMed) (25) (Russia). Notably, these additional cases are only included in the training dataset, and as such the validation and test datasets are identical to those of COVIDx CT-2A.

In both COVIDx CT-2 benchmark datasets, the findings for the chest CT volumes correspond to three different infection types: (1) novel coronavirus pneumonia due to SARS-CoV-2 viral infection (NCP), (2) common pneumonia (CP), and (3) normal controls. The image and patient distributions for the three infection types across training, validation, and test partitions are shown in Tables 1, 2 for COVIDx CT-2A and COVIDx CT-2B, respectively. Note that the data is partitioned at the patient level, and as such each patient appears in a single partition. For CT volumes labeled as NCP or CP, slices containing abnormalities were identified and assigned the same labels as the CT volumes. Notably, patient age was not available for all cases, and as such the age ranges and median ages reported above are based on patient cases for which age was available. The given range alludes to the inclusion of pediatric images in the COVIDx CT-2 datasets, which possess significant visual differences from adult images. However, we argue that leveraging COVID-19 pediatric cases may allow for the trained models to be more robust, and moreover only 26 of the included patients are under the age of 18. For images which were originally in Hounsfield units (HU), a standard lung window centered at −600 HU with a width of 1,500 HU was used to map the images to unsigned 8-bit integer range (i.e., 0–255).

TABLE 1
www.frontiersin.org

Table 1. Distribution of chest CT slices and patient cases (in parentheses) by data partition and infection type in the COVIDx CT-2A dataset.

TABLE 2
www.frontiersin.org

Table 2. Distribution of chest CT slices and patient cases (in parentheses) by data partition and infection type in the COVIDx CT-2B dataset.

The rationale for creating two different COVIDx CT-2 benchmark datasets stems from the availability of weakly verified findings (i.e., findings not based on RT-PCR test results or final radiology reports), which can be useful for further increasing the quantity and diversity of patient cases that a neural network can be exposed to and can be of great interest for researchers, clinicians, and citizen scientists to explore and build upon while being made aware of the fact some of the CT scans do not have strongly verified findings available. Select patient cases from the benchmark datasets were reviewed and reported on by two board-certified radiologists with 10 and 30 years of experience, respectively. Both COVIDx CT-2A and COVIDx CT-2B benchmark datasets are publicly available2 as part of the COVID-Net initiative, with example CT images from each type of infection shown in Figure 1.

FIGURE 1
www.frontiersin.org

Figure 1. Example CT images from the COVIDx CT-2 benchmark datasets from each type of infection: (A) novel coronavirus pneumonia due to SARS-CoV-2 infection (NCP), (B) common pneumonia (CP), and (C) normal controls.

2.2. COVID-Net CT-2 Construction and Learning

By leveraging the COVIDx CT-2 benchmark datasets introduced in the previous section, we train a variety of CNNs in a way that better generalizes to a wide range of clinical scenarios. More specifically, six COVID-Net CT-2 deep neural networks are built based on six different deep CNN architecture designs: SqueezeNet (26), MobileNetV2 (27), EfficientNet-B0 (28), NASNet-A-Mobile (29), COVID-Net CT (15) (denoted COVID-Net CT L in this work), and a novel lightweight architecture called COVID-Net CT S. COVID-Net CT S shares the same macroarchitecture design as COVID-Net CT L, but with significantly more efficient microarchitecture designs. This greatly improved efficiency was achieved by leveraging machine-driven design exploration via generative synthesis (30), with the COVID-Net CT L architecture utilized as the initial design prototype. The COVID-Net CT S architecture is shown in Figure 2, and is made publicly available3.

FIGURE 2
www.frontiersin.org

Figure 2. COVID-Net CT-2 S architecture design and COVIDx CT-2 benchmark. We leverage the COVID-Net CT network architecture (15) as the basis of the COVID-Net CT S network, which was discovered automatically via machine-driven design exploration.

The various COVID-Net CT-2 CNNs were trained on the COVIDx CT-2A dataset. For comparison purposes, we also trained each COVID-Net CT-2 network on the original COVIDx CT dataset (15) (referred to from here on as COVIDx CT-1 for clarity). To differentiate between networks trained on different datasets, we indicate CT-1 or CT-2 in the network names (e.g., MobileNetV2 CT-1, MobileNetV2 CT-2, COVID-Net CT-1 L, COVID-Net CT-2 L, etc.). Given the unsigned 8-bit integer format of the datasets, CT slices were normalized to the range [0, 1] through division by 255. Optimization was performed via stochastic gradient descent with momentum (31) using cross-entropy loss with L2 regularization, and the following hyperparameters were selected: learning rate = 5e-4, momentum = 0.9, λL2 = 1e-4, batch size = 8. To further increase data diversity beyond what is provided by the large multinational cohort, we leveraged random data augmentation in the form of cropping box jitter (±7.5%), rotation (±20o), horizontal and vertical shear (±0.2), horizontal flip, and intensity shift (±15 gray levels) and scaling (±10%), with all augmentations being applied with 50% probability. Training was stopped once the validation accuracy plateaued (i.e., early stopping). During training, we leveraged the batch re-balancing strategy used in (15) to ensure a balanced distribution of each infection type at the batch level.

We evaluate the various COVID-Net CT-2 networks using the COVIDx CT-2 test dataset. To assess performance, we report test accuracy as well as class-wise sensitivity, positive predictive value (PPV), specificity, and negative predictive value (NPV). Additionally, to assess the efficiency-performance trade-offs of the models, we report NetScore (32) which takes into account test accuracy, architectural complexity, as well as computational complexity within a unified metric. Qualitative evaluation through explainability was also performed, and is discussed in the next section.

2.3. Explainability-Driven Performance Validation

As with COVID-Net CT (15), we utilize GSInquire (33) to conduct explainability-driven performance validation. Using GSInquire, we audit the trained models to better understand and verify their decision-making behavior when analyzing CT images to predict the condition of a patient. This form of performance validation is particularly important in a clinical context, as the decisions made about patients' conditions can affect the health of patients via treatment and care decisions made using a model's predictions. Therefore, examining the decision-making behavior through model auditing is key to ensuring that the right visual indicators in the CT scans (e.g., ground-glass opacities) are leveraged for making a prediction as opposed to irrelevant visual cues (e.g., synthetic padding, artifacts, patient tables, etc.). Furthermore, incorporating interpretability in the validation process also increases the level of trust that a clinician has in leveraging such models for clinical decision support by adding an extra degree of algorithmic transparency.

To facilitate explainability-driven performance validation via model auditing, GSInquire provides an explanation of how a model makes a decision based on input data by identifying a set of critical factors within the input data that impact the decision-making process of the neural network in a quantitatively significant way. This is accomplished by probing the model with an input signal (in this case, a CT image) as the targeted stimulus signal and observing the reactionary response signals throughout the model, thus enabling quantitative insights to be derived through the inquisition process. These quantitative insights are then transformed and projected into the same space as the input signal to produce an interpretation (in this case, a set of critical factors in the CT image that quantitatively led to the prediction of the patient's condition). These interpretations can be visualized spatially relative to the CT images for greater insights into the decision-level behavior of COVID-Net CT-2. Compared to other explainability methods (3438), this interesting nature of GSInquire in identifying quantitative impactful critical factors enables it to achieve explanations that better reflect the decision-making process of models (33). This makes it particularly suitable for quality assurance of models prior to clinical deployment to identify errors, biases, and anomalies that can lead to “right decisions for the wrong reasons.” The results obtained from GSInquire for select patient cases are further reviewed and reported on by two board-certified radiologists (AS and DK). The first radiologist (AS) has over 10 years of experience, while the second radiologist (DK) has over 30 years of radiology experience.

3. Results

3.1. Quantitative Analysis

To explore the efficacy of the COVID-Net CT-2 networks for COVID-19 detection from CT images, we conducted a quantitative evaluation of the trained networks using the COVIDx CT-2 test dataset. For comparison purposes, results for networks trained on the COVIDx CT-1 dataset (15) are also given. The training data used for each network is denoted by either CT-1 or CT-2 in the network name, as previously mentioned.

The test accuracy and NetScore (32) of each COVID-Net CT-2 network is shown in Table 3. It can be observed that all architectures achieve high test accuracy, with the best accuracy (99%) obtained using the MobileNetV2 and EfficientNet-B0 architectures trained on COVIDx CT-2A. Moreover, for all tested architectures, training on COVIDx CT-2A yields significant gains in test accuracy over training on COVIDx CT-1, with improvements ranging from +3.3 to +7.3% test accuracy. In terms of architectural and computational complexity, COVID-Net CT S possesses the fewest parameters (0.45 M) and floating-point operations (FLOPs, 1.94 G) while achieving an accuracy of 98.3%. The efficiency-performance trade-off of each model is assessed via its NetScore, of which COVID-Net CT-2 S' score of 83.3 is the highest and is significantly higher than the other tested networks. In contrast, the two COVID-Net CT-2 networks with the highest accuracies achieved significantly lower NetScores (74.1 and 70.7 for MobileNetV2 CT-2 and EfficientNet-B0 CT-2, respectively) owing to their significantly higher architectural and computational complexities. In resource-limited environments, balancing performance with efficiency is an important consideration.

TABLE 3
www.frontiersin.org

Table 3. Comparison of parameters, FLOPs, accuracy (image-level), and NetScore (32) for the tested networks on the COVIDx CT-2 benchmark test dataset.

The sensitivity and PPV for each infection type on the COVIDx CT-2 test dataset is shown in Table 4. Examining the differences in sensitivity between COVID-Net CT-2 networks trained on COVIDx CT-1 and COVIDx CT-2A, it can be observed that significant gains in sensitivity are achieved through training on COVIDx CT-2A (+4.9 to +23.9%), with the best sensitivity (99.1%) obtained using the EfficientNet-B0 architecture. Notably, for the COVID-Net CT L and S architectures, increased sensitivity comes at the cost of a slight reduction in COVID-19 PPV, whereas for the other four architectures sensitivity and PPV are both improved. From a clinical perspective, high sensitivity ensures few false negatives which would lead to missed patients with COVID-19 infections, whereas high PPV ensures few false positives which add an unnecessary burden on the healthcare system, which is already stressed due to the ongoing pandemic.

TABLE 4
www.frontiersin.org

Table 4. Sensitivity and positive predictive value (PPV) for each infection type at the image level on the COVIDx CT-2 benchmark test dataset.

The specificity and NPV for each infection type on the COVIDx CT-2 test dataset is shown in Table 5. Examining the differences in specificity between models trained on COVIDx CT-1 and COVIDx CT-2A, it can be observed that specificity does not change significantly in most cases, with a mix of improvements and reductions when switching from COVIDx CT-1 training to COVIDx CT-2 training. In contrast, when considering NPV, we observe consistent improvements ranging from no change (for COVID-Net CT L) to +6.4% (for EfficientNet-B0) when training on COVIDx CT-2A. The high specificity and NPV achieved by these models are important from a clinical perspective to ensure that COVID-19-negative predictions are indeed true negatives in the vast majority of cases, which facilitates rapid identification of COVID-19-negative patients.

TABLE 5
www.frontiersin.org

Table 5. Specificity and negative predictive value (NPV) for each infection type at the image level on the COVIDx CT-2 benchmark test dataset.

Given the diverse nature of the COVIDx CT-2 test dataset, these experimental results are particularly promising in terms of network generalization and applicability for use in different clinical environments. Additionally, the reduced performance observed when COVIDx CT-1 is used for training illustrates the value of larger, diverse training data for improving model performance in a variety of clinical scenarios.

3.2. Qualitative Analysis

To audit the decision-making behavior of COVID-Net CT-2 and ensure that it is leveraging relevant visual indicators when predicting the condition of a patient, we conducted explainability-driven performance validation using the COVIDx CT-2 benchmark test dataset, and the results obtained using COVID-Net CT-2 L for select patient cases are further reviewed and reported on by two board-certified radiologists. The critical factors identified by GSInquire for example chest CT images from the four COVID-19-positive cases that were reviewed are shown in Figure 3, and additional examples for COVID-Net CT-2 S are shown in Figure 4.

FIGURE 3
www.frontiersin.org

Figure 3. Example chest CT images from four COVID-19 cases reviewed and reported on by two board-certified radiologists, and the associated critical factors (highlighted in red) as identified by GSInquire (33) for COVID-Net CT-2 L. Based on the observations made by two expert radiologists, it was found that the critical factors leveraged by COVID-Net CT-2 L are consistent with radiologist interpretation.

FIGURE 4
www.frontiersin.org

Figure 4. Example chest CT images from four COVID-19 cases, and the associated critical factors (highlighted in red) as identified by GSInquire (33) for COVID-Net CT-2 S.

Overall, it can be observed from the GSInquire-generated visual explanations that both COVID-Net CT-2 L and COVID-Net CT-2 S are mainly utilizing visible lung abnormalities to distinguish between COVID-19-positive and COVID-19-negative cases. As such, this auditing process allows us to determine that COVID-Net CT-2 is indeed leveraging relevant visual indicators in the decision-making process as opposed to irrelevant visual indicators such as imaging artifacts, artificial padding, and patient tables. This performance validation process also reinforces the importance of utilizing explainability methods to confirm proper decision-making behavior in neural networks designed for clinical decision support.

3.3. Radiologist Findings

The expert radiologist findings and observations with regards to the critical factors identified by GSInquire for each of the four patient cases shown in Figure 3 are as follows. In all four cases, COVID-Net CT-2 L detected them to be novel coronavirus pneumonia due to SARS-CoV-2 viral infection, which was clinically confirmed.

Case 1 (top-left of Figure 3). It was observed by one of the radiologists that there are bilateral peripheral mixed ground-glass and patchy opacities with subpleural sparing, which is consistent with the identified critical factors leveraged by COVID-Net CT-2 L. The absence of large lymph nodes and effusion further helped the radiologist point to novel coronavirus pneumonia due to SARS-CoV-2 viral infection. The degree of severity is observed to be moderate to high. It was confirmed by the second radiologist that the identified critical factors leveraged by COVID-Net CT-2 L are correct areas of concern and represent areas of consolidation with a geographic distribution that is in favor of novel coronavirus pneumonia due to SARS-CoV-2 viral infection.

Case 2 (top-right of Figure 3). It was observed by one of the radiologists that there are bilateral peripherally-located ground-glass opacities with subpleural sparing, which is consistent with the identified critical factors leveraged by COVID-Net CT-2 L. As in Case 1, the absence of large lymph nodes and large effusion further helped the radiologist point to novel coronavirus pneumonia due to SARS-CoV-2 viral infection. The degree of severity is observed to be moderate to high. It was confirmed by the second radiologist that the identified critical factors leveraged by COVID-Net CT-2 L are correct areas of concern and represent areas of consolidation with a geographic distribution that is in favor of novel coronavirus pneumonia due to SARS-CoV-2 viral infection.

Case 3 (bottom-left of Figure 3). It was observed by one of the radiologists that there are peripheral bilateral patchy opacities, which is consistent with the identified critical factors leveraged by COVID-Net CT-2 L. Unlike the first two cases, there is small right effusion. However, as in Cases 1 and 2, the absence of large effusion further helped the radiologist point to novel coronavirus pneumonia due to SARS-CoV-2 viral infection. Considering that the opacities are at the base, a differential of atelectasis change was also provided. The degree of severity is observed to be moderate. It was confirmed by the second radiologist that the identified critical factors leveraged by COVID-Net CT-2 L are correct areas of concern and represent areas of consolidation.

Case 4 (bottom-right of Figure 3). It was observed by one of the radiologists that there are peripherally located asymmetrical bilateral patchy opacities, which is consistent with the identified critical factors leveraged by COVID-Net CT-2 L. As in Cases 1 and 2, the absence of lymph nodes and large effusion further helped the radiologist point to novel coronavirus pneumonia due to SARS-CoV-2 viral infection, but a differential of bacterial pneumonia was also provided considering the bronchovascular distribution of patchy opacities. In addition, there is no subpleural sparing. This highlights the potential difficulties in differentiating between novel coronavirus pneumonia and common pneumonia. It was confirmed by the second of the radiologists that the identified critical factors leveraged by COVID-Net CT-2 L are correct areas of concern and represent areas of consolidation with a geographic distribution that is in favor of novel coronavirus pneumonia due to SARS-CoV-2 viral infection.

Therefore, it can be observed that the explainability-driven validation process shows consistency between the decision-making process of COVID-Net CT-2 and radiologist interpretation, which suggests strong potential for computer-aided COVID-19 assessment within a clinical environment.

4. Discussion

In this work, we introduced COVID-Net CT-2, enhanced CNNs tailored for the purpose of COVID-19 detection from chest CT images. Two new CT benchmark datasets were introduced and used to facilitate the training of COVID-Net CT-2, and these datasets represent the largest, most diverse multinational cohorts of their kind available in open-access form, spanning cases from at least 16 countries. Experimental results show that the COVID-Net CT-2 networks are capable of not only achieving strong quantitative results, but also doing so in a manner that is consistent with radiologist interpretation via explainability-driven performance validation. The results are promising and suggest the strong potential of neural networks as an effective tool for computer-aided COVID-19 assessment.

Given the severity of the COVID-19 pandemic and the potential for deep learning to facilitate computer-assisted COVID-19 clinical decision support, a number of deep learning systems have been proposed in research literature for detecting SARS-CoV-2 infections using CT images (1416, 21, 3950), with a comprehensive review performed by Islam et al. (51). While some proposed deep learning systems focus on binary detection (SARS-CoV-2 positive vs. negative) (50), several proposed systems operate at a finer level of granularity by further identifying whether SARS-CoV-2-negative cases are normal control (16, 39, 47, 48), SARS-CoV-2 negative pneumonia [e.g., bacterial pneumonia, viral pneumonia, community-acquired pneumonia (CAP), etc.] (16, 3942, 48, 49), or non-pneumonia (41).

The majority of the proposed deep learning systems for COVID-19 detection from CT images rely on pre-existing network architectures that were designed for other image classification tasks. A large number of proposed systems additionally rely on segmentation of the lung region and/or lung lesions (14, 16, 3941, 44, 45, 47, 48). Some proposed systems also modify pre-existing network architectures; for example, Xu et al. (39) add location-attention modules to a ResNet-18 (52) backbone architecture, and Li et al. (41) and Bai et al. (40) add pooling operations to 2D architectures for volume-driven detection. Of the deep learning systems that proposed new neural network architectures, Shah et al. (43) proposed a 10-layer CNN architecture named CTnet-10, which ultimately showed lower detection performance than pre-existing architectures in literature. Zheng et al. (45) proposed a 3D CNN architecture named DeCovNet which is capable of volume-driven detection. Gunraj et al. (15) used machine-driven design to construct a tailored architecture, which was found to outperform three existing architectures. Hasan et al. (53) combined features obtained from a Q-deformed entropy model and custom CNN and used them to train a long short-term memory-based classifier. This feature fusion approach was found to outperform either of the feature sets alone. Finally, Javaheri et al. (54) use a U-Net-based architecture known as BCDU-Net to pre-process CT images before performing classification using a custom CNN architecture. The proposed BCDU-Net was trained using artificial data based on control cases and the 3D CNN classifier was then trained to classify CT volumes as COVID-19, community-acquired pneumonia, or control.

While the concept of leveraging deep learning for COVID-19 detection from CT images has been previously explored, even the largest studies in research literature have been limited in terms of quantity and/or diversity of patients, with many limited to single-nation cohorts (1416, 21, 39, 42, 54, 55). For example, the studies by Mei et al. (14), Gunraj et al. (15), Ning et al. (21), and Zhang et al. (16) were all limited to Chinese patient cohorts consisting of 905 patients, 1,489 patients, 1,521 patients, and 3,777 patients, respectively. Moreover, the studies by Ardakani et al. (42) and Javaheri et al. (54) leveraged patient cohorts from Iran including 108 and 335 patients, respectively. Multinational patient cohorts have been leveraged in several studies, but have typically been limited to few patients or few countries. For example, Hasan et al. (53) leveraged a multinational cohort (countries unknown) of 321 patients, Harmon et al. (50) leveraged a cohort of 2,617 patients across four countries, and Jin et al. (47) leveraged a cohort of 9,025 patients across at least three countries. To the best of the authors' knowledge, the multinational patient cohort introduced in this study represents the most diverse multinational patient cohort at 4,501 patients across at least 16 countries, and is the largest available in open-access form. By building the COVID-Net CT-2 deep neural networks using a large multinational patient cohort, we can better study the generalization capabilities and applicability of deep learning for computer-assisted assessment in a wide variety of clinical scenarios and demographics.

With the tremendous burden the ongoing COVID-19 pandemic has put on healthcare systems and healthcare workers around the world, the hope is that research such as COVID-Net CT-2 and open-source initiatives such as the COVID-Net initiative can accelerate the advancement and adoption of deep learning solutions within a clinical setting to aid front-line health workers and healthcare systems in improving clinical workflow efficiency and effectiveness in the fight against the COVID-19 pandemic. While to the best of the authors' knowledge this research does not put anyone at a disadvantage, it is important to note that COVID-Net CT-2 is not a production-ready solution and is meant for research purposes. As such, predictions made by COVID-Net CT-2 should not be utilized blindly and should instead be built upon and leveraged in a human-in-the-loop fashion by researchers, clinicians, and citizen data scientists alike. Future work involves leveraging the pre-trained networks for downstream tasks such as lung function prediction, severity assessment, and actionable predictions for guiding personalized treatment and care for SARS-CoV-2 positive patients.

Data Availability Statement

Publicly available datasets were analyzed in this study. This data can be found at: https://www.kaggle.com/hgunraj/covidxct.

Ethics Statement

This study was reviewed and approved by the University of Waterloo Ethics Board (42235). Written informed consent from the participants or their legal guardian/next of kin was not required to participate in this study in accordance with national legislation and institutional requirements.

Author Contributions

HG and AW conceived the experiments. HG conducted the experiments. DK and AS reviewed and reported on select patient cases and corresponding explainability results. All authors analyzed the results and reviewed the manuscript.

Conflict of Interest

AW was affiliated with DarwinAI Corp. DarwinAI Corp. provided computing support for this work, specifically through access to their deep learning development platform.

The remaining authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's Note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Acknowledgments

We would like to thank the Natural Sciences and Engineering Research Council of Canada (NSERC), the Canada Research Chairs program, the Canadian Institute for Advanced Research (CIFAR), DarwinAI Corp., Justin Kirby of the Frederick National Laboratory for Cancer Research, and the various organizations and initiatives from around the world collecting valuable COVID-19 data.

Footnotes

References

1. Wang W, Xu Y, Gao R, Lu R, Han K, Wu G, et al. Detection of SARS-CoV-2 in different types of clinical specimens. JAMA. (2020) 323:1843–4. doi: 10.1001/jama.2020.3786

PubMed Abstract | CrossRef Full Text | Google Scholar

2. Fang Y, Zhang H, Xie J, Lin M, Ying L, Pang P, et al. Sensitivity of chest CT for COVID-19: comparison to RT-PCR. Radiology. (2020) 296:E115–7. doi: 10.1148/radiol.2020200432

PubMed Abstract | CrossRef Full Text | Google Scholar

3. Li Y, Yao L, Li J, Chen L, Song Y, Cai Z, et al. Stability issues of RT-PCR testing of SARS-CoV-2 for hospitalized patients clinically diagnosed with COVID-19. J Med Virol. (2020) 92:903–8. doi: 10.1002/jmv.25786

PubMed Abstract | CrossRef Full Text | Google Scholar

4. Yang Y, Yang M, Shen C, Wang F, Yuan J, Li J, et al. Laboratory diagnosis and monitoring the viral shedding of SARS-CoV-2 infection. Innovation. (2020) 1:100061. doi: 10.1016/j.xinn.2020.100061

PubMed Abstract | CrossRef Full Text | Google Scholar

5. Ai T, Yang Z, Hou H, Zhan C, Chen C, Lv W, et al. Correlation of chest CT and RT-PCR testing for coronavirus disease 2019 (COVID-19) in China: a report of 1014 cases. Radiology. (2020) 296:E32–40. doi: 10.1148/radiol.2020200642

PubMed Abstract | CrossRef Full Text | Google Scholar

6. Xie X, Zhong Z, Zhao W, Zheng C, Wang F, Liu J. Chest CT for typical coronavirus disease 2019 (COVID-19) pneumonia: relationship to negative RT-PCR testing. Radiology. (2020) 296:E41–5. doi: 10.1148/radiol.2020200343

PubMed Abstract | CrossRef Full Text | Google Scholar

7. Tian S, Hu W, Niu L, Liu H, Xu H, Xiao SY. Pulmonary pathology of early-phase 2019 novel coronavirus (COVID-19) pneumonia in two patients with lung cancer. J Thorac Oncol. (2020) 15:700–4. doi: 10.20944/preprints202002.0220.v2

PubMed Abstract | CrossRef Full Text | Google Scholar

8. Shatri J, Tafilaj L, Turkaj A, Dedushi K, Shatri M, Bexheti S, et al. The role of chest computed tomography in asymptomatic patients of positive coronavirus disease 2019: a case and literature review. J Clin Imaging Sci. (2020) 10:35. doi: 10.25259/JCIS_58_2020

PubMed Abstract | CrossRef Full Text | Google Scholar

9. Guan Wj, Ni Zy, Hu Y, Liang Wh, Ou Cq, He Jx, et al. Clinical characteristics of coronavirus disease 2019 in China. N Engl J Med. (2020) 382:1708–20. doi: 10.1056/NEJMoa2002032

PubMed Abstract | CrossRef Full Text | Google Scholar

10. Wang D, Hu B, Hu C, Zhu F, Liu X, Zhang J, et al. Clinical characteristics of 138 hospitalized patients with 2019 novel coronavirus–infected pneumonia in Wuhan, China. JAMA. (2020) 323:1061–9. doi: 10.1001/jama.2020.1585

PubMed Abstract | CrossRef Full Text | Google Scholar

11. Chung M, Bernheim A, Mei X, Zhang N, Huang M, Zeng X, et al. CT Imaging features of 2019 novel coronavirus (2019-nCoV). Radiology. (2020) 295:202–7. doi: 10.1148/radiol.2020200230

PubMed Abstract | CrossRef Full Text | Google Scholar

12. Pan F, Ye T, Sun P, Gui S, Liang B, Li L, et al. Time course of lung changes at chest CT during recovery from coronavirus disease 2019 (COVID-19). Radiology. (2020) 295:715–21. doi: 10.1148/radiol.2020200370

PubMed Abstract | CrossRef Full Text | Google Scholar

13. Bai HX, Hsieh B, Xiong Z, Halsey K, Choi JW, Tran TML, et al. Performance of radiologists in differentiating COVID-19 from Non-COVID-19 viral pneumonia at chest CT. Radiology. (2020) 296:E46–54. doi: 10.1148/radiol.2020200823

PubMed Abstract | CrossRef Full Text | Google Scholar

14. Mei X, Lee HC, Diao Ky, Huang M, Lin B, Liu C, et al. Artificial intelligence–enabled rapid diagnosis of patients with COVID-19. Nat Med. (2020) 26:1224–85. doi: 10.1038/s41591-020-0931-3

PubMed Abstract | CrossRef Full Text | Google Scholar

15. Gunraj H, Wang L, Wong A. COVIDNet-CT: a tailored deep convolutional neural network design for detection of COVID-19 cases from chest CT images. Front Med. (2020) 7:608525. doi: 10.3389/fmed.2020.608525

PubMed Abstract | CrossRef Full Text | Google Scholar

16. Zhang K, Liu X, Shen J, Li Z, Sang Y, Wu X, et al. Clinically applicable AI system for accurate diagnosis, quantitative measurements, and prognosis of COVID-19 pneumonia using computed tomography. Cell. (2020) 18:1423–33. doi: 10.1016/j.cell.2020.04.045

PubMed Abstract | CrossRef Full Text | Google Scholar

17. Wang L, Lin ZQ, Wong A. COVID-Net: a tailored deep convolutional neural network design for detection of COVID-19 Cases from chest X-ray images. Sci Rep. (2020) 10:19549. doi: 10.1038/s41598-020-76550-z

PubMed Abstract | CrossRef Full Text | Google Scholar

18. Wong A, Lin ZQ, Wang L, Chung AG, Shen B, Abbasi A, et al. COVIDNet-S: towards computer-aided severity assessment via training and validation of deep neural networks for geographic extent and opacity extent scoring of chest X-rays for SARS-CoV-2 lung disease severity. Sci Rep. (2020) 11:9315. doi: 10.1038/s41598-021-88538-4

PubMed Abstract | CrossRef Full Text | Google Scholar

19. An P, Xu S, Harmon SA, Turkbey EB, Sanford TH, Amalou A, et al. CT images in Covid-19 [Data set]. In: The Cancer Imaging Archive. (2020). doi: 10.7937/tcia.2020.gqry-nc81

CrossRef Full Text | Google Scholar

20. Rahimzadeh M, Attar A, Sakhaei SM. A fully automated deep learning-based network for detecting COVID-19 from a new and large lung CT scan dataset. Biomed. Signal Process. Control. (2021) 68:102588. doi: 10.1016/j.bspc.2021.102588

PubMed Abstract | CrossRef Full Text | Google Scholar

21. Ning W, Lei S, Yang J, Cao Y, Jiang P, Yang Q, et al. Open resource of clinical data from patients with pneumonia for the prediction of COVID-19 outcomes via deep learning. Nat Biomed Eng. (2020) 4:1197–207. doi: 10.1038/s41551-020-00633-5

PubMed Abstract | CrossRef Full Text | Google Scholar

22. Jun M, Yixin W, Xingle A, Cheng G, Ziqi Y, Jianan C, et al. Towards Data-Efficient Learning: A Benchmark for COVID-19 CT Lung and Infection Segmentation. arXiv [Preprint]. (2020) arXiv:2004.12537.

Google Scholar

23. Armato S III, McLennan G, Bidaut L, McNitt-Gray M, Meyer C, Reeves A, et al. Data from LIDC-IDRI. In: The Cancer Imaging Archive. (2015). doi: 10.7937/K9/TCIA.2015.LO9QL9SX

CrossRef Full Text | Google Scholar

24. COVID-19. Radiopaedia. (2020). Available online at: https://radiopaedia.org/articles/covid-19-4 (accessed January 19, 2021).

Google Scholar

25. Morozov SP, Andreychenko AE, Pavlov NA, Vladzymyrskyy AV, Ledikhova NV, Gombolevskiy VA, et al. MosMedData: chest CT scans with COVID-19 related findings dataset. medRxiv [Preprint]. (2020). doi: 10.1101/2020.05.20.20100362

CrossRef Full Text | Google Scholar

26. Iandola FN, Han S, Moskewicz MW, Ashraf K, Dally WJ, Keutzer K. SqueezeNet: AlexNet-Level Accuracy With 50x Fewer Parameters and <0.5MB Model Size. arXiv. (2019) arXiv.1602.07360.

Google Scholar

27. Sandler M, Howard A, Zhu M, Zhmoginov A, Chen LC. MobileNetV2: inverted residuals and linear bottlenecks. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City, UT (2018). p. 4510–20. doi: 10.1109/CVPR.2018.00474

CrossRef Full Text | Google Scholar

28. Tan M, Le Q. EfficientNet: rethinking model scaling for convolutional neural networks. In: 2019 International Conference on Machine Learning (ICML). Long Beach, CA (2019).

Google Scholar

29. Zoph B, Vasudevan V, Shlens J, Le QV. Learning transferable architectures for scalable image recognition. In: 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition. Salt Lake City, UT (2018). p. 8697–710. doi: 10.1109/CVPR.2018.00907

CrossRef Full Text | Google Scholar

30. Wong A, Shafiee MJ, Chwyl B, Li F. GenSynth: a generative synthesis approach to learning generative machines for generate efficient neural networks. Electron Lett. (2019) 55:986–9. doi: 10.1049/el.2019.1719

CrossRef Full Text | Google Scholar

31. Qian N. On the momentum term in gradient descent learning algorithms. Neural Netw. (1999) 12:145–51. doi: 10.1016/S0893-6080(98)00116-6

PubMed Abstract | CrossRef Full Text | Google Scholar

32. Wong A. NetScore: Towards universal metrics for large-scale performance analysis of deep neural networks for practical usage. In: 16th International Conference on Image Analysis and Recognition (ICIAR). (2019) 15–26. doi: 10.1007/978-3-030-27272-2_2

CrossRef Full Text | Google Scholar

33. Lin ZQ, Shafiee MJ, Bochkarev S, Jules MS, Wang XY, Wong A. Do explanations reflect decisions? A machine-centric strategy to quantify the performance of explainability algorithms. arXiv. (2019) 1910.07387.

Google Scholar

34. Kumar D, Wong A, Taylor GW. Explaining the unexplained: a class-enhanced attentive response (CLEAR) approach to understanding deep neural networks. In: IEEE Conference on Computer Vision and Pattern Recognition Workshops. Honolulu, HI (2017). doi: 10.1109/CVPRW.2017.215

CrossRef Full Text | Google Scholar

35. Lundberg S, Lee SI. A unified approach to interpreting model predictions. In: 31st International Conference on Neural Information Processing Systems (NIPS). Long Beach, CA (2017). p. 4768–77.

Google Scholar

36. Ribeiro MT, Singh S, Guestrin C. “Why should i trust you?”: explaining the predictions of any classifier. In: 22nd ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD). San Francisco, CA (2017). p. 1135–1144.

Google Scholar

37. Erion G, Janizek JD, Sturmfels P, Lundberg S, Lee SI. Improving performance of deep learning models with axiomatic attribution priors and expected gradients. Nat Mach Intell. (2020) 3:620–31. doi: 10.1038/s42256-021-00343-w

CrossRef Full Text | Google Scholar

38. Kumar D, Taylor GW, Wong A. Discovery radiomics with CLEAR-DR: interpretable computer aided diagnosis of diabetic retinopathy (2017).

Google Scholar

39. Xu X, Jiang X, Ma C, Du P, Li X, Lv S, et al. A deep learning system to screen novel coronavirus disease 2019 pneumonia. Engineering. (2020) 6:1122–9. doi: 10.1016/j.eng.2020.04.010

PubMed Abstract | CrossRef Full Text | Google Scholar

40. Bai HX, Wang R, Xiong Z, Hsieh B, Chang K, Halsey K, et al. AI augmentation of radiologist performance in distinguishing COVID-19 from pneumonia of other etiology on chest CT. Radiology. (2020) 296:201491. doi: 10.1148/radiol.2020201491

PubMed Abstract | CrossRef Full Text | Google Scholar

41. Li L, Qin L, Xu Z, Yin Y, Wang X, Kong B, et al. Using artificial intelligence to detect COVID-19 and community-acquired pneumonia based on pulmonary CT: evaluation of the diagnostic accuracy. Radiology. (2020) 296:E65–71. doi: 10.1148/radiol.2020200905

PubMed Abstract | CrossRef Full Text | Google Scholar

42. Ardakani AA, Kanafi AR, Acharya UR, Khadem N, Mohammadi A. Application of deep learning technique to manage COVID-19 in routine clinical practice using CT images: results of 10 convolutional neural networks. Comput Biol Med. (2020) 121:103795. doi: 10.1016/j.compbiomed.2020.103795

PubMed Abstract | CrossRef Full Text | Google Scholar

43. Shah V, Keniya R, Shridharani A, Punjabi M, Shah J, Mehendale N. Diagnosis of COVID-19 using CT scan images and deep learning techniques. Emerg. Radiology. (2021) 28:497-505. doi: 10.1007/s10140-020-01886-y

PubMed Abstract | CrossRef Full Text | Google Scholar

44. Chen J, Wu L, Zhang J, Zhang L, Gong D, Zhao Y, et al. Deep learning-based model for detecting 2019 novel coronavirus pneumonia on high-resolution computed tomography: a prospective study. medRxiv. (2020). doi: 10.1101/2020.02.25.20021568

CrossRef Full Text | Google Scholar

45. Zheng C, Deng X, Fu Q, Zhou Q, Feng J, Ma H, et al. A Weakly-Supervised Framework for COVID-19 Classification and Lesion Localization From Chest CT. IEEE Trans Med Imaging.. (2020) 39:2615–25. doi: 10.1109/TMI.2020.2995965

PubMed Abstract | CrossRef Full Text | Google Scholar

46. Wang B, Jin S, Yan Q, Xu H, Luo C, Wei L, et al. AI-assisted CT imaging analysis for COVID-19 screening: Building and deploying a medical AI system. Appl Soft Comput. (2021) 98:106897. doi: 10.1016/j.asoc.2020.106897

PubMed Abstract | CrossRef Full Text | Google Scholar

47. Jin C, Chen W, Cao Y, Xu Z, Tan Z, Zhang X, et al. Development and evaluation of an artificial intelligence system for COVID-19 diagnosis. Nat Commun. (2020) 11:5088. doi: 10.1038/s41467-020-18685-1

PubMed Abstract | CrossRef Full Text | Google Scholar

48. Song Y, Zheng S, Li L, Zhang X, Zhang X, Huang Z, et al. Deep learning enables accurate diagnosis of novel coronavirus (COVID-19) with CT images. IEEE/ACM Trans Comput Biol Bioinform. (2021) 18:2775–80. doi: 10.1109/TCBB.2021.3065361

PubMed Abstract | CrossRef Full Text | Google Scholar

49. Wang S, Kang B, Ma J, Zeng X, Xiao M, Guo J, et al. A deep learning algorithm using CT images to screen for Corona Virus Disease (COVID-19). Eur Radiol. (2021) 31:6096–104. doi: 10.1007/s00330-021-07715-1

PubMed Abstract | CrossRef Full Text | Google Scholar

50. Harmon SA, Sanford TH, Xu S, Turkbey EB, Roth H, Xu Z, et al. Artificial intelligence for the detection of COVID-19 pneumonia on chest CT using multinational datasets. Nat Commun. (2020) 11:4080. doi: 10.1038/s41467-020-17971-2

PubMed Abstract | CrossRef Full Text | Google Scholar

51. Islam MM, Karray F, Alhajj R, Zeng J. A review on deep learning techniques for the diagnosis of novel coronavirus (COVID-19). IEEE Access. (2021) 9:30551–72. doi: 10.1109/ACCESS.2021.3058537

PubMed Abstract | CrossRef Full Text | Google Scholar

52. He K, Zhang X, Ren S, Sun J. Deep residual learning for image recognition. In: 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR). Las Vegas, NV (2016). p. 770–8. doi: 10.1109/CVPR.2016.90

PubMed Abstract | CrossRef Full Text | Google Scholar

53. Hasan AM, AL-Jawad MM, Jalab HA, Shaiba H, Ibrahim RW, AL-Shamasneh AR. Classification of COVID-19 coronavirus, pneumonia and healthy lungs in CT scans using q-deformed entropy and deep learning features. Entropy. (2020) 22:517. doi: 10.3390/e22050517

PubMed Abstract | CrossRef Full Text | Google Scholar

54. Javaheri T, Homayounfar M, Amoozgar Z, Reiazi R, Homayounieh F, Abbas E, et al. CovidCTNet: an open-source deep learning approach to diagnose covid-19 using small cohort of CT images. NPJ Digit Med. (2021) 4:29. doi: 10.1038/s41746-021-00399-3

PubMed Abstract | CrossRef Full Text | Google Scholar

55. Wu X, Hui H, Niu M, Li L, Wang L, He B, et al. Deep learning-based multi-view fusion model for screening 2019 novel coronavirus pneumonia: a multicentre study. Eur J Radiol. (2020) 128:109041. doi: 10.1016/j.ejrad.2020.109041

PubMed Abstract | CrossRef Full Text | Google Scholar

Keywords: COVID-19, computed tomography, deep learning, image classification, radiology, SARS-CoV-2, pneumonia

Citation: Gunraj H, Sabri A, Koff D and Wong A (2022) COVID-Net CT-2: Enhanced Deep Neural Networks for Detection of COVID-19 From Chest CT Images Through Bigger, More Diverse Learning. Front. Med. 8:729287. doi: 10.3389/fmed.2021.729287

Received: 22 June 2021; Accepted: 31 December 2021;
Published: 10 March 2022.

Edited by:

Wu Yuan, The Chinese University of Hong Kong, China

Reviewed by:

Saman Motamed, University of Toronto, Canada
Savvakis Nicolaou, University of British Columbia, Canada

Copyright © 2022 Gunraj, Sabri, Koff and Wong. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Hayden Gunraj, hayden.gunraj@uwaterloo.ca

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.