The Combination of Data-Driven Machine Learning Approaches and Prior Knowledge for Robust Medical Image Processing and Analysis

28.3K
views
106
authors
13
articles
Cover image for research topic "The Combination of Data-Driven Machine Learning Approaches and Prior Knowledge for Robust Medical Image Processing and Analysis"
Editors
4
Impact
Loading...
1,812 views
4 citations
2,843 views
4 citations
Original Research
14 April 2023
MIB-ANet: A novel multi-scale deep network for nasal endoscopy-based adenoid hypertrophy grading
Mingmin Bi
5 more and 
Linlin Shen
(A) The overview of the proposed MIB-ANet architecture. (B) The architecture of ANet. (C) The architecture of Inception Block and Modified Inception Block.

Introduction: To develop a novel deep learning model to automatically grade adenoid hypertrophy, based on nasal endoscopy, and asses its performance with that of E.N.T. clinicians.

Methods: A total of 3,179 nasoendoscopic images, including 4-grade adenoid hypertrophy (Parikh grading standard, 2006), were collected to develop and test deep neural networks. MIB-ANet, a novel multi-scale grading network, was created for adenoid hypertrophy grading. A comparison between MIB-ANet and E.N.T. clinicians was conducted.

Results: In the SYSU-SZU-EA Dataset, the MIB-ANet achieved 0.76251 F1 score and 0.76807 accuracy, and showed the best classification performance among all of the networks. The visualized heatmaps show that MIB-ANet can detect whether adenoid contact with adjacent tissues, which was interpretable for clinical decision. MIB-ANet achieved at least 6.38% higher F1 score and 4.31% higher accuracy than the junior E.N.T. clinician, with much higher (80× faster) diagnosing speed.

Discussion: The novel multi-scale grading network MIB-ANet, designed for adenoid hypertrophy, achieved better classification performance than four classical CNNs and the junior E.N.T. clinician. Nonetheless, further studies are required to improve the accuracy of MIB-ANet.

3,973 views
5 citations
Original Research
09 March 2023

The heart is a relatively complex non-rigid motion organ in the human body. Quantitative motion analysis of the heart takes on a critical significance to help doctors with accurate diagnosis and treatment. Moreover, cardiovascular magnetic resonance imaging (CMRI) can be used to perform a more detailed quantitative analysis evaluation for cardiac diagnosis. Deformable image registration (DIR) has become a vital task in biomedical image analysis since tissue structures have variability in medical images. Recently, the model based on masked autoencoder (MAE) has recently been shown to be effective in computer vision tasks. Vision Transformer has the context aggregation ability to restore the semantic information in the original image regions by using a low proportion of visible image patches to predict the masked image patches. A novel Transformer-ConvNet architecture is proposed in this study based on MAE for medical image registration. The core of the Transformer is designed as a masked autoencoder (MAE) and a lightweight decoder structure, and feature extraction before the downstream registration task is transformed into the self-supervised learning task. This study also rethinks the calculation method of the multi-head self-attention mechanism in the Transformer encoder. We improve the query-key-value-based dot product attention by introducing both depthwise separable convolution (DWSC) and squeeze and excitation (SE) modules into the self-attention module to reduce the amount of parameter computation to highlight image details and maintain high spatial resolution image features. In addition, concurrent spatial and channel squeeze and excitation (scSE) module is embedded into the CNN structure, which also proves to be effective for extracting robust feature representations. The proposed method, called MAE-TransRNet, has better generalization. The proposed model is evaluated on the cardiac short-axis public dataset (with images and labels) at the 2017 Automated Cardiac Diagnosis Challenge (ACDC). The relevant qualitative and quantitative results (e.g., dice performance and Hausdorff distance) suggest that the proposed model can achieve superior results over those achieved by the state-of-the-art methods, thus proving that MAE and improved self-attention are more effective and promising for medical image registration tasks. Codes and models are available at https://github.com/XinXiao101/MAE-TransRNet.

3,747 views
3 citations

Introduction: It is critical to identify the stroke onset time of patients with acute ischemic stroke (AIS) for the treatment of endovascular thrombectomy (EVT). However, it is challenging to accurately ascertain this time for patients with wake-up stroke (WUS). The current study aimed to construct a deep learning approach based on computed tomography perfusion (CTP) or perfusion weighted imaging (PWI) to identify a 6-h window for patients with AIS for the treatment of EVT.

Methods: We collected data from 377 patients with AIS, who were examined by CTP or PWI before making a treatment decision. Cerebral blood flow (CBF), time to maximum peak (Tmax), and a region of interest (ROI) mask were preprocessed from the CTP and PWI. We constructed the classifier based on a convolutional neural network (CNN), which was trained by CBF, Tmax, and ROI masks to identify patients with AIS within a 6-h window for the treatment of EVT. We compared the classification performance among a CNN, support vector machine (SVM), and random forest (RF) when trained by five different types of ROI masks. To assess the adaptability of the classifier of CNN for CTP and PWI, which were processed respectively from CTP and PWI groups.

Results: Our results showed that the CNN classifier had a higher performance with an area under the curve (AUC) of 0.935, which was significantly higher than that of support vector machine (SVM) and random forest (RF) (p = 0.001 and p = 0.001, respectively). For the CNN classifier trained by different ROI masks, the best performance was trained by CBF, Tmax, and ROI masks of Tmax > 6 s. No significant difference was detected in the classification performance of the CNN between CTP and PWI (0.902 vs. 0.928; p = 0.557).

Discussion: The CNN classifier trained by CBF, Tmax, and ROI masks of Tmax > 6 s had good performance in identifying patients with AIS within a 6-h window for the treatment of EVT. The current study indicates that the CNN model has potential to be used to accurately estimate the stroke onset time of patients with WUS.

1,909 views
4 citations
Open for submission
Frontiers Logo

Frontiers in Medicine

Recent developments in artificial intelligence and radiomics
Edited by Luca Urso, Luigi Manco, Salvatore Annunziata
Deadline
30 June 2025
Submit a paper
Recommended Research Topics
115.8K
views
128
authors
18
articles
Frontiers Logo

Frontiers in Medicine

Artificial Intelligence in Positron Emission Tomography
Edited by XIAOLI LAN, Kuangyu Shi, Xiu Ying Wang, Chuantao Zuo
62.9K
views
96
authors
13
articles
Frontiers Logo

Frontiers in Medicine

Small Animal Imaging: Technological and Methodological Advances to Improve the Translational Power
Edited by Monique R. Bernsen, TERRY JONES, Wendy McDougald, Laura Mezzanotte, Carmel Mary Moran, Adriana Tavares, Louise Van Der Weerd
46.6K
views
84
authors
13
articles
Frontiers Logo

Frontiers in Medicine

Radiomics and Artificial Intelligence in Radiology and Nuclear Medicine
Edited by Giorgio Treglia, Salvatore Annunziata
41.9K
views
120
authors
13
articles
Frontiers Logo

Frontiers in Medicine

Workflow Optimisation for Radiological Imaging
Edited by Jie-Zhi Cheng, Minjeong Kim, Xin Yang
32K
views
130
authors
17
articles