
94% of researchers rate our articles as excellent or good
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.
Find out more
ORIGINAL RESEARCH article
Front. Oncol. , 11 March 2025
Sec. Cancer Imaging and Image-directed Interventions
Volume 15 - 2025 | https://doi.org/10.3389/fonc.2025.1508525
Objective: To explore the application of a deep learning model based on lateral nasopharyngeal X-rays in diagnosing tonsillar and adenoid hypertrophy.
Methods: A retrospective study was conducted using DICOM images of lateral nasopharyngeal X-rays from pediatric outpatients aged 2-12 at our hospital from July 2014 to July 2024. The study included patients exhibiting varying degrees of respiratory obstruction symptoms (disease group). Initially, 1006 images were collected, but after excluding low-quality images and standardizing the imaging phase, 819 images remained. These images were divided into training and validation sets in an 8:2 ratio. The independent test set is consisted of 484 images. We delineated the target areas for tonsils and adenoids and used a YOLOv8n-based model for object detection and use various convolutional neural network models to classify the cropped images, assessing the severity of tonsillar and adenoid hypertrophy. We compared the performance of these models on the training and validation sets using metrics such as ROC-AUC, accuracy, precision, recall, and F1 score.
Results: The combined model, incorporating YOLOv8 for object detection and secondary classification, demonstrated excellent performance in diagnosing tonsillar and adenoid hypertrophy, significantly improving diagnostic accuracy and consistency. The ResNet18 model, due to its lightweight nature and minimal computational resource requirements, performed exceptionally well in the YOLOv8-ResNet fusion model for detecting and classifying tonsils and adenoids, making it our preferred model.
Conclusion: The deep learning model combining YOLOv8n and ResNet18 based on lateral nasopharyngeal X-rays demonstrates significant advantages in diagnosing pediatric tonsillar and adenoid hypertrophy.
Tonsillar and adenoid hypertrophy are common upper respiratory tract diseases in children, significantly affecting their health and quality of life. The adenoids and tonsils are important lymphoid tissues in the pharynx, responsible for filtering pathogens entering the respiratory tract. However, when these tissues become excessively hypertrophic, they can cause symptoms such as mouth breathing, nasal obstruction, difficulty breathing, snoring, and sleep apnea. If left untreated, this can lead to facial deformities, growth retardation, cognitive impairment, increased cardiovascular risk (1), and behavioral problems (2). Adenoid and tonsillar hypertrophy are independently associated with the risk of pediatric obstructive sleep apnea syndrome (OSAS), with the prevalence of OSAS increasing with the size of the adenoids and tonsils (3). Adenotonsillectomy and medication are common treatments for children with adenoid and tonsillar hypertrophy who suffer from OSAS.
Lateral cephalometric radiographs, a standard orthodontic method for evaluating craniofacial morphology, provide orthodontists with readily available references for assessing airway obstruction and hypertrophic adenoids and tonsils (4). Numerous studies have reported a reasonable correlation between cephalometric measurements of adenoids and their size (5). Lateral cephalometric radiographs are accurate in diagnosing adenoid hypertrophy (6), but there is a lack of sufficient guidelines for diagnosing tonsillar hypertrophy using cephalometric measurements (7). Currently, the Fujioka method (8) is used to measure adenoid size on lateral radiographs by calculating the adenoid/nasopharyngeal (A/N) ratio (Figure 1). This ratio is derived by dividing the adenoid measurement (A) by the distance from the posterior edge of the hard palate to the anterior inferior edge of the spheno-occipital junction (N). An A/N ratio greater than 0.67 is considered indicative of adenoid hypertrophy.
Figure 1. The Fujioka method for measuring adenoid size on lateral radiographs by calculating the adenoid/nasopharyngeal (A/N) ratio. The black line marks the lateral contour of the occipital slope of the skull on the lateral radiograph. The green line (A) represents the vertical distance from the most prominent point of the adenoid to the lateral contour of the occipital slope of the skull. The red line (N) represents the width of the nasopharyngeal cavity at the level of the most prominent point of the adenoid.
The standard grading system for diagnosing tonsillar size and hypertrophy is based on clinical oropharyngeal examination (9), grading tonsils by the percentage they occupy in the oropharyngeal airway’s diameter (10). However, this widely used method is imperfect as it cannot show anterior-posterior obstruction of the oropharynx (11–13). Some literature suggests using the T/O ratio to evaluate tonsillar hypertrophy (7), but these previous evaluation metrics have certain limitations (14). They are also relatively cumbersome to operate and require a high level of diagnostic and assessment proficiency from different physicians.
In recent years, deep learning has made significant progress in medical image analysis and diagnosis (15), with broad applications in image segmentation for trauma fractures (16) and lung nodule diagnosis (17). This paper proposes a deep learning method combining YOLO (You Only Look Once) object detection and image classification for diagnosing tonsillar and adenoid hypertrophy in lateral nasopharyngeal X-rays. Tonsillar and adenoid hypertrophy are common pediatric health issues that can lead to airway obstruction, sleep apnea syndrome, and recurrent infections, among other serious health problems. Traditional diagnostic methods rely on the clinical experience of physicians, which can be subjective and have limitations in diagnostic accuracy. In contrast, deep learning models can reduce diagnostic bias (18). Therefore, developing an automated computer-aided diagnostic method is of paramount importance. Currently, there is no research that utilizes a combination of YOLO for object detection and CNNs for secondary classification to diagnose tonsillar and adenoid hypertrophy on the same lateral nasopharyngeal X-ray images.
Tonsils and adenoids reach their maximum size at around ages 7-9 and 12-13, respectively (19). In this study (Figure 2), we included children aged 2-13, who are the most common candidates for adenoidectomy and tonsillectomy. We collected lateral nasopharyngeal X-ray DICOM images of pediatric outpatients from our hospital from July 2014 to July 2024. These images were used as training, validation, and independent test sets. Initially, there were 1006 images in the training and validation sets, which were reduced to 819 images after excluding low-quality images and standardizing the imaging phase. The validation set included 164 images, evenly distributed among 104 cases of simple tonsillar hypertrophy, 268 cases of simple adenoid hypertrophy, 223 cases of both tonsillar and adenoid hypertrophy, and 224 normal cases. An additional independent test set was collected, consisting of 484 images, including 74 cases of simple tonsillar hypertrophy, 153 cases of both tonsillar and adenoid hypertrophy, 151 cases of simple adenoid hypertrophy, and 154 normal cases. The inclusion criteria for the collected images were: (1) adequate radiographic image quality, (2) children in the disease group with varying degrees of respiratory obstruction symptoms, and (3) children in the normal group without any clinical symptoms. Exclusion criteria included: (1) poor image quality or low contrast blurry images, (2) abnormal postures of the children during imaging (e.g., excessive neck extension or flexion), and (3) images taken in the inverse phase. The images were initially diagnosed by radiologists with 3 years of experience, and the final diagnoses were confirmed by two attending radiologists with 5-8 years of experience. Any discrepancies in diagnoses were resolved through discussion.
This study obtained parental consent and was approved by the Institutional Ethics Committee of our hospital.
YOLOv8 (20) (You Only Look Once) models use Darknet-53 (21) as their backbone, which includes residual and convolutional blocks. This model transforms object detection into a regression problem by directly predicting the class and location of the object. Previous studies have utilized this model for facial landmark detection (22), and Wenting Xie et al. applied an improved YOLOv8 model for ovarian cancer diagnosis (23). While YOLO models perform well in object detection and localization, they may not be as effective in distinguishing different states of the same organ (Figure 3), particularly in medical imaging. CNN models, composed of multiple stacked convolutional layers, are advanced deep learning technologies. They typically consist of an input layer (image input), a convolutional layer (which convolves the input image with filters to generate feature maps), a Rectified Linear Unit (ReLU) activation layer (which activates neurons above a threshold), a pooling layer (which reduces image size while retaining high-level features), and a fully connected (FC) layer (which produces the final results). These models can automatically extract features from image data and classify the images, and have been applied to various radiology tasks, achieving high performance in image-based disease classification. We found that combining CNN models with YOLO models can enhance the classification performance, especially in distinguishing subtle features. This approach integrates feature extraction and classification within the same network, providing a streamlined and efficient process. However, CNNs can struggle with imbalanced datasets, which often requires various techniques to balance the data. Data augmentation is commonly used in the medical field to increase the size of datasets. This method generates additional labeled images without altering the semantics of the images, thereby mitigating dataset imbalance. In this study, we used several data augmentation methods, such as gamma transformation and horizontal flipping, which do not affect the vertical orientation of the nasopharyngeal X-rays. Although CNNs require large labeled medical datasets for training, which can be challenging to create due to time and labor costs, recent research suggests that transfer learning can be a solution for small dataset issues.
Figure 3. Performance metrics of direct object detection and inference classification using the YOLOv8 model. (A) Calibration plots for each class. (B) Confusion matrix. (C) ROC curves.
In transfer learning (TL), convolutional neural networks (CNNs) are first trained to learn features from a broad domain, such as ImageNet. The trained features and network parameters are then transferred to a more specific domain. In CNN models, low-level features like edges, curves, and corners are learned in the initial layers, while more specific high-level features are learned in the final layers. Among various TL models, we selected seven CNN models, including ResNet18, ResNet34, ResNet50, DenseNet121, EfficientNet-B0, VGG16, and AlexNet, to compare their performance using different metrics. Ultimately, we chose to combine the ResNet18 with the YOLOv8n model. ResNet-18 is a convolutional neural network with 18 layers that addresses the issue of training deep networks by introducing residual blocks. Its architecture includes an initial convolutional layer, four stages of residual blocks (each stage containing two 3x3 convolutional layers), a global average pooling layer, and a fully connected layer. This design allows the network to effectively train deep models while maintaining relatively low computational complexity. A recent study (24) evaluated the performance of several neural networks with a softmax output layer and ReLU activation, confirming the superior performance of softmax with ReLU in classification tasks. Therefore, to obtain probabilistic predictions, we used a fully connected layer as the output layer and modified it for four-class classification. We chose the cross-entropy function as the loss function, which inherently includes the Softmax operation, eliminating the need for an explicit softmax definition. The structure of ResNet18 incorporates a key design feature known as residual blocks. These blocks allow the model to use ‘skip connections’ to pass inputs directly to subsequent layers, bypassing some intermediate layers of neurons. This design helps to mitigate the vanishing and exploding gradient problems that can arise as the network depth increases, which makes it easier for the network to learn deep representations without degradation. Through this structure, ResNet18 can effectively capture both low-level and high-level features in images, leading to excellent performance in image classification tasks. In the process of transfer learning and fine-tuning, we initially froze the first four modules of the model and gradually unfroze all layers during training. During training, we calculated the output and loss through forward propagation and updated the weights via backpropagation. We also dynamically adjusted the learning rate to avoid catastrophic forgetting. Additionally, we modified the output of the fully connected (FC) layer to accommodate a four-class classification task. After testing various optimizers, we selected the “Adam” optimizer for its superior performance among all those studied. Consequently, we applied this optimizer during the model training process. To understand the model’s attention to different regions of the images, we employed a visualization tool called Gradient-weighted Class Activation Mapping (2D-Grad-CAM) (25) to analyze the interpretability of the five fused deep learning models. This tool visually identifies the correspondence between the regions of interest in the pathological area and the model’s prediction attention. In this study, we focused on whether the models not only paid attention to the tonsil and adenoid areas but also evaluated the regions of the airway.
The collected lateral nasopharyngeal X-rays were first converted from DICOM files to JPG images using Python’s PIL package. Then, experienced radiologists annotated the target detection boxes using the LabelImg software, marking the tonsil and adenoid detection regions on each X-ray. The annotations were completed collaboratively by two experts. For images where there was disagreement, the experts engaged in discussions to reach a consensus and ensure the accuracy of the classifications. The adenoid region was defined from the upper edge of the nasopharynx to the palatal plane, from the posterior edge of the hard palate to the posterior pharyngeal wall and the anterior aspect of the C1 vertebra. The tonsil detection region was defined from the palatal uvula level to the posterior aspect of the tongue, from the anterior aspect of the C3 vertebra to the aryepiglottic fold. The detection boxes were rectangular regions determined by vertical and horizontal lines passing through these anatomical landmarks, ensuring that the entire airway cross-sectional content of the tonsil or adenoid was included in the training images and improving the localization accuracy of the YOLO model’s detection boxes.
After defining the detection boxes, we devised a two-stage detection + diagnosis method to diagnose whether children had tonsillar or adenoid hypertrophy on the same lateral nasopharyngeal X-ray. First, we fine-tuned the pre-trained YOLOv8m model on ImageNet to learn the detection box locations, adjusting model parameters and hyperparameters through iterative training. After 135 epochs, we controlled the box loss error range within 0.5%, using IOU-based evaluations to improve detection accuracy. The model’s final weights were saved corresponding to the lowest box loss during the last training iteration. For the training and evaluation of the secondary classification model, we first used the trained YOLO model to predict and crop the original images from the training and validation sets. The highest-confidence detection boxes for the predicted tonsil and adenoid regions were selected as the final detection boxes. These final detection boxes were then used to crop the original images, resulting in predicted cropped images. These cropped images were categorized into four classes: normal tonsils, enlarged tonsils, normal adenoids, and enlarged adenoids. We then re-established the training and validation sets, and the cropped images were fed into the secondary classification models, pretrained on ImageNet, to learn image features for classification. This process generated the optimal model weight files for different base models and organs.
During the evaluation on the independent test set, we repeated the same procedure, using the secondary classification models to predict and obtain the labels for the cropped images. These predicted labels were compared with the true labels for the independent test set. We utilized two methods for correction: first, we evaluated the predicted cropped images and true labels from the independent test set; second, we cropped the original images using the true labels and input these into the secondary classification models for performance evaluation.
The results from these two evaluation methods were used to assess the secondary classification models, determining the optimal models for classifying tonsil and adenoid images. This approach ensured the overall performance of the combined YOLO model and secondary classification model was optimal.
We use the following indicators to evaluate the performance of the model and select the best model: ROC AUC,accuracy, sensitivity, specificity, F1 score, Confusion Matrix Visualization images. This study used the following tools: Python 3.7.16 (https://www.python.org/downloads/release/python-3716/) and PyTorch third-party libraries(Version: 1.13.1)on Windows 11 operating system ([MSC v.1916 64 bit (AMD64)]).
By comparing performance metrics across the training, validation (Figure 4, Table 1), and independent test sets (Figure 5, Table 2), along with visualizing confusion matrices and the regions of interest highlighted by the models, we concluded that the fusion model consisting of YOLOv8n as the front-end model for object detection and localization, combined with the fine-tuned ResNet18 as the back-end secondary classification model, demonstrates significant advantages in diagnosing conditions using lateral nasopharyngeal X-ray images.
Figure 4. Performance metrics of the ResNet18 model on training and validation sets. (A) Training and validation accuracy curves. (B) Training and validation loss curves. (C) Calibration curves. (D) Confusion matrices. (E) ROC curves.
Figure 5. Performance metrics of the ResNet18 model on test sets. (A) Calibration curves. (B) Confusion matrices. (C) Visualization heatmaps of the tonsil region. (D) Visualization heatmaps of the adenoid region.
In this experiment, we explored the use of YOLOv8n for object detection and localization, combined with various deep learning models like ResNet18 for secondary classification. YOLOv8n accurately identifies the regions of interest, while the subsequent classification is performed by models such as ResNet18. The goal of this approach was to improve classification accuracy for four categories: normal adenoid, normal tonsil, enlarged adenoid, and enlarged tonsil. The results showed that the combination of YOLOv8n and ResNet18 performed exceptionally well across several key metrics. On the independent test set, our combined model achieved an impressive classification accuracy of 97%, with AUC values above 0.95 for each category. These results were further corroborated by the ROC curves, which demonstrated strong discriminatory power between the different classes. While other models like AlexNet, VGG16, DenseNet121, EfficientNet-B0, ResNet34, and ResNet50 also performed well, especially in certain complex classification tasks, they generally fell short of ResNet18’s performance. Despite having ROC AUC values close to or reaching 0.95, the confusion matrices and calibration curves revealed that these models had slightly higher misclassification rates and calibration inaccuracies in some categories. In particular, DenseNet121 and EfficientNet-B0, while stable in classification tasks, showed some limitations in fine-grained classification compared to ResNet18. In summary, through a comprehensive analysis of the classification reports, ROC curves, calibration curves, and confusion matrices across the training, validation, and independent test sets, we found that the combination of YOLOv8n and ResNet18 outperformed the other models in this four-classification task. This model structure not only improved classification accuracy and stability but also demonstrated good generalization capability, making it suitable for practical applications. Therefore, we ultimately chose the YOLOv8n and ResNet18 fusion model as the integrated approach for diagnosing tonsil and adenoid enlargement based on lateral nasopharyngeal radiographs. Currently, the treatment standards for tonsillar and adenoid hypertrophy are based on the presence of clinical symptoms. For patients with clinically significant symptoms or recurrent inflammation due to hypertrophic tonsils or adenoids, tonsillectomy and adenoidectomy are recommended. Our deep learning model can play a role in large-scale screenings and help avoid unnecessary surgical interventions caused by diagnostic variability among individuals.
Lateral nasopharyngeal X-rays offer convenience, low cost, high repeatability, and ease of operation. Compared to maxillofacial CT scans (14), they have a lower radiation dose, making them less harmful to children. Additionally, they provide better standardization and operability compared to ultrasound examinations, which often depend on the skill and subjective judgment of the ultrasonographer and lack the standardization of nasopharyngeal X-rays (26, 27). Currently, clinical practice often involves diagnosing based on nasopharyngeal X-rays combined with clinical symptoms. We selected children who already exhibited symptoms and underwent X-ray imaging to ensure diagnostic accuracy.
Previous studies on deep learning have primarily focused on the diagnosis of adenoid enlargement, without including tonsillar hypertrophy. In contrast, our study addresses the diagnosis of both tonsillar and adenoid enlargement. Many of the earlier studies on diagnosing adenoid enlargement using X-ray nasopharyngeal lateral films relied on manual annotation (28), which could lead to inconsistent diagnostic quality due to variability in annotation skills, and the process is often time-consuming. Some studies have used CT images for automated adenoid enlargement diagnosis (29); however, CT scans involve higher radiation doses than X-rays, making them less suitable for routine screening. More recent studies have employed region-of-interest (ROI) delineation followed by deep learning models for automated diagnosis (30), but this still requires preprocessing of the X-ray nasopharyngeal lateral films. Other research has focused on detecting anatomical landmarks on nasopharyngeal lateral images to automatically diagnose adenoid enlargement (31–33). Although these methods have demonstrated strong diagnostic performance, they are less intuitive in their presentation and still require preprocessing of the nasopharyngeal lateral films.
In this stepwise model architecture, errors in the target detection phase can propagate to the classification phase. If YOLOv8 fails to accurately locate the target area or mistakenly detects background regions, the subsequent classification model will process incorrect input data, reducing classification accuracy. To mitigate this impact, we corrected the performance evaluation results using real cropped images during the evaluation phase, ensuring accurate prediction and classification by the model.
In medical images, the positions and shapes of tonsils and adenoids may vary due to patient differences and imaging angles. YOLOv8, as a powerful object detection tool, can adapt to these variations and accurately locate targets, which is crucial for handling complex scenarios and ensuring precise localization. The subsequent convolutional neural network can then perform fine-grained classification based on these accurate localization results. This ability to handle complex scenarios enhances the robustness and reliability of the entire system in practical applications.
This study has several limitations. It is a single-center study with a relatively small sample size, and the sample type (lateral nasopharyngeal X-rays) limits the amount of information that can be obtained from the images. Additionally, the 8:2 ratio for training and validation sets results in a small number of validation images. Accurate diagnosis and treatment of tonsillar and adenoid hypertrophy require consideration of clinical symptoms, and relying solely on X-rays is inappropriate. Although Resnet18 performed well compared to other models, the accuracy curve showed some fluctuations in the later stages of training. An ideal model would have high and closely aligned accuracy curves, ensuring both fitting capability and generalization ability. This fluctuation might indicate variable performance during each training epoch, especially when using small batch sizes.
In the future, we plan to expand the sample size, closely monitor changes in loss and accuracy curves, and collaborate with multiple centers to collect more lateral nasopharyngeal X-rays. This will increase the number of images in the training and validation sets and add additional test sets to validate the model’s prediction ability on unseen images, aiming to develop a more accurate disease diagnosis model. We will also explore using more advanced front-end models to replace YOLOv8 for target detection, further reducing error transmission. Additionally, we will attempt to use more lightweight models and optimize them for operation on mobile devices, making the model more convenient and effective for use in primary healthcare settings.
The deep learning model combining YOLOv8n and ResNet18 based on lateral nasopharyngeal X-rays demonstrates significant advantages in diagnosing pediatric tonsillar and adenoid hypertrophy.
The original contributions presented in the study are included in the article/Supplementary Material. Further inquiries can be directed to the corresponding author.
The studies involving humans were approved by the Institutional Ethics Committee of Children’s Hospital of Soochow University. The studies were conducted in accordance with the local legislation and institutional requirements. Written informed consent for participation in this study was provided by the participants’ legal guardians/next of kin.
ZW: Methodology, Writing – original draft. RZ: Conceptualization, Writing – original draft. YY: Data curation, Writing – review & editing. XL: Writing – original draft. BW: Funding acquisition, Writing – review & editing. JW: Project administration, Writing – review & editing.
The author(s) declare that no financial support was received for the research, authorship, and/or publication of this article.
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
The author(s) declare that no Generative AI was used in the creation of this manuscript.
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
The Supplementary Material for this article can be found online at: https://www.frontiersin.org/articles/10.3389/fonc.2025.1508525/full#supplementary-material
1. Marin JM, Carrizo SJ, Vicente E, Agusti AG. Long-term cardiovascular outcomes in men with obstructive sleep apnoea-hypopnoea with or without treatment with continuous positive airway pressure: an observational study. Lancet. (2005) 365:1046–53. doi: 10.1016/S0140-6736(05)71141-7
2. Baldassari CM, Mitchell RB, Schubert C, Rudnick EF. Pediatric obstructive sleep apnea and quality of life: a meta-analysis. Otolaryngol Head Neck Surg. (2008) 138:265–73. doi: 10.1016/j.otohns.2007.11.003
3. Xiao L, Su S, Liang J, Jiang Y, Shu Y, Ding L. Analysis of the risk factors associated with obstructive sleep apnea syndrome in chinese children. Front Pediatr. (2022) 10:900216. doi: 10.3389/fped.2022.900216
4. Franco LP, Souki BQ, Cheib PL, Abrao M, Pereira TB, Becker HM, et al. Are distinct etiologies of upper airway obstruction in mouth-breathing children associated with different cephalometric patterns? Int J Pediatr Otorhinolaryngol. (2015) 79:223–8. doi: 10.1016/j.ijporl.2014.12.013
5. Major MP, Flores-Mir C, Major PW. Assessment of lateral cephalometric diagnosis of adenoid hypertrophy and posterior upper airway obstruction: a systematic review. Am J Orthod Dentofacial Orthop. (2006) 130:700–8. doi: 10.1016/j.ajodo.2005.05.050
6. Duan H, Xia L, He W, Lin Y, Lu Z, Lan Q. Accuracy of lateral cephalogram for diagnosis of adenoid hypertrophy and posterior upper airway obstruction: A meta-analysis. Int J Pediatr Otorhinolaryngol. (2019) 119:1–9. doi: 10.1016/j.ijporl.2019.01.011
7. Lv C, Yang L, Ngan P, Xiao W, Zhao T, Tang B, et al. Role of the tonsil-oropharynx ratio on lateral cephalograms in assessing tonsillar hypertrophy in children seeking orthodontic treatment. BMC Oral Health. (2023) 23:836. doi: 10.1186/s12903-023-03573-z
8. Fujioka M, Young LW, Girdany BR. Radiographic evaluation of adenoidal size in children: adenoidal-nasopharyngeal ratio. AJR Am J Roentgenol. (1979) 133:401–4. doi: 10.2214/ajr.133.3.401
9. Brodsky L. Modern assessment of tonsils and adenoids. Pediatr Clin North Am. (1989) 36:1551–69. doi: 10.1016/S0031-3955(16)36806-7
10. Kumar DS, Valenzuela D, Kozak FK, Ludemann JP, Moxham JP, Lea J, et al. The reliability of clinical tonsil size grading in children. JAMA Otolaryngol Head Neck Surg. (2014) 140:1034–7. doi: 10.1001/jamaoto.2014.2338
11. Patel NA, Carlin K, Bernstein JM. Pediatric airway study: Endoscopic grading system for quantifying tonsillar size in comparison to standard adenotonsillar grading systems. Am J Otolaryngol. (2018) 39:56–64. doi: 10.1016/j.amjoto.2017.10.013
12. Wang JH, Chung YS, Jang YJ, Lee BJ. Palatine tonsil size and its correlation with subjective tonsil size in patients with sleep-disordered breathing. Otolaryngol Head Neck Surg. (2009) 141:716–21. doi: 10.1016/j.otohns.2009.09.007
13. Lai CC, Friedman M, Lin HC, Wang PC, Hsu CM, Yalamanchali S, et al. Objective versus subjective measurements of palatine tonsil size in adult patients with obstructive sleep apnea/hypopnea syndrome. Eur Arch Otorhinolaryngol. (2014) 271:2305–10. doi: 10.1007/s00405-014-2944-3
14. Iwasaki T, Sugiyama T, Yanagisawa-Minami A, Oku Y, Yokura A, Yamasaki Y. Effect of adenoids and tonsil tissue on pediatric obstructive sleep apnea severity determined by computational fluid dynamics. J Clin Sleep Med. (2020) 16:2021–8. doi: 10.5664/jcsm.8736
15. Yao F, Ding J, Hu Z, Cai M, Liu J, Huang X, et al. Ultrasound-based radiomics score: a potential biomarker for the prediction of progression-free survival in ovarian epithelial cancer. Abdom Radiol (NY). (2021) 46:4936–45. doi: 10.1007/s00261-021-03163-z
16. Hardalac F, Uysal F, Peker O, Ciceklidag M, Tolunay T, Tokgöz N, et al. Fracture detection in wrist X-ray images using deep learning-based object detection models. Sensors (Basel). (2022) 22(3):1285. doi: 10.3390/s22031285
17. You S, Park JH, Park B, Shin HB, Ha T, Yun JS, et al. The diagnostic performance and clinical value of deep learning-based nodule detection system concerning influence of location of pulmonary nodule. Insights Imaging. (2023) 14:149. doi: 10.1186/s13244-023-01497-4
18. Chen H, Yang BW, Qian L, Meng YS, Bai XH, Hong XW, et al. Deep learning prediction of ovarian Malignancy at US compared with O-RADS and expert assessment. Radiology. (2022) 304:106–13. doi: 10.1148/radiol.211367
19. Bhattacharjee R, Kheirandish-Gozal L, Spruyt K, Mitchell RB, Promchiarak J, Simakajornboon N, et al. Adenotonsillectomy outcomes in treatment of obstructive sleep apnea in children: a multicenter retrospective study. Am J Respir Crit Care Med. (2010) 182:676–83. doi: 10.1164/rccm.200912-1930OC
20. Terven J, Córdova-Esparza D-M, Romero-González J-A. A comprehensive review of YOLO architectures in computer vision: from YOLOv1 to YOLOv8 and YOLO-NAS. Mach Learn Knowledge Extraction. (2023) 5:1680–716. doi: 10.3390/make5040083
21. Varghese R, Sambath M. YOLOv8: A novel object detection algorithm with enhanced performance and robustness, in: 2024 International Conference on Advances in Data Engineering and Intelligent Computing Systems (ADICS). Chennai, India: IEEE. (2024).
22. Yuan Z, Shao P, Li J, Wang Y, Zhu Z, Qiu W, et al. YOLOv8-ACU: improved YOLOv8-pose for facial acupoint detection. Front Neurorobot. (2024) 18:1355857. doi: 10.3389/fnbot.2024.1355857
23. Xie W, Lin W, Li P, Lai H, Wang Z, Liu P, et al. Developing a deep learning model for predicting ovarian cancer in Ovarian-Adnexal Reporting and Data System Ultrasound (O-RADS US) Category 4 lesions: A multicenter study. J Cancer Res Clin Oncol. (2024) 150:346. doi: 10.1007/s00432-024-05872-6
24. Naz J, Sharif MI, Sharif MI, Kadry S, Rauf HT, Ragab AE. A comparative analysis of optimization algorithms for gastrointestinal abnormalities recognition and classification based on ensemble xcepNet23 and resNet18 features. Biomedicines. (2023) 11(6):1723. doi: 10.3390/biomedicines11061723
25. Selvaraju RR, Cogswell M, Das A, Vedantam R, Parikh D, Batra D. Grad-CAM: visual explanations from deep networks via gradient-based localization, in: 2017 IEEE International Conference on Computer Vision (ICCV), Venice, Italy: IEEE. (2017).
26. Bandarkar AN, Adeyiga AO, Fordham MT, Preciado D, Reilly BK. Tonsil ultrasound: technical approach and spectrum of pediatric peritonsillar infections. Pediatr Radiol. (2016) 46:1059–67. doi: 10.1007/s00247-015-3505-7
27. Aydin S, Uner C. Normal palatine tonsil size in healthy children: a sonographic study. Radiol Med. (2020) 125:864–9. doi: 10.1007/s11547-020-01168-0
28. Orenc S, Acar E, Ozerdem MS, Sahin S, Kaya A. Automatic identification of adenoid hypertrophy via ensemble deep learning models employing X-ray adenoid images. J Imaging Inform Med. (2025). doi: 10.1007/s10278-025-01423-8
29. Dong W, Chen Y, Li A, Mei X, Yang Y. Automatic detection of adenoid hypertrophy on cone-beam computed tomography based on deep learning. Am J Orthod Dentofacial Orthop. (2023) 163:553–60 e3. doi: 10.1016/j.ajodo.2022.11.011
30. Guo W, Gao Y, Yang Y. Automatic detection of adenoid hypertrophy on lateral nasopharyngeal radiographs of children based on deep learning. Transl Pediatr. (2024) 13:1368–77. doi: 10.21037/tp-24-194
31. Shen Y, Li X, Liang X, Xu H, Li C, Yu Y, et al. A deep-learning-based approach for adenoid hypertrophy diagnosis. Med Phys. (2020) 47:2171–81. doi: 10.1002/mp.14063
32. Rao Y, Zhang Q, Wang X, Xue X, Ma W, Xu L, et al. Automated diagnosis of adenoid hypertrophy with lateral cephalogram in children based on multi-scale local attention. Sci Rep. (2024) 14:18619. doi: 10.1038/s41598-024-69827-0
Keywords: tonsillar, adenoid, artificial intelligence in medicine, ResNet18, YOLOv8, diagnostic imaging
Citation: Wu Z, Zhuo R, Yang Y, Liu X, Wu B and Wang J (2025) Optimized deep learning model for diagnosing tonsil and adenoid hypertrophy through X-rays. Front. Oncol. 15:1508525. doi: 10.3389/fonc.2025.1508525
Received: 14 October 2024; Accepted: 20 February 2025;
Published: 11 March 2025.
Edited by:
Alin Horatiu Nedelcu, Grigore T. Popa University of Medicine and Pharmacy, RomaniaReviewed by:
Mehmet Düzlü, Gazi University, TürkiyeCopyright © 2025 Wu, Zhuo, Yang, Liu, Wu and Wang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Jian Wang, lx97238@sina.com
†These authors have contributed equally to this work
Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.
Research integrity at Frontiers
Learn more about the work of our research integrity team to safeguard the quality of each article we publish.