Skip to main content

ORIGINAL RESEARCH article

Front. Vet. Sci., 06 June 2024
Sec. Veterinary Imaging
This article is part of the Research Topic Cutting-Edge Technology in Veterinary Medicine View all 9 articles

Deep learning models for interpretation of point of care ultrasound in military working dogs

  • 1Organ Support and Automation Technologies Group, U.S. Army Institute of Surgical Research, JBSA Fort Sam Houston, San Antonio, TX, United States
  • 2Hemorrhage Control and Vascular Dysfunction Group, U.S. Army Institute of Surgical Research, JBSA Fort Sam Houston, San Antonio, TX, United States
  • 3Texas A&M University, School of Veterinary Medicine, College Station, TX, United States
  • 4Veterinary Support Group, U.S. Army Institute of Surgical Research, JBSA Fort Sam Houston, San Antonio, TX, United States

Introduction: Military working dogs (MWDs) are essential for military operations in a wide range of missions. With this pivotal role, MWDs can become casualties requiring specialized veterinary care that may not always be available far forward on the battlefield. Some injuries such as pneumothorax, hemothorax, or abdominal hemorrhage can be diagnosed using point of care ultrasound (POCUS) such as the Global FAST® exam. This presents a unique opportunity for artificial intelligence (AI) to aid in the interpretation of ultrasound images. In this article, deep learning classification neural networks were developed for POCUS assessment in MWDs.

Methods: Images were collected in five MWDs under general anesthesia or deep sedation for all scan points in the Global FAST® exam. For representative injuries, a cadaver model was used from which positive and negative injury images were captured. A total of 327 ultrasound clips were captured and split across scan points for training three different AI network architectures: MobileNetV2, DarkNet-19, and ShrapML. Gradient class activation mapping (GradCAM) overlays were generated for representative images to better explain AI predictions.

Results: Performance of AI models reached over 82% accuracy for all scan points. The model with the highest performance was trained with the MobileNetV2 network for the cystocolic scan point achieving 99.8% accuracy. Across all trained networks the diaphragmatic hepatorenal scan point had the best overall performance. However, GradCAM overlays showed that the models with highest accuracy, like MobileNetV2, were not always identifying relevant features. Conversely, the GradCAM heatmaps for ShrapML show general agreement with regions most indicative of fluid accumulation.

Discussion: Overall, the AI models developed can automate POCUS predictions in MWDs. Preliminarily, ShrapML had the strongest performance and prediction rate paired with accurately tracking fluid accumulation sites, making it the most suitable option for eventual real-time deployment with ultrasound systems. Further integration of this technology with imaging technologies will expand use of POCUS-based triage of MWDs.

1 Introduction

Ultrasound is commonly used in canines with suspected abdominal or thoracic injuries following trauma, to identify free fluid which may require surgical intervention. Different standardized exams are used in veterinary medicine such as the abdominal focused assessment with sonography for trauma (AFAST®), thoracic FAST (TFAST®), or the Veterinary Bedside Lung Ultrasound Exam (Vet BLUE®) (13). These are often performed together and referred to as GlobalFAST® which can be used for civilian trauma cases, but also for working dog casualties (4). Working dogs cover a wide range of occupations including military working dogs (MWDs) which go anywhere soldiers are deployed and aid with a wide range of tasks (5). The ever increasing high risk mission that MWDs share with their handlers puts them at risk for similar injuries as their Service member counterparts (6, 7). Unfortunately, in the early roles of care, where MWD casualties are first managed, veterinary expertise may not be present to properly acquire ultrasound images and to interpret images making GlobalFAST® inaccessible for treatment of MWDs at these early stages of care (8).

This is further complicated on the future battlefield where medical evacuation will be limited and more medical care and triage will need to be provided in theater, at early roles of care (9). In fact, this is already being experienced with the Ukraine-Russia conflict, where limited medical evacuation opportunities arise due to challenged airspace, which is requiring far forward surgical teams to treat and manage a larger number of casualties for up to 72 h in theater (10). This is further complicated by precise long-range weaponry minimizing the relative safety of CASEVAC even at distances above 500 km away from enemy lines. In addition, more than 70% of Ukraine casualties stem from more advanced rocket or artillery injuries, which often result in complex polytrauma to multiple organ systems (10). Thus, as we look towards the future battlefield, it is even more imperative to have accurate triage procedures for prioritizing injured warfighters for access to limited evacuation opportunities.

Towards addressing this critical capability gap for canine and human casualties on the future battlefield, artificial intelligence (AI) can be utilized to automate medical triage image interpretation (11, 12). AI for image interpretation often relies on deep convolutional neural network models containing millions of trainable parameters to extract features from images for making categorical predictions (13, 14). For medical applications, AI has been widely used for tumor detection (15, 16), COVID-19 diagnosis (17, 18), and obstetric ultrasound applications (19, 20). In addition, AI has been applied to interpret radiographs in thoracic (21, 22), cardiac (23, 24), and orthopedic (25) settings. Our research team has previously developed an ultrasound image AI interpretation model for detecting shrapnel in tissue, termed ShrapML (26, 27). We have recently expanded this work to the enhanced FAST (eFAST) exam commonly used for human emergency triage applications (28). This application resulted in different AI models for detecting pneumothorax, hemothorax, and abdominal hemorrhage injuries in tissue phantom image sets. In this presented work, we hypothesize if AI image interpretation models are trained on canine image datasets, they will be able to automatically identify injuries at each POCUS scan point. By doing so, the skill threshold for POCUS interpretation will be lowered so that this critical triage task can be available at early echelons of care where emergency intervention is most needed for MWDs.

2 Materials and methods

2.1 Imaging protocol

Research was conducted in compliance with the Animal Welfare Act, implementing Animal Welfare regulations, and the principles of the Guide for the Care and Use for Laboratory Animals. The Institutional Animal Care and Use Committee at the Department of Defense Military Working Dog Veterinary Services approved all research conducted in this study. The facility where this research was conducted is fully accredited by the AAALAC International. The POCUS protocol used mirrored the GlobalFAST® procedure in a total of five (1.5 to 10 years old) healthy canine subjects (20 to 55 kgs weight) under general anesthesia or deep sedation for other medical procedures, as prescribed by the attending veterinarian. Ultrasound (US) clips were collected in 8 scan points (Table 1) using a C11 transducer (Fujifilm, Bothell, WA, United States) with a Sonosite Edge ultrasound system (Fujifilm, Bothell, WA, United States). The subject was positioned in right lateral, left lateral, sternal or dorsal recumbency for ease of access to each scan point. A minimum of three 15 s clips were collected at each scan point with the probe orientation held in the coronal plane for the first 6 s and then rotated to the transverse plane for the remainder of each clip. All clips collected from the live subjects were used as baseline (negative for injury) data. The same scanning protocol was used to obtain US imaging data from a cadaver canine model. A total of five frozen cadavers (Skulls Unlimited, Oklahoma City, OK, United States) were received and stored at −20°C until ready for use. Once thawed, an endotracheal tube (McKesson Medical-Surgical, Irving, TX, United States) was placed into the trachea of each subject and secured to a bag valve mask (EMS Safety Services, Eugene, OR, United States) for ventilation. At this time thoracic and abdominal CT scans (Toshiba Aquilion CT Scanner, Cannon Medical Systems, Tustin, CA, United States) were collected to identify any pre-existing injuries. Then, data was collected at each scan point, using the same protocol as the live subjects. After collecting the first round of data, if the subject was positive for any injury, e.g., a pneumothorax, a needle decompression was performed to remove air and obtain a negative scan. Another round of data was collected with the scan points that were negative for injury. Next, controlled injuries were performed by adding blood or saline to the pleural space (up to 300 mL) or the abdomen (up to 400 mL) for a final round of positive injury image collection in the cadaver subjects.

Table 1
www.frontiersin.org

Table 1. Scan point description for the POCUS imaging protocol.

2.2 Preprocessing images

All clips were exported from the US machine as MP4 format and then renamed to reflect the scan point, subject ID, and recumbency of each subject. Frames were extracted from each clip using ffmpeg tool, via a Ruby script, and then sorted by positive or negative for injury by scan point. Each frame was then cropped to remove the user interface information from the US system and the images were resized to 512 × 512 pixels. Additional steps were taken with images collected at the chest tube site, to recreate M-mode images. Briefly, clips were processed to extract a pixel-wide image over time for visualizing the lung-pleura interface movement. These custom-M-mode images were then cropped and resized to 512 × 512 as well.

Before images were ready for training, they were augmented to prevent model overfitting and improve performance. While data augmentation is useful to prevent overfitting, it can result in poor model performance and more computationally intensive training if not setup optimally for the application (29). A representative image was chosen from each scan point, including M-mode reconstructions, to match histogram values across all the other images using “imhistmatch” function by MATLAB (MathWorks, Natick, MA, United States). Then, contrast and brightness were randomly adjusted by ±20% to add training noise using the “jitterColorHSV” function by MATLAB. Both MATLAB functions were applied to all images for every scan point using Image Batch Processor on MATLAB. Augmented US images were imported at a 512 × 512 × 3 image size and were randomly assigned to training, validation or testing datasets at a 70:15:15 ratio. Image sets were set up so that an even number of positive or negative images were selected in each dataset for each split. Next, training images were augmented randomly by affine transformations: random scaling, random X and Y reflections, random rotation, random X and Y shear, and random X and Y translation. However, for the CTS M-mode scan point only X reflection and translation affine transformations were applied given how these images were constructed. Due to DH scan point images being unable to train with all augmentations (data not shown), only reflection and translation augmentations were applied for both the X and Y direction.

2.3 Training AI models

Three different AI models were evaluated for this application that have previously been used for ultrasound image interpretation successfully – MobileNetV2 (30), DarkNet-19 (31), and ShrapML (26). MobileNetV2 has 53 convolutional layers, 3.5 million parameters, and was optimized for use on mobile devices. We have previously shown this architecture to perform at the highest accuracy for identifying shrapnel in a custom tissue phantom. The second-best performing architecture, DarkNet-19, has 19 convolutional layers, 20.8 million parameters, and utilizes global average pooling for making predictions. The last model used, ShrapML, was purpose built and Bayesian optimized for identifying shrapnel in ultrasound images at a high accuracy and much more rapid than conventional models. In addition, we have shown it to be successful at identifying pneumothorax, hemothorax, and abdominal hemorrhage injuries in eFAST images captured in human tissue phantom models (28). ShrapML consists of 8 convolutional layers with only 430,000 trainable parameters.

Training for all scan points consisted of a learning rate of 0.001 with a batch size of 32 images and RMSprop (root mean squared propagation) as the optimizer. A maximum of 100 epochs was allowed for training with a validation patience of 5 epochs if the overall validation loss did not improve. The model with the lowest validation loss was selected for use with blind predictions. All training was performed using MATLAB R2022b run on a Microsoft Windows workstation with a NVIDIA GeForce RTX 3090 Ti 24Gb VRAM graphics card, Intel i9-12900k and 64 GB RAM.

2.4 Performance metrics

Testing image sets were used to assess blind performance in multiple ways. First, confusion matrices were generated to categorize prediction as either true positive (TP), true negative (TN), false positive (FP), or false negative (FN) results. These results were used to generate performance metrics for accuracy Eq. 1, precision Eq. 2, recall Eq. 3, specificity Eq. 4, and F1 scores Eq. 5 using commonly used formulas for each.

Accuracy = T P + T N T P + T N + F P + F N     (1)
Precision = T P T P + F P     (2)
Recall = T P T P + F N     (3)
Specificity = T N T N + F P     (4)
F 1 score = 2 × Precision × Recall Precision + Recall     (5)

Then, we constructed receiver operating characteristic (ROC) plots to further classify performance for a number of confidence thresholds for the predictions. ROC plots were used to calculate the area under the ROC curve or AUROC, which tells you how well the model differentiates between categories. Next, inference time for test image predictions were quantified for each trained model to assess differences in computational efficiency of the three different AI models used. Lastly, Gradient-weighted Class Activation Mapping (GradCAM) overlays were generated for test predictions to highlight the regions of images where the AI predictions were focused (32). These were used as an explainable-AI methodology to verify the AI models were accurately tracking the image regions where injury differences were present (16, 33, 34).

3 Results

3.1 MobileNetV2

MobileNetV2 was successfully trained for each POCUS scan point, with an average accuracy across all locations of 98.8% (Table 2). In addition, strong performance was evident for other conventional metrics across each POCUS scan point. However, upon closer inspection using GradCAM mask overlays, the MobileNetV2 trained model was not always properly tracking the injury site, but instead was focused on image artifacts that will likely not be consistent for additional canine subjects not included in the current datasets (Figure 1). CTS scan sites for both M- and B-mode were accurately tracking injuries, other scan sites such as HR, DH, and SR were not tracking correctly. Average inference times across all MobileNetV2 scan site models was 6.21 ms per prediction.

Table 2
www.frontiersin.org

Table 2. Summary of performance metrics for MobileNetV2.

Figure 1
www.frontiersin.org

Figure 1. Prediction results by scan point for MobileNetV2. Results for each scan site showing (column 1) confusion matrix test prediction results, (column 2–3) negative and (column 4–5) positive representative images without and with the GradCAM overlay. Regions in the images with high relevance to model predictions have red-yellow overlays, while those of lower relevance have blue-green overlays.

3.2 DarkNet-19

The DarkNet-19 models had similar inference speeds compared to MobileNetV2 at 5.93 ms per prediction, but overall performance was reduced for a number of the scan sites, resulting in an average accuracy across all scan points of 86.4% (Table 3). Certain scan points like chest-tube M-mode images resulted only in predictions of negative (TN or FN) and the GradCAM overlays identified no obvious tracked features in the image (Figure 2). While this was the worst performing dataset trained against, the Cystocolic scan site was also only at 69.2% accuracy. While performance was reduced compared to MobileNetV2 across nearly all metrics, the GradCAM overlays were more accurately tracking image features consistent with locations where free fluid was or could be identified. These results indicated that while performance was overall reduced for DarkNet-19, the predictions were more often tracking the proper image features. More images and subject variability may improve on training performance.

Table 3
www.frontiersin.org

Table 3. Summary of performance metrics for DarkNet-19.

Figure 2
www.frontiersin.org

Figure 2. Prediction results by scan point for DarkNet-19. Results for each scan site showing (column 1) confusion matrix test prediction results, (column 2–3) negative and (column 4–5) positive representative images without and with the GradCAM overlay. Regions in the images with high relevance to model predictions have red-yellow overlays, while those of lower relevance have blue-green overlays.

3.3 ShrapML

The last model evaluated was ShrapML, which resulted in an accuracy across all scan sites of 93.4% (Table 4). Unlike DarkNet-19, no trained model resulted in an instance of 100% positive or negative guesses. However, performance metrics were consistently worse than MobileNetV2. Given the smaller model size of ShrapML, the inference times were much quicker compared to the other models with prediction rates at an average of 3.43 ms per image. GradCAM overlays more closely resembled DarkNet-19 in that many of the heat map intensity points were focused on regions where free fluid was likely to be found or near organs present in the ultrasound scan (Figure 3), except for the HR site. Overall, ShrapML was successful at performing similarly well to these large network structures for this GlobalFAST application, model overfitting was less evident in the results, and overall prediction speed outperformed the other models tested.

Table 4
www.frontiersin.org

Table 4. Summary of performance metrics for ShrapML.

Figure 3
www.frontiersin.org

Figure 3. Prediction results by scan point for ShrapML. Results for each scan site showing (column 1) confusion matrix test prediction results, (column 2–3) negative and (column 4–5) positive representative images without and with the GradCAM overlay. Regions in the images with high relevance to model predictions have red-yellow overlays, while those of lower relevance have blue-green overlays.

A summary table of average performance metrics for each scan site across all three model architectures is shown in Table 5.

Table 5
www.frontiersin.org

Table 5. Summary of performance metrics for each POCUS site.

4 Discussion

Medical imaging-based triage is critical for both human and veterinary emergency medicine to identify issues early on and ensure resources are properly distributed. In remote or military medicine situations, the lack of skilled personnel makes imaging based-triage less relied upon, but AI prediction models can simplify this for the end user. Here, we focus on the POCUS procedure GlobalFAST®, a widely used triage exam to look for abdominal or thoracic free fluid in injured dogs. The AI models shown in this work can automate predictions for ultrasound results if properly tuned for the application.

Three different AI architectures were evaluated to see which was capable of being trained to distinguish positive injury cases from baseline images. While all models were generally successful at being trained for these applications, strong test performance may not indicate properly trained models. For instance, MobileNetV2 had the highest accuracy, but heat map overlays indicating where the AI was focused were not tracking proper image locations. Model overfit was combatted with the various image augmentation techniques used for the training, but this was insufficient to mimic proper subject variability to create a more robust model for this architecture. This issue was less evident for the other two model architectures, highlighting the importance of AI model selection and validation on ultrasound image applications such as this. However, without more subjects and the variability that those bring, it is hard to fully verify if the developed DarkNet-19 or ShrapML models are suitable. Preliminarily, ShrapML had the strongest performance and prediction rate, making it the most suitable going forward as well as eventual integration for real-time deployment with ultrasound machines.

Focusing on the various scan points in the used POCUS exam, there were obvious differences in the AI model training. Training image sets were not equally sized, but that did not correlate to what scan sites performed the best. The DH site was the overall strongest performing site across all performance metrics. However, this could be due to this scan site having the largest difference between live and cadaveric tissue resulting in a well-trained model. In addition, less augmentation steps were used for this site due to training issues using all affine transformations. More images are needed to address this issue from a wider range of subjects. CTS and HR views also performed well across the three models trained. Worst performing was the M-mode reconstructed chest tube images which could be influenced by the minimal training data used for this model, and thus may be improved with more training data. The CC site was also a lower performing scan site even though more than 10,000 images were used in the training dataset. However, this is mostly influenced by DarkNet-19 having lower performance for this scan site while the other two models had accuracies greater than 96%. Overall, each scan site for this POCUS application was successful as an input for an injury prediction model.

5 Conclusion

Artificial intelligence has the potential to simplify triage and injury diagnosis for emergency veterinary medicine. The results shown in this work highlight how AI can be used for automating US detection of intrabdominal and intrathoracic injury detection for veterinary applications. Each scan point reached greater than 80% injury detection accuracy, with most surpassing 90% accuracy. However, more data is still needed to be able to ensure that the AI models are not overfitting the training data and can accurately predict for new subject data. Next steps for this work will expand training datasets so that blind subject testing is possible for confirming generalized models are developed. With more data, these models can be set up for real-time integration with ultrasound devices allowing for early detection of thoracic and abdominal injuries for military working dogs and other canine trauma situations. This will lower the skill threshold for medical imaging-based triage so that these techniques can be more widely used.

Data availability statement

The datasets presented in this article are not readily available because they have been collected and maintained in a government-controlled database that is located at the US Army Institute of Surgical Research. As such, this data can be made available through the development of a Cooperative Research & Development Agreement (CRADA) with the corresponding author. Requests to access the datasets should be directed to ES, eric.j.snider3.civ@health.mil.

Ethics statement

The animal study was approved by Research was conducted in compliance with the Animal Welfare Act, the implementing Animal Welfare regulations, and the principles of the Guide for the Care and Use for Laboratory Animals. The Institutional Animal Care and Use Committee at the Department of Defense Military Working Dog Veterinary Services approved all research conducted in this study. The facility where this research was conducted is fully accredited by the AAALAC International. The study was conducted in accordance with the local legislation and institutional requirements.

Author contributions

SH: Conceptualization, Data curation, Formal analysis, Methodology, Visualization, Writing – original draft, Writing – review & editing. LH: Data curation, Formal analysis, Methodology, Software, Writing – review & editing. TE: Funding acquisition, Writing – review & editing. EV: Writing – original draft, Methodology, Funding acquisition, Data curation, Conceptualization. ES: Conceptualization, Data curation, Formal analysis, Methodology, Writing – original draft, Writing – review & editing.

Funding

The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. This work was funded through the Restoral program by the Defense Health Agency. This project was supported in part by an appointment to the Science Education Programs at National Institutes of Health (NIH), administered by ORAU through the U.S. Department of Energy Oak Ridge Institute for Science and Education (LH).

Acknowledgments

The authors would like to acknowledge Dr. Joanna Hourani, MAJ (Dr.) Richard Brooksby, and MAJ (Dr.) Erin Hennessey for their assistance with image capture in Military Working Dogs.

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher’s note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

Author disclaimer

The views expressed in this article are those of the authors and do not reflect the official policy or position of the U.S. Army Medical Department, Department of the Army, DOD, or the U.S. Government.

References

1. Boysen, SR, and Lisciandro, GR. The use of ultrasound for dogs and cats in the emergency room: AFAST and TFAST. Vet Clin North Am Small Anim Pract. (2013) 43:773–97. doi: 10.1016/j.cvsm.2013.03.011

Crossref Full Text | Google Scholar

2. Cole, L, Pivetta, M, and Humm, K. Diagnostic accuracy of a lung ultrasound protocol (vet BLUE) for detection of pleural fluid, pneumothorax and lung pathology in dogs and cats. J Small Anim Pract. (2021) 62:178–86. doi: 10.1111/jsap.13271

Crossref Full Text | Google Scholar

3. Boatright, K. (2020). Up your imaging game: The power of AFAST 52. Available at: https://www.dvm360.com/view/up-your-imaging-game-the-power-of-afast (Accessed January 5, 2024).

Google Scholar

4. Lisciandro, GR, and Lisciandro, SC. Global FAST for patient monitoring and staging in dogs and cats. Vet Clin North Am Small Anim Pract. (2021) 51:1315–33. doi: 10.1016/j.cvsm.2021.07.011

Crossref Full Text | Google Scholar

5. Green, R. (2021). What do military working dogs do? Am. Kennel Club Available at: https://www.akc.org/expert-advice/news/what-are-military-working-dogs/ (Accessed January 5, 2024).

Google Scholar

6. Edwards, T, Scott, LLF, Gonyeau, KE, Howard, EH, Parker, JS, and Hall, K. Comparison of trauma sustained by civilian dogs and deployed military working dogs. J Vet Emerg Crit Care (San Antonio). (2021) 31:498–507. doi: 10.1111/vec.13064

Crossref Full Text | Google Scholar

7. McGraw, AL, and Thomas, TM. Military working dogs: an overview of veterinary Care of these Formidable Assets. Vet Clin North Am Small Anim Pract. (2021) 51:933–44. doi: 10.1016/j.cvsm.2021.04.010

Crossref Full Text | Google Scholar

8. Lagutchik, M, Baker, J, Balser, J, Burghardt, W, Enroth, M, Flournoy, S, et al. Trauma Management of Military Working Dogs. Mil Med. (2018) 183:180–9. doi: 10.1093/milmed/usy119

Crossref Full Text | Google Scholar

9. Townsend, S, and Lasher, W. The U.S. Army in multi-domain operations 2028. Arlington, VA, USA: U.S. Army (2018).

Google Scholar

10. Epstein, A, Lim, R, Johannigman, J, Fox, CJ, Inaba, K, Vercruysse, GA, et al. Putting medical boots on the ground: lessons from the war in Ukraine and applications for future conflict with near-peer adversaries. J Am Coll Surg. (2023) 237:364–73. doi: 10.1097/XCS.0000000000000707

Crossref Full Text | Google Scholar

11. Latif, J., Xiao, C., Imran, A., and Tu, S. (2019). Medical imaging using machine learning and deep learning algorithms: a review, in 2019 2nd international conference on computing, mathematics and engineering technologies (iCoMET), (IEEE), 1–5.

Google Scholar

12. Liu, R, Rong, Y, and Peng, Z. A review of medical artificial intelligence. Glob Health J. (2020) 4:42–5. doi: 10.1016/j.glohj.2020.04.002

Crossref Full Text | Google Scholar

13. Komatsu, M, Sakai, A, Dozen, A, Shozu, K, Yasutomi, S, Machino, H, et al. Towards clinical application of artificial intelligence in ultrasound imaging. Biomedicines. (2021) 9:720. doi: 10.3390/biomedicines9070720

Crossref Full Text | Google Scholar

14. Liu, S, Wang, Y, Yang, X, Lei, B, Liu, L, Li, SX, et al. Deep learning in medical ultrasound analysis: a review. Engineering. (2019) 5:261–75. doi: 10.1016/j.eng.2018.11.020

Crossref Full Text | Google Scholar

15. Chiang, T-C, Huang, Y-S, Chen, R-T, Huang, C-S, and Chang, R-F. Tumor detection in automated breast ultrasound using 3-D CNN and prioritized candidate aggregation. IEEE Trans Med Imaging. (2019) 38:240–9. doi: 10.1109/TMI.2018.2860257

Crossref Full Text | Google Scholar

16. Song, D, Yao, J, Jiang, Y, Shi, S, Cui, C, Wang, L, et al. A new xAI framework with feature explainability for tumors decision-making in ultrasound data: comparing with grad-CAM. Comput Methods Prog Biomed. (2023) 235:107527. doi: 10.1016/j.cmpb.2023.107527

Crossref Full Text | Google Scholar

17. Diaz-Escobar, J, Ordóñez-Guillén, NE, Villarreal-Reyes, S, Galaviz-Mosqueda, A, Kober, V, Rivera-Rodriguez, R, et al. Deep-learning based detection of COVID-19 using lung ultrasound imagery. PLoS One. (2021) 16:e0255886. doi: 10.1371/journal.pone.0255886

Crossref Full Text | Google Scholar

18. Gil-Rodríguez, J, Pérez de Rojas, J, Aranda-Laserna, P, Benavente-Fernández, A, Martos-Ruiz, M, Peregrina-Rivas, J-A, et al. Ultrasound findings of lung ultrasonography in COVID-19: a systematic review. Eur J Radiol. (2022) 148:110156. doi: 10.1016/j.ejrad.2022.110156

Crossref Full Text | Google Scholar

19. Baumgartner, CF, Kamnitsas, K, Matthew, J, Fletcher, TP, Smith, S, Koch, LM, et al. SonoNet: real-time detection and localisation of Fetal standard scan planes in freehand ultrasound. IEEE Trans Med Imaging. (2017) 36:2204–15. doi: 10.1109/TMI.2017.2712367

Crossref Full Text | Google Scholar

20. Iriani Sapitri, A, Nurmaini, S, Naufal Rachmatullah, M, Tutuko, B, Darmawahyuni, A, Firdaus, F, et al. Deep learning-based real time detection for cardiac objects with fetal ultrasound video. Inform Med Unlocked. (2023) 36:101150. doi: 10.1016/j.imu.2022.101150

Crossref Full Text | Google Scholar

21. Banzato, T, Wodzinski, M, Burti, S, Osti, VL, Rossoni, V, Atzori, M, et al. Automatic classification of canine thoracic radiographs using deep learning. Sci Rep. (2021) 11:3964. doi: 10.1038/s41598-021-83515-3

Crossref Full Text | Google Scholar

22. Müller, TR, Solano, M, and Tsunemi, MH. Accuracy of artificial intelligence software for the detection of confirmed pleural effusion in thoracic radiographs in dogs. Vet Radiol Ultrasound. (2022) 63:573–9. doi: 10.1111/vru.13089

Crossref Full Text | Google Scholar

23. Kim, E, Fischetti, AJ, Sreetharan, P, Weltman, JG, and Fox, PR. Comparison of artificial intelligence to the veterinary radiologist’s diagnosis of canine cardiogenic pulmonary edema. Vet Radiol Ultrasound. (2022) 63:292–7. doi: 10.1111/vru.13062

Crossref Full Text | Google Scholar

24. Li, S, Wang, Z, Visser, LC, Wisner, ER, and Cheng, H. Pilot study: application of artificial intelligence for detecting left atrial enlargement on canine thoracic radiographs. Vet Radiol Ultrasound. (2020) 61:611–8. doi: 10.1111/vru.12901

Crossref Full Text | Google Scholar

25. McEvoy, FJ, Proschowsky, HF, Müller, AV, Moorman, L, Bender-Koch, J, Svalastoga, EL, et al. Deep transfer learning can be used for the detection of hip joints in pelvis radiographs and the classification of their hip dysplasia status. Vet Radiol Ultrasound. (2021) 62:387–93. doi: 10.1111/vru.12968

Crossref Full Text | Google Scholar

26. Boice, EN, Hernandez-Torres, SI, and Snider, EJ. Comparison of ultrasound image classifier deep learning algorithms for shrapnel detection. J Imaging. (2022) 8:140. doi: 10.3390/jimaging8050140

Crossref Full Text | Google Scholar

27. Snider, EJ, Hernandez-Torres, SI, and Boice, EN. An image classification deep-learning algorithm for shrapnel detection from ultrasound images. Sci Rep. (2022) 12:8427. doi: 10.1038/s41598-022-12367-2

Crossref Full Text | Google Scholar

28. Hernandez-Torres, SI, Bedolla, C, Berard, D, and Snider, EJ. An extended focused assessment with sonography in trauma ultrasound tissue-mimicking phantom for developing automated diagnostic technologies. Front Bioeng Biotechnol. (2023) 11:1244616. doi: 10.3389/fbioe.2023.1244616

Crossref Full Text | Google Scholar

29. Xu, Y., Noy, A., Lin, M., Qian, Q., Li, H., and Jin, R. (2020). Wemix: How to better utilize data augmentation. arXiv preprint. arXiv:2010.01267. Available at: https://arxiv.org/abs/2010.01267

Google Scholar

30. Sandler, M., Howard, A., Zhu, M., Zhmoginov, A., and Chen, LC. (2018). Mobilenetv2: Inverted residuals and linear bottlenecks. In: Proceedings of the IEEE conference on computer vision and pattern recognition. IEEE publisher. 4510–4520. Available at: https://ieeexplore.ieee.org/abstract/document/8578572

Google Scholar

31. Redmon, J., and Farhadi, A. (2016). YOLO9000: Better, Faster, Stronger. ArXiv161208242 Cs. Available at: http://arxiv.org/abs/1612.08242 (Accessed April 22, 2022).

Google Scholar

32. Selvaraju, R. R., Cogswell, M., Das, A., Vedantam, R., Parikh, D., and Batra, D. (2017). Grad-cam: visual explanations from deep networks via gradient-based localization, in Proceedings of the IEEE international conference on computer vision, 618–626.

Google Scholar

33. Hsu, S-T, Su, Y-J, Hung, C-H, Chen, M-J, Lu, C-H, and Kuo, C-E. Automatic ovarian tumors recognition system based on ensemble convolutional neural network with ultrasound imaging. BMC Med Inform Decis Mak. (2022) 22:298. doi: 10.1186/s12911-022-02047-6

Crossref Full Text | Google Scholar

34. Yang, J, Shi, X, Wang, B, Qiu, W, Tian, G, Wang, X, et al. Ultrasound image classification of thyroid nodules based on deep learning. Front Oncol. (2022) 12:905955. doi: 10.3389/fonc.2022.905955

Crossref Full Text | Google Scholar

Keywords: ultrasound imaging, military medicine, canine, deep learning, triage, abdominal hemorrhage, pneumothorax, hemothorax

Citation: Hernandez Torres SI, Holland L, Edwards TH, Venn EC and Snider EJ (2024) Deep learning models for interpretation of point of care ultrasound in military working dogs. Front. Vet. Sci. 11:1374890. doi: 10.3389/fvets.2024.1374890

Received: 23 January 2024; Accepted: 20 May 2024;
Published: 06 June 2024.

Edited by:

Blaz Cugmas, University of Latvia, Latvia

Reviewed by:

Juan Claudio Gutierrez, University of California, Davis, United States
Martin Ceballos, University of Buenos Aires, Argentina

Copyright © 2024 Hernandez Torres, Holland, Edwards, Venn and Snider. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Eric J. Snider, eric.j.snider3.civ@health.mil

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.