Skip to main content

ORIGINAL RESEARCH article

Front. Radiol., 21 November 2024
Sec. Artificial Intelligence in Radiology

SenseCare: a research platform for medical image informatics and interactive 3D visualization

\r\nGuotai Wang,Guotai Wang1,2Qi DuanQi Duan2Tian ShenTian Shen2Shaoting Zhang,
\r\nShaoting Zhang1,2*
  • 1School of Mechanical and Electrical Engineering, University of Electronic Science and Technology of China, Chengdu, China
  • 2SenseTime Research, Shanghai, China

Introduction: Clinical research on smart health has an increasing demand for intelligent and clinic-oriented medical image computing algorithms and platforms that support various applications. However, existing research platforms for medical image informatics have limited support for Artificial Intelligence (AI) algorithms and clinical applications.

Methods: To this end, we have developed SenseCare research platform, which is designed to facilitate translational research on intelligent diagnosis and treatment planning in various clinical scenarios. It has several appealing functions and features such as advanced 3D visualization, concurrent and efficient web-based access, fast data synchronization and high data security, multi-center deployment, support for collaborative research, etc.

Results and discussion: SenseCare provides a range of AI toolkits for different tasks, including image segmentation, registration, lesion and landmark detection from various image modalities ranging from radiology to pathology. It also facilitates the data annotation and model training processes, which makes it easier for clinical researchers to develop and deploy customized AI models. In addition, it is clinic-oriented and supports various clinical applications such as diagnosis and surgical planning for lung cancer, liver tumor, coronary artery disease, etc. By simplifying AI-based medical image analysis, SenseCare has a potential to promote clinical research in a wide range of disease diagnosis and treatment applications.

1 Introduction

With the development of medical imaging techniques and computer science, computer-aided systems for medical image analysis and downstream diagnosis and treatment decision have been playing an increasing role in clinic practices. In recent years, Artificial Intelligence (AI) has lead to a revolution of image analysis and pattern recognition, and has a huge potential to be applied for more efficient and intelligent medical image computing in a wide range of medical departments towards smart healthcare. However, before AI is used in clinic practice, extensive research studies are needed through the collaboration between clinicians, radiologists, pathologists, surgeons, AI scientists and engineers, which can validate the effectiveness, robustness, reliability and security of AI systems. To this end, a research platform that supports different medical image processing tasks and intelligent medical image computing algorithms for various clinic applications are highly desirable.

Despite the availability of several existing platforms for developing AI algorithms for general image recognition or medical image computing, they are not clinic-oriented and have limited support for clinical research. For example, TensorFlow (1), Pytorch (2) and Keras (3) are general deep learning libraries that provide low-level functions to develop complex deep learning models without specific functionality for medical image computing. Some other libraries such as NiftyNet (4), DLTK (5) and PyMIC (6) are developed for deep learning with medical images, but they are mainly designed for AI algorithm developers and there is no graphic user interface, which is difficult for clinicians and radiologists to use in specific clinic applications.

On the other hand, several medial image analysis solutions have also been developed in the past decades. Tools that solve a specific medical image processing task such as segmentation (7), registration (8, 9) and visualization (10) can be used for a part of a clinic application pipeline, but still not ready-to-use for clinic researchers. Some research platforms such as MITK (11), NifTK (12) and 3D Slicer (13) provide several traditional medical image analysis tools and 3D visualization for image guided intervention. However, these platforms have limited support for AI models. Recently, some AI-based plugins have been added to 3D Slicer for image segmentation tasks, but the AI models for other tasks are limited, and it does not support model training and dealing with pathological images. Recently, some virtual reality visualization platforms such as COVI3D (14, 15) have been proposed for quantification and interpretation of lung tumor and lesions (16, 17) with the help of AI models, but their extensibility to other applications is limited. In the era of multi-modal medical data, a desired clinical research platform based on AI does not only need to provide AI models for various clinical applications, but also require data exchange with other image management systems, functionality across diverse image modalities, availability of sophisticated 3D visualization and high extensibility and portability for different clinic scenarios.

To solve these problems, we introduce the SenseCare Smart Health Platform (SenseCare for short) that aims to take advantages of state-of-the-art AI techniques to foster researchers from different clinical departments to implement innovations for improvement through the whole process of clinical diagnosis, treatment planning and rehabilitation management. Compared with some existing medical image computing and visualization platforms, it has the following functionalities and advantages: (1) Integrating a wide range of AI algorithms based on deep learning for extensive image processing tasks, including classification, detection, segmentation, registration and 3D visualization; (2) Supporting various data modalities ranging from structural data like radiological and pathological images to time series data like ECG, with real-time data synchronization among different systems; (3) Offering web-based access and multi-center deployment with high concurrency across different devices and operating systems; (4) Facilitating collaborative research for various clinical applications. SenseCare has increasingly assisted to achieve outputs in several research projects including quantitative analysis of cardiac function (18), assessment of knee articular cartilages (19), pathological image analysis (20, 21), lung cancer diagnosis (22, 23), quantitative brain tumor assessment (24, 25), spine image analysis (26, 27) and radiotherapy planning for head and neck cancers (28, 29), etc.

The following sections of this paper are organized as follows: In Section 2.1, we give a brief summary of the architecture of SenseCare, which is followed by detailed descriptions of basic functional modules in Section 2.2. We then introduce SenseCare’s AI toolkits in Section 2.3. In Section 3, we show several examples of clinical applications powered by SenseCare. Finally, discussions and conclusions are given in Sections 4, 5, respectively.

2 Material and methods

2.1 Architecture of sensecare

As shown in Figure 1, SenseCare provides a wide range of artificial intelligence algorithms based on deep learning for learning from and analyze different kinds of medical data. It also provides advanced visualization of medical images that enables users to analyze complex anatomies and segmented structures. These modules are combined with a browser/server architecture and multi-center deployment so that they are accessible on different kinds of devices and at various locations.

Figure 1
www.frontiersin.org

Figure 1. Architecture of SenseCare platform.

The architecture of SenseCare follows a modular structure that consists of three layers: (1) basic functional modules such as data management and visualization, web-based access and multi-center deployment, (2) advanced AI toolkits that include libraries for model training and many built-in AI algorithms for image segmentation, registration, lesion detection, etc., and (3) application scenarios that adapt the basic functional modules and AI algorithms to deal with different clinical tasks such as computer assisted diagnosis of the lung and surgical planning for bone tumors.

2.2 Basic functional modules

2.2.1 Data support

Support of various imaging modalities. SenseCare supports intelligent analysis of in images in various modalities ranging from radiological images to pathological images. For radiological images, it allows efficient import, query, retrieval, and storage of clinical images using DICOM protocols and structures. Major radiological images including Computed Tomography (CT), Magnetic Resonance Imaging (MRI), Digital Radiology (DR) and Positron Emission Tomography (PET) are all well supported by SenseCare in different clinical applications. For pathological images, SenseCare supports several image formats including SVS, TIFF, VMS, NPDI, KFB and others. It also provides a series of functions of import, query, retrieval, storage, management, common measurements and analysis to help pathologists perform diagnosis in a more efficient and intuitive way.

Data synchronization and security. Medical data are commonly stored in different systems such as the Picture Archiving and Communication System (PACS) and Radiology Information System (RIS). Synchronization of data between these systems and the image computing workstation is critical in a clinical research environment. To facilitate more efficient and functional workflow for clinical researchers and medical practitioners, SenseCare can be seamlessly integrated into existing information systems in the hospital, and provides users with improved access by efficient data synchronization. It is capable of synchronizing data from PACS, RIS and other common information systems in hospitals without disturbing the original clinical workflows. When newly acquired data are transmitted or changes in status take place in these information systems, SenseCare will synchronize the updated information and present the users with the latest information automatically. It also supports pulling data directly from PACS/RIS based on user-defined rules. For example, users can designate image modality and type or time range, and send queries to fetch the data they want from a database.

Several strategies are employed by SenseCare to ensure data security in clinical environment. First, the sever of SenceCare is located in the hospital, and is only accessible in the local area network within the hospital, avoiding the risk of leaking data to the internet. Secondly, the cloud-based image computing and visualization in SenseCare avoids data transmission from the server to users’ devices, and only the outputs of image analysis and rendering are sent to the client side. Thirdly, encryption algorithms are embedded in SenseCare to reliably manage and protect user accounts and data from potential risk factors.

2.2.2 Advanced visualization

SenseCare provides advanced 3D reconstruction and visualization of medical images to facilitate the analysis of complex anatomies and segmented structures, which presents data and information in a more explicit way thus improving the information interpretation. Comprehensive methods such as Maximum Intensity Projection (MIP), Minimum Intensity Projection (MinIP), Multi-Planar Reconstruction (MPR), Curved Planar Reformation (CPR) and 3D volume rendering are available for users to perform 3D visualization and enhance interactivity. These capabilities play an important role in clinical diagnosis meanwhile contributing to surgical planning, simulation and navigation, as well as radiotherapy planning, etc, as shown in Figures 3c, 4.

2.2.3 Concurrent and efficient web-based access

SenseCare is designed with high performance concurrency. The network communication structure of SenseCare enables hundreds of users to concurrently perform high-resolution image rendering and 3D post-progressing thus satisfying multiple usage needs. With servers deployed, SenseCare platform allows over 1,000 concurrent users to review and retrieve medical images while its comprehensive toolkits for advanced image post-processing are designed to support more than 160 users simultaneously.

SenseCare can be used in different devices and operating systems and no plugins are needed. By adopting a browser/server architecture, it offers a truly seamless user experience and eliminates the need for multiple logons. Users can access not only MPR, MIP/MinIP, CPR tools, but also the full range of three-dimensional capabilities through HTML5 websites. SenseCare grants radiologists and physicians with efficient access and workflow-boosting benefits even when working from iPad, smartphone or laptop.

2.2.4 Multi-center deployment

Collecting data from multiple centers is important for developing robust algorithms and large-scale clinic studies. To support such research that requires a collaboration between researchers from different locations, SenseCare can be deployed at multiple centers. This favorable feature distinguishes SenseCare from traditional image computing workstations that are located at a single institution.

In addition, SenseCare’s multi-center deployment facilitates the data collection and access process. Various types of data from different centers can be collected and cleaned under standardized rules. Data scientists, clinical researchers, and pharmacists are all able to participate in the process while respecting the original clinical process.

2.2.5 Support for collaborative research

To facilitate the collaboration between researchers, SenseCare provides a specific Document Management System (DMS) to make it easier to organize, secure, capture, digitize, tag, approve and complete tasks with research-related files. It can handle a large amount of papers and images and cope with different workflows by supporting advanced user permission management and task management. This efficient tool enables researchers to manage multiple sophisticated research projects easily, and helps a ground of researchers collaborate with each other and exchange knowledge for efficient accomplishment of comprehensive research projects.

For user permission management, SenseCare supports different levels of users. Admin users can create normal users and set permissions for other normal users to access, view, manipulate, share and remove folders and files. Nonetheless, Admin users can change the permission if needed. For task management, SenseCare’s DMS allows admin users to break down tasks and delegate tasks to normal users. It also allows users to track task progress from the beginning to the end and set small milestones to make sure the whole project will be finished on time. Users can prioritize, organize and set deadlines for themselves and are able to draw together the resources they need to achieve their research goals.

2.3 Artificial intelligence toolkits

Recent years have seen a fast growing of novel deep learning algorithms for medical image computing tasks (30), which play an important role for more accurate and efficient diagnosis and treatment planning and assessment. In this section, we introduce SenseCare’s AI toolkits for medical image computing, including tools for users to develop new algorithms, and built-in deep learning models for image segmentation, registration and lesion detection, etc.

2.3.1 Data annotation and model training

Deep learning models require a large amount of annotated data for training, and both annotation of medical images and model training are time-consuming and complex for common clinical researchers. To facilitate developing AI models for a range of medical image analysis tasks, SenseCare not only provides AI-assisted image annotation tools to improve annotation efficiency, but also offers a framework to train deep learning models easily, and support docker integration to avoid complex environment configuration by users.

Efficient data annotation tools. To facilitate the annotation of a large set of training images for developing AI models, SenseCare provides a set of off-the-shelf tools for efficient image annotation, such as contouring tumors and organs for segmentation tasks, bounding box annotation for object and lesion detection tasks. The annotation tools in SenseCare are available for both radiological images and pathological images, and they provide a variety of annotation types. For example, users can choose from different interactive styles including mouse click points, rectangles, circles, ellipses, polygons or hand-drawn shapes based on the characteristics of the target.

Since manual annotation is time-consuming and annotation may vary from different annotators’ inputs, SenseCare also supports semi-automatic annotation with minimal user interactions and high efficiency and accuracy. For medical image segmentation, the interactive annotation tool not only generates high-accuracy boundaries based on the click/bounding box prompt given by the user (31), but also automatically suggests the most informative samples for annotation (32), which largely reduces the annotation burdens and improves the efficiency. For object detection tasks, the annotator can start from annotations automatically generated by algorithms, and only needs to provide few interactions to obtain refined annotations. As an example, Figure 2 shows efficient annotation of signet ring cell carcinoma in pathological images using SenseCare, where algorithms suggested some annotations in Figure 2a, and the annotator only provides manual refinement to obtain accurate annotations for a slide in Figure 2b.

Figure 2
www.frontiersin.org

Figure 2. Annotation of signet ring cell carcinoma. (a) Suggested annotations. (b) Manual refinement.

Model training. SenseCare’s AI toolkits are constructed on the basis of SenseParrots, a deep learning framework independently developed by SenseTime. While matching mainstream frameworks such as TensorFlow (1) and PyTorch (2), it also has advantages in ultra-deep networks and ultra-large-scale data model training. The integrated optimization from the underlying layer to the SenseParrots framework enables the platform to outperform others under the same level of configuration. By using more than 1,000 GPU cards for parallel training, it takes less than 1.5 min to complete the AlexNet neural network training in a benchmark task, surpassing the previous fastest 4 min record (33).

For medical image analysis, SenseCare supports a wide range of neural network structures including CNN-based and ViT-based networks for image classification, detection and segmentation, etc. Considering that the there are a large amount of unannotated images or imperfect annotations in many applications, SenseCare also provides Semi-Supervised Learning (SSL) and Weakly Supervised Learning (WSL) methods for training. Especially, SSL leverages a small set of annotated images with a large set of unannotated images for training, and can better mine information from unannotated cases based on pseudo labels or consistency regularization (34). WSL leverages image-level or sparse annotations for training high-performance models (35), which reduces the annotation requirement. Additionally, noise-robust loss functions and training paradigms are implemented in SenseCare to learn from noisy labels (36), which prevents the model being corrupted by incorrect or noisy labels in the training set.

SenseCare provides a human-in-the-loop strategy for data annotation and model training. For a set of unannotated images, it first allows the user to label some images using the interactive annotation tools. With a set of labeled images (and optionally a set of unlabeled images for SSL), it then invokes the model training process through a configuration file, where configurations include the paths of training data and output model, training strategies (e.g., fully supervised learning or SSL), network structure, loss function, optimizer and hyper-parameters such as batch size, learning rate and training epoch number. Additionally, users can select image pre-processing and augmentation methods in the configuration file. After training, the model can be used to predict labels for remaining unlabeled images with uncertainty estimation, which is used to query the user for active learning-based annotation. The training process is iteratively conducted with more annotations obtained. The human-in-the-loop strategy avoids labeling all the images from scratch and improves the annotation efficiency. In each round of training, SenceCare allows the user to use a set of validation dataset to validate the performance and assess the degree of over-fitting, and the training can be early stopped if the performance on validation set does not increase for several epochs. Note that a single round of training is also allowed if the user has provided annotations for the training images in advance.

Docker integration. Users of SenseCare can employ the built-in deep learning models mentioned above for several image computing tasks, and they can also develop their own models with the help of SenseCare’s training and testing pipelines. When the user creates a new algorithm, SenseCare provides a dockerized version of the algorithm, and can package it with all the dependencies together in the form of containers, so that the algorithm can work seamlessly in any new environment. Therefore, researchers can focus on the development of AI algorithms without worrying about the testing and production environment. Since dockers are lightweight, SenseCare makes it more convenient and efficient for users to develop, test and deploy algorithms and deep learning models for various clinical applications.

2.3.2 Object detection

Automatic detection of objects of interest is a common task for computer assisted diagnosis systems. SenseCare has several built-in deep learning models for object detection tasks, such as landmark and lesion detection in 3D radiological images and cancer cell detection in histopathological images.

For example, automatic localization of vertebrae in CT is important for image-guided diagnosis, pre-operative planning and post-operative evaluation. Deep learning models such as 3D Fully Convolutional Neural Networks (FCN) are embedded in SenseCare for accurate vertebrae localization (26). The model also takes advantages of prior knowledge such as spatial and sequential constraints to obtain high robustness in challenging cases. In addition, SenseCare is able to predict tumor invasiveness and malignant in Ground Glass Opacity (GGO) on the basis of its lung nodule detection model (22). Figure 3a shows an example of lung nodule detection by deep learning models in SenseCare. For signet ring cell carcinoma detection from Hematoxylin and Eosin (H&E) stained Whole Slide Images (WSI), SenceCare is integrated with bottom-up approaches (37) that obtain cell instance masks first and then derive bounding boxes for each instance, which is more accurate than the general RCNN-based detection methods.

Figure 3
www.frontiersin.org

Figure 3. Nodule detection, image registration and 3D visualization of lung images in SenseCare. (a) Lung nodule detection. (b) 3D visualization. (c) Image registration.

2.3.3 Image segmentation

Image segmentation is essential for most clinical applications such as accurate modeling of anatomical structures, quantitative measurement of tumor volumes, planning of radiotherapy and surgical treatment. Its output has a large impact on the downstream workflows. However, due to the low contrast between the target tissue and its surroundings, inhomogeneous appearance, complex shape variation and image noise, accurate segmentation is extremely challenging and traditional image segmentation algorithms are often faced with large regions of over- and under-segmentation.

Supported by its deep learning-based image segmentation models, SenseCare can overcome these challenges and has obtained state-of-the-art performance in a range of segmentation tasks. For example, to segment the complex structures of pulmonary vessels from CT images, SenseCare is equipped with a multi-view-based 2.5D network with a low complexity, which outperformed other contemporary networks by a large margin on the LIDC dataset (38). To segment intervertebral discs from MR images, a novel multi-resolution path network with deep supervision is included in SenseCare and it achieved superior performance on the MICCAI 2018 IVDM3Seg challenge dataset (27). Several specific CNN models are also developed for other applications, such as multiple Organs-at-Risk (OAR) segmentation from CT for radiotherapy planning (28), segmentation of optic disc and cup for glaucoma diagnosis (39), nuclei instance segmentation in histopathological images (20, 21) and cartilage segmentation from MR images for osteoarthritis assessment (19). These different models are ready-to-use and serve as strong baselines for the above specific applications, and they can be easily adapted to new segmentation tasks based on the user’s research interests.

2.3.4 Image registration

Images acquired in different modalities are often need to be fused to provide sufficient information for diagnosis and treatment decision. In addition, a patient may be scanned several times at different stages of a disease to obtain a better understanding of the evolution of the disease. Therefore, it is necessary to register two or more images into a common spatial coordinate system for better interpretation of anatomical and pathological characteristics thus improving diagnosis and treatment for patients.

SenseCare provides several registration algorithms for different tasks. For example, the combination of MR and CT is quite useful since the former is better suited for delineation of tumor regions while the latter is needed for accurate computation of the radiation dose. Registration between MR and CT images supported by SenseCare makes it more efficient and accurate to obtain radiotherapy planning. Registration algorithms embedded in SenseCare include both rigid and non-rigid registration between images from a single modality or multiple modalities. Rigid methods are useful for the registration in the presence of rigid bodies such as bones. Non-rigid registration is used for applications such as correcting soft-tissue deformation during imaging and modeling the dynamic heart. Figure 3c shows an example of non-rigid lung image registration in SenseCare.

2.3.5 Foundation models

Recently, foundation models such as vision-language models (40) and Segment Anything Model (SAM) (41) have shown great potential in reducing data and annotation cost and improving performance in applications with limited dataset size (42). SenseCare currently supports using foundation models mainly in two ways for researchers. First, it allows researchers to use foundation models to efficiently annotate data in image classification and segmentation tasks. For example, by leveraging the zero-shot inference ability of vision-language models such as variants of CLIP (40), users can first obtain pseudo-labels of training images. Then, SenseCare uses uncertainty estimation to select reliable pseudo-labels, and leverages semi-supervised learning methods to learn from those with reliable pseudo-labels and other unlabeled samples. SenseCare also supports source-free domain adaptation of medical image segmentation models with the help of SAM (43). Specifically, pseudo-labels obtained by the source model is used to generate prompts for SAM that outputs improved pseudo-labels for target-domain images. The new pseudo-labels are then used to fine-tune the source model with prior knowledge-based constraints (44), which makes the source model be well adapted to a new domain without annotations.

Second, SenseCare provides a foundation model that improves the performance of downstream tasks and reduces the annotation cost (45). A hybrid architecture combining the advantage of CNN and Transformers is designed, and it is trained on a large-scale CT dataset based on a volume fusion-based self-supervised learning framework. Then the pretrained model can be adapted to different downstream segmentation targets including head and neck organs, thoracic/abdominal organs and lesions in CT and other modalities.

3 Clinical applications

The most important high-level goal of SenseCare is to serve as a platform for clinical research in various applications. In this section, we give several examples of user scenarios where SenseCare is adapted for various clinical applications.

3.1 Lung-oriented application

3.1.1 Chest radiograph multi-disease detection

Chest Radiography (CR) serves as the most common imaging tool for lung disease screening and diagnosis. SenseCare offers a comprehensive CR-based multi-disease detection system, encompassing foreign body detection, screening for negative findings, multi-disease detection, rib counting, cardiothoracic ratio measurement, scoliosis calculation, and automated report generation, as shown in Figures 4a,b. Specifically, SenseCare begins by performing quality control on the input chest radiograph. It then detects 14 diseases, including Abnormal Aortic Knob, Atelectasis, Cardiomegaly, Edema, Emphysema, Fibrosis, Hilar Enlargement, Nodule, Pleural Effusion, Pleural Thickening, Pneumonia, Pneumothorax, Rib Fracture, and Tuberculosis, providing bounding box and confidence score for each disease. For pneumothorax, it further calculates the pneumothorax compression ratio by segmenting the pneumothorax and lung regions using UNet (46). Rib instance segmentation and detangling are implemented to achieve rib counting and determine the corresponding regions for disease boxes. The cardiothoracic ratio is calculated by detecting key points that define the chest contour and heart width. Cobb angle measurement for scoliosis is achieved through spine segmentation and extraction of the spinal centerline (47). Through the coordination of multiple modules, SenseCare offers an automated, rapid, and comprehensive fine-grained diagnosis for CR.

Figure 4
www.frontiersin.org

Figure 4. Chest X-Ray image and Lung CT analysis in SenseCare. (a) User interface of Chest X-ray analysis. (b) Multi-disease detection and automated report generation. (c) Intelligent lung nodule analysis and follow-up in CT. (d) Intelligent analysis of pulmonary embolism. (e) Lung surgical planning with visualized traced tube structures. (f) Lung surgical planning with lung segment analysis.

3.1.2 Lung nodule analysis and follow-up

Lung cancer is a malignant disease with a five-year survival rate of 16%–17%. Researches have shown that early diagnosis and intervention will improve the five-year survival rate up to 54%. Although computer-assisted image analysis have been adopted to facilitate the lung cancer detection and diagnosis in an early stage, the various sizes, shapes and types of pulmonary nodules are placing great burdens on clinical physicians, thus causing misdiagnosis and missed diagnosis due to their fatigue and overwork.

Based on its leading algorithms, SenseCare supports a thorough research and analysis of pulmonary nodules and lesions (36) by automatic detection, segmentation, and quantitative analysis. It can automatically detect and locate the nodules (22, 48) and then provide further quantitative information of each nodule such as its volume and density, in addition to qualitative estimation of its type and malignancy. As shown in Figure 3, four kinds of nodules are automatically detected and distinguished by SenseCare (23). Furthermore, it equips clinicians with an advanced nodule follow-up feature. Utilizing registration technology, it facilitates the longitudinal analysis between any two patient scans, automatically pinpointing and correlating the positions of existing, vanished, or emergent nodules. It meticulously gauges alterations in nodule size, density, and malignancy risk, estimating the lesion’s doubling time, which is a crucial metric that informs clinical decision-making. Figure 4c illustrates the user interface for thoracic CT imaging analysis and smart follow-up within SenseCare. Compared with COVI3D (14, 15) that is specifically designed for lung lesion analysis, SenseCare provides a more detailed analysis and visualization of other important thoracic structures such as different lung lobes, vessels and airways.

3.1.3 Pulmonary embolism analysis

Pulmonary Embolism (PE) is formed when a portion of a blood clot breaks off from the wall of a vein and travels through the blood stream, passes through the heart (right atrium and right ventricle), and becomes lodged in a pulmonary artery, causing a partial or complete obstruction. The best available diagnostic technique of PE is Computed Tomography Pulmonary Angiography (CTPA) (49), where PEs usually appear as dark spots among the bright regions of blood arteries.

Leveraging CTPA data, SenseCare employs advanced algorithms for detection, segmentation, and localization, enabling the automated identification, positioning, and quantitative analysis of pulmonary embolisms (50). The system proactively issues high-risk alerts when blockages are present in the principal pulmonary artery. Furthermore, it performs curved plane reconstruction for each individual embolus, streamlining the examination of the embolism’s overall state. The curved plane reconstruction technology vividly illustrates the embolus’s position and morphology within the vessel, enhancing physicians’ capacity to evaluate the condition. In addition, SenseCare provides intelligent 3D reconstruction for clinicians. By offering a 3D visualization of various anatomical structures such as the lungs, pulmonary arteries and veins, pericardium, and embolism, it eases the assessment of the proximity of emboli to neighboring tissues, furnishing a crucial benchmark for subsequent thrombolytic treatments. Figure 4b illustrates the user interface for CTPA imaging PE analysis within SenseCare.

Table 1 shows the performance of SenseCare and related works on lung CT analysis. It obtains a sensitivity of 0.93 at 6 False Positives (FPs) per scan for lung nodule detection, and a Dice of 0.945 for lung lobe segmentation. It also supports pulmonary vessel and airway segmentation with an average Dice of 0.914, and the sensitivity for pulmonary embolism detection is 0.90 at 2 FPs per scan.

Table 1
www.frontiersin.org

Table 1. Performance of SenseCare and related works on lung CT analysis.

3.1.4 Lung surgical planning

Surgical operations play an important role in effective treatment of lung cancer. Due to complex anatomical structures as well as breathing movements in chest, visualized surgical planning is essential for improving the efficiency and success rate of surgeries. SenseCare has provided the Lung Surgical Planning system that can automatically build 3D quantitative pulmonary structures from CT images, shown in Figures 4c,d. The whole pipeline of extracting various structures is completed in three minutes without manual interactions. Soft tissues including pulmonary parenchyma, pulmonary lobes and segments are extracted as analysis basis. Besides, the bronchi tree is traced and analyzed according to the position and depth, acting as important landmarks during resection. Furthermore, pulmonary arteries and veins are obtained following the vessel tree structures, which ensures the precise segmentation results even in non-contrast CT images. Base on above results, the system supports watershed analysis for precise lung resection, as well as precise needle placement suggestions for tumors and surrounding tissues in ablation treatment. The system helps clinicians obtain both global visualizations and important details, leading to more efficient and accurate operations in clinical practice.

3.2 Pathology-related application

Pathological diagnosis is regarded as the most reliable criteria for diagnosis of cancers such as Gastrointestinal cancer. However, it is labor-insensitive and time-consuming to manually discern lesion areas and pathological cells from up to 100,000 × 100,000-pixel whole slide pathological images, which easily leads to fatigue of human analysts and cannot satisfy the increasing demand for pathological diagnosis due to the lack of adept experts in developing countries.

SenseCare-Pathology is designed to support pathological diagnosis across a spectrum of anatomical regions, including gastrointestinal, cervical, lymph nodes, lung, prostate and pleural effusion & ascites diseases, and it facilitates both tissue and cellular level analysis (37, 60). This innovative platform empowers pathology department of hospitals and third-party pathological diagnosis centers to perform efficient and intelligent analysis of pathological images by providing AI-based functional modules such as lesion and cell detection (37), nuclei/gland segmentation (21) and cell segmentation (61), etc. Especially, SenseCare supports efficient retrieval of pathological images from a large-scale database (6163). Users can employ the retrieval module to search for similar images for a given input image. These modules help clinical researchers build capability to conduct large-scale cancer screening projects. Figure 5 shows an example of lesion localization from pathological images.

Figure 5
www.frontiersin.org

Figure 5. Pathological image analysis in SenseCare.

Table 2 shows the performance of SenseCare compared with some recent works on pathological image classification tasks. For gastric cancer malignancy detection from WSIs, it obtains an sensitivity of 0.95 and specificity of 0.91. For colorectal cancer malignancy detection, the sensitivity and specificity is 0.93 and 0.85, respectively. For Thin-prep Cytologic Test (TCT) analysis, the sensitivity and specificity of cervical cancer malignancy detection is 0.95 and 0.83, respectively. SenseCare also supports classification of mold from TCT (sensitivity 0.90 and specificity 0.99), classification of clue cells (sensitivity 0.92 and specificity 0.99), segmentation of cancer region in WSI (Dice 0.87) and detection of low-quality WSIs including the presence of bubble, glue, folding, contaminant with an overall sensitivity of 0.90.

Table 2
www.frontiersin.org

Table 2. Performance of SenseCare and related works on pathological image analysis.

3.3 Heart disease diagnosis

Cardiovascular diseases (CVDs) are the most common causes of death throughout the world. Non-invasive morphological and functional assessment of cardiovascular structures plays an important role in diagnosis and treatment of CVDs. Computed Tomography Angiography (CTA) and electrocardiogram (ECG) are two common means for diagnosis of CVDs. However, manual analysis of CTA and ECG is laborious and time-consuming. SenseCare provides a full-stack solution towards automatic diagnosis of coronary artery disease based CTA and ECG.

3.3.1 Coronary artery disease diagnosis using CTA

For automatic diagnosis of coronary artery disease using CTA, SenseCare provides a full-stack solution, including fully automatic segmentation of 3D whole heart and coronary arteries, extraction of coronary artery centerlines, labeling of important artery branches, reconstruction of MPR and CPR, real-time volumetric rendering, detection of plaques and quantification of stenosis, and automatic generation of diagnose reports.

More specifically, SenseCare employs a cascaded method for automatic segmentation of 3D whole heart (70) and coronary arteries, which is integrated into a semi-supervised framework (71) and has achieved state-of-the-art performance using very few manual annotations. For robust and accurate coronary artery segmentation, an Artery and Vein Disentanglement Network (AVDNet) is proposed by incorporating the coronary vein into the segmentation task (72). A unified deep reinforcement learning (73) framework is proposed to automatically traverse tree-structure centerlines of coronary arteries. Pixel-level segmentation followed by 3D classification and segmentation of point sets (74) is proposed for coronary artery labeling. We utilize a recurrent CNN to automatic detect and classify the type of coronary artery plaque (75), and the degree of coronary artery stenosis is quantified by a multi-class segmentation of plaque and vessel lumen from the reconstructed probe images. Finally, a structured diagnose report is summarized based on the results obtained from all steps mentioned above, and the whole procedure is finished within one minute. Figure 6a shows a snapshot of the coronary artery application in SenseCare.

Figure 6
www.frontiersin.org

Figure 6. Heart disease and liver cancer diagnosis and surgical planning system in SenseCare. (a) Conroy artery disease diagnosis using CTA. (b) Cardiac disease diagnosis using ECG. (c) Liver cancer diagnosis system. (d) Liver surgical planning system.

Table 3 shows the performance of coronary artery and cerebral artery CTA analysis obtained by SenseCare compared with recent related works. For coronary artery analysis, SenseCare obtains a patient-level accuracy of 0.94 for stenosis detection ( 50% stenosis), with an accuracy of 0.97 for calcium score calculation and 0.96 for coronary artery labeling, respectively. For cerebral artery analysis, the accuracy of artery labeling is 0.95, and the sensitivity of intracranial aneurysm detection is 0.94. The sensitivity of cerebral hemorrhage detection is 0.99, with a specificity of 0.92. The sensitivity and specificity for cerebral ischemia detection is 0.88 and 0.95, respectively.

Table 3
www.frontiersin.org

Table 3. Performance of SenseCare and related works on CTA analysis.

3.3.2 Automatic diagnosis of cardiac diseases using ECG

As a common means of heart disease detection, ECG has the characteristics of being inexpensive, non-invasive, and easily available. The manual measurement of parameters and waveform recognition are essential steps in the traditional ECG diagnostic process, posing significant challenges for doctors when analyzing long signal sequences. As shown in Figure 6b, SenseCare provides an automatic solution for ECG diagnosis, including parameter measurement and diseases classification.

3.4 Liver cancer diagnosis and surgical planning

Primary liver cancer, mainly Hepatocellular Carcinoma (HCC), is the sixth most prevalent malignancy and third leading cause of cancer-related death worldwide. The prognosis is poor owing to the high recurrence in 60%–70% of patients within 5 years after curative surgery (84). Diagnosis of liver cancer based on multi-modal medical images is important for treatment decision.

SenseCare provides a closed-loop platform for intelligent diagnosis of liver cancer using multi-phase dynamic CT scans and multi-sequence MRI images. As shown in Figure 6c, it has an integrated pipeline for automatic sequence alignment, detection and localization of tumors (85). Moreover, it provides morphology and density measurements for assessment of cancer type, malignancy and LI-RADS symptom (86). It also implements automatic segmentation and volume measurements of abdomen organs, and risk assessment of cirrhosis and fatty liver. Finally, a structured and standardized diagnostic report is generated. Therefore, SenseCare serves as a promising tool to help radiologists locate and identify the suspected liver tumors more efficiently and effectively.

Table 4 shows the performance of SenseCare and recent relevant works on structure segmentation and lesion detection in liver CT images. For abdominal organ segmentation, the average Dice of liver, left kidney, right kidney and spleen from CT and MRI is 0.96 and 0.95, respectively. For hepatic vessel segmentation, the recall of center line is 0.88 in both CT and MRI. The sensitivity and specificity of lesion detection are 0.95 and 0.70 for CT images, and 0.92 and 0.70 for MRI images, respectively. For classification between benign and malignant liver lesions, the accuracy is 0.90, with an AUC of 0.95. SenseCare also supports automatic recognition dynamic phases of contrast enhanced CT with an accuracy of 0.97, and the classification accuracy for different MRI dynamic sequences is 0.97.

Table 4
www.frontiersin.org

Table 4. Performance of SenseCare and related works on liver image analysis.

For liver surgical planning, SenseCare provides a quantitative 3D modeling of the liver for highly automated liver-specific analysis, which is shown in Figure 6d. It helps hepatobiliary surgeons perform interactive analysis and treatment planning. In detail, it supports efficient automatic segmentation and quantitation of tumors, vessels, ducts and abdomen organs within four minutes, and interactive refinement of segmentation results. It supports pre-surgical planning for liver resection in multiple paradigms including anatomical, vessel-territory-based and curvature-based methods. The system also aides in precise needle placement relative to tumors and surrounding tissues for ablation treatment. It helps clinicians get a whole picture of the abdomen in a quicker and more comprehensive paradigm, which improves the safety of operations and ensures the treatment outcome.

3.5 Pelvic tumor surgical planning

Pelvic tumor is one of highly malignant tumors, and surgical treatment is the most effective treatment for pelvis tumors, where an accurate preoperative simulation and planning based on segmentation and modeling of the tumor is critical for the success of surgery.

SenseCare provides an intelligent preoperative surgical planning for for limb salvage surgery of malignant pelvic tumors, where the accuracy and efficiency is improved by our deep learning-based algorithms. The surgical planning workflow mainly consists of three parts. First, the pelvic tumor is segmented from MR scans with a U-Net like model (46). Then the pelvic bone is segmented from CT scans with a CNN combined with self-attention blocks (92). Finally, a robust rigid/affine inverse-consistency registration method that is an extension of SymMirorr (93) is conducted to align the corresponding MR-CT pair. Based on the registered CT-MR pair and the corresponding segmentation results, surgeons and radiologists could make accurate preoperative surgical planning rapidly. Figure 7a shows the user interface of pelvic tumor surgical planning system in SenseCare. The system efficiently cuts down the multi-party communication costs between radiology, orthopedics and 3D printing centers, and ultimately reduces doctor’s workload and facilitates patient-tailored treatment. It took only 15 min to complete the surgical planning for pelvic tumor resection, which is a dramatic acceleration compared with the 2-day time span in a traditional workflow (94).

Figure 7
www.frontiersin.org

Figure 7. Pelvic tumor and skeleton-oriented applications in SenseCare. (a) Pelvic tumor surgical planning. (b) Lower limb surgical planning. (c) Bone age assessment. (d) Skeleton reconstruction.

3.6 Skeleton-oriented application

Skeleton-relevant diseases such as fracture, development disorder and osteoarthritis could largely reduce the patients’ quality of life, and lead to several complications. Skeleton analysis based on medical imaging is essential for accurate diagnosis and treatment planning of such diseases. SenceCare provides several intelligent tools for diagnosis and assessment of skeleton diseases, including lower limb surgical planning, bone age assessment and whole body skeleton reconstruction.

3.6.1 Lower limb surgical planning

Knee Osteoarthritis (OA) is an important public health issue that causes chronic disability. Lower limb osteotomies is a well-established and commonly utilized technique in medial knee osteoarthritis secondary to different forms of knee joint malalignments, trying to establish a better alignment by passing the load bearing leg axis through hip, knee and ankle joint. The measurement of lower limb axial alignment with radiographs is a critical step in the preoperative planning to exactly define the characteristic of the osteotomy.

SenseCare’s lower limb surgical planning system provides advanced AI algorithms for automatic radiograph analysis such as segmentation of femur and tibia and detection of lower limb landmarks for accurate measurement of the critical angles and distances of lower limbs. The results are used for planning of surgical treatments such as High Tibial Osteotomy (HTO), as shown in Figure 7b. Localization of landmarks is based on a coarse-to-fine recurrent network (95) combined with Gaussian heatmap regression (96). The heatmap corresponding to a landmark location is a sum of multiple Gaussian functions centered at that landmark. At inference time, the largest values in the heatmap are taken as the detected landmark positions. Such a strategy helps to achieve stable and accurate limb landmark detection results that ensure the reliability of downstream surgical planning.

3.6.2 Bone age assessment

Bone age is an important and widely used quantitative metric to estimate the development of child’s skeleton in the field of pediatrics. Greulich and Pyle (GP) and Tanner–Whitehouse (TW) are the two most used bone age assessment methods in clinical practice. The GP method requires comparing the radiograph with the reference atlases and taking the bone age of the closest one as the evaluation result. Yet TW method is much more complex due to its scoring mechanism. The radiologist needs to manually evaluate the developmental stages of 20 specific bones. Each developmental stage of each bone corresponds to a score. By adding the scores of all bones, the final bone maturity score is obtained, and the bone age is calculated. Obviously, this is a time-consuming and error-prone task, and a fully automated workflow is desirable.

SenseCare provides a fast and fully automated solution that integrates multiple bone age assessment methods, including TW3 RUS, TW3 Carpal, and GP. As shown in Figure 7c, it can automatically detect 20 specific bones in the hand X-ray image, evaluate the development status of each bone, and calculate the bone age based on different methods. Meanwhile, an assessment report containing descriptions of primary developmental characteristics is also presented.

3.6.3 Whole body skeleton reconstruction

The SenseCare intelligent Virtual Reality (VR) application is oriented to the whole body’s skeleton and joints, and fully automatically performs VR and standard plane reconstruction to help users interpret scans. For the VR function, the target skeleton are fully automatically distinguished from the background including non-interesting bones and plaster fixation, thereby eliminating irrelevant information during VR display. For standard plane reconstruction, the shape and posture of the target joint (including shoulder, elbow, wrist, knee, ankle, foot, etc.) in the image can be automatically recognized, and then the scan is reconstructed along a preset standard direction, as shown in Figure 7d.

3.7 Stroke diagnosis

Stroke is a severe cerebrovascular disease globally, with high incidence, high disability rate and mortality (97). Ischemic stroke is the most common type of stroke, accounting for 75%–85% of all stroke cases. The detection and quantitative evaluation of stroke lesions by medical imaging is of great significance for accurate diagnosis and treatment decision.

The SenseCare platform provides a one-stop stroke diagnosis solution based on different types of CT imaging. First, it provides automatic analysis of intracerebral hemorrhage and ischemic stroke using non-contrast CT. As shown in Figure 8a, for intracerebral hemorrhage, it can detect the hemorrhage and calculate the volume of bleeding automatically, and classify the hemorrhage into five types. For ischemic stroke, the ASPECT score is automatically measured for preliminary assessment of the severity of ischemia, as shown in Figure 8b. Second, for ischemic stroke, the CT perfusion analysis system can decode raw 4D CT perfusion images into perfusion parameter maps (CBV, CB, MTT and TMAX) for accurate diagnosis, as shown in Figure 8c. It can quantify the volume (98) and mismatch between the infarct core and the ischemic penumbra. Thirdly, for thrombectomy patients, SenseCare features a head and neck CTA reconstruction and analysis system (99) for rapid vascular reconstruction and localization of plaque occluded areas and aneurysm, as shown in Figure 8d. In terms of quantitative performance, the sensitivity and specificity for infarcted region detection is 0.92 and 0.94, respectively, and the relative volume error for low-perfusion region segmentation is 5.0%. The holistic solution and high performance in CT perfusion image analysis can be applied to stroke centers to improve their treatment capacity.

Figure 8
www.frontiersin.org

Figure 8. Stroke diagnosis in SenseCare. (a) Hemorrhage diagnosis. (b) ASPECT score measurement for ischemic stroke. (c) CT perfusion analysis. (d) Head and neck CTA analysis.

3.8 Radiotherapy contouring

Radiotherapy is an important and widely-adopted treatment for cancers. In radiotherapy treatment planning, one of the most critical step is the delineation of the organs-at-risk (OAR) and target volumes, the quality of which is a core factor affecting the efficacy and side effects of radiotherapy. Clinically, radiologists have to spend several hours for manual delineation, due to the large number of OARs, complex shapes of cancers and large size of 3D volumes. As a result, it requires great time and efforts, as well as a high level of professionalism for the radiologist.

As shown in Figure 9, SenseCare Radiotherapy Contouring system aids in the radiologists by automatically contouring OARs across the whole body, including the head and neck (100103), chest (104) and abdomen (105, 106), as well as common targets including breast cancer, rectal cancer (107), etc. The system can be seamlessly connected to different kinds of CT machines and Treatment Planning Systems (TPS), automatically run AI calculation in background, support a 3D interactive view and editing the delineated structures, and export the results back via the standard RT structure DICOM protocol, in a complete closed loop of workflow.

Figure 9
www.frontiersin.org

Figure 9. Radiotherapy contouring system in SenseCare.

4 Discussion

With the development of medical imaging techniques and artificial intelligence, AI-based medical image computing systems have an increasing potential for more intelligent diagnosis and treatment in clinic practice. However, existing deep learning platforms rarely support specific clinical applications while most current medical image analysis platforms are not well supported by advanced AI algorithms. A desirable research platform for AI-based medical image informatics should have easy-to-use and extensive AI modules, be accessible to clinical researchers at multiple centers and on various devices, and support a range of clinical applications with different data modalities. To achieve this goal and boost clinical research towards smart healthcare, SenseCare is developed as a generic research platform for intelligent medical image computing that can support various research needs across different medical disciplines.

To analyze how SenseCare meets the above goals, we collected some feedback from users. 53 clinical users from 9 hospitals at 4 different cities were asked to assess to which extent they agreed or disagreed (a 7-point scale) with 8 statements: (Q1) SenseCare helps a lot in image analysis for diagnosis; (Q2) SenseCare is easy to learn and operate; (Q3) SenseCare helps to improve the diagnosis performance of users; (Q4) It effectively reduces the risk of missed diagnosis for lesions especially small lung nodules and rib fracture; (Q5) SenseCare has a good performance for 3D vessel reconstruction for coronary artery CTA, head and neck CTA and brain Magnetic Resonance Angiography (MRA); (Q6) SenseCare’s 3D reconstruction of anatomical structures based on segmentation is helpful for surgical planning in lung and liver cancers; (Q7) SenseCare improves the efficiency and performance in follow-up of patients; (Q8) Expect to use more AI tools in SenseCare to assist solving problems in healthcare. The summary of agreement level is shown in Table 5. 51 out of 53 users (96.23%) agreed or strongly agreed that SenseCare is helpful for diagnosis (Q1), and 52 out of 53 users agreed or strongly agreed that SenseCare is easy to learn and use (Q2). More than 46 (86.79%) users agreed or strongly agreed that SenseCare helps to improve the capability of users (Q3), reduces missed diagnosis (Q4), obtains good results in vessel reconstruction (Q5) and leads to better follow-up (Q7). In terms of being helpful for surgical planing, 51 users gave the first three agreement levels, accounting for 96.23%. In average, 90.81% users agreed or strongly agreed with the 8 statements.

Table 5
www.frontiersin.org

Table 5. Agreement level of 53 clinical users on 8 statements of SenseCare.

As shown in Table 6, compared with existing platforms of medical image analysis and visualization for clinical applications, SenseCare has several advantages: (1) On-the-shelf AI toolkits for different medical image computing tasks such as image segmentation, registration, lesion and landmark detection, which makes it easy for users to not only directly use these built-in AI models for specific tasks, but also develop new customized AI models based on SenseCare’s tools for intelligent image annotation, model training and validation; (2) Extensive clinical applications that are developed for a range of clinical diagnosis and treatment planning tasks, and the comprehensive AI functional modules and advanced 3D visualization make SenseCare scalable for new clinical applications; (3) Easy to access with high concurrency and low requirement on users’ device or operating system, due to its cloud-based service and multi-center deployment; and (4) Efficient synchronization of data from different information systems and cross multiple centers, with data privacy ensured by encryption algorithms.

Table 6
www.frontiersin.org

Table 6. Comparison between SeneseCare and existing medical image computing and visualization platforms for clinical applications.

Despite that a range of AI toolkit have been available in SenseCare for many clinical applications, there are still some aspects that need to be improved. First, the current foundation models in SenseCare are mainly for general radiology and pathology image analysis, and foundation model for other modalities such as surgical videos will be considered in the future (108). Second, SenseCare currently focuses mainly on structured medical data. However, many healthcare scenarios require the integration of unstructured data such as clinical notes and patient histories. We will expand its support for unstructured data to enhance the platform’s ability to provide comprehensive diagnostic insights in the future. Thirdly, to better support precision medicine, prediction of prognosis, treatment responses and survival rates of patients will be supported by SenseCare, which will contribute to more effective treatment decision-making for patients with serious diseases such as malignant cancers.

5 Conclusion

In this paper, we present a one-stop research platform SenseCare that provides a large set of AI algorithms for medical image segmentation, registration, detection and other tasks that can help clinicians and radiologists conduct various clinic-oriented translational research programs, such as lung cancer diagnosis and surgical planning, efficient pathological image analysis, pelvic tumor and limb surgical planning, coronary artery disease diagnosis and modeling, etc. In addition to the built-in AI algorithms, SenseCare also provides several tools for users to develop and deploy customized AI models efficiently. The AI toolkits and other appearing functional modules such as advanced visualization, web-based access and multi-center deployment in SenseCare can efficiently boost clinical research programs and applications towards smart healthcare.

Data availability statement

The raw data supporting the conclusions of this article will be made available by the authors, without undue reservation.

Author contributions

GW: Formal Analysis, Writing – original draft, Funding acquisition. QD: Writing – original draft, Software, Project administration. TS: Resources, Writing – review & editing, Visualization, Methodology, Conceptualization. SZ: Writing – review & editing, Supervision, Resources, Investigation.

Funding

The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported in part by the National Key Research and Development Program of China (2022ZD0160705), and National Natural Science Foundation of China (62271115).

Conflict of interest

The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.

Publisher's note

All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.

References

1. Abadi M, Barham P, Chen J, Chen Z, Davis A, Dean J, et al.. Tensorflow: a system for large-scale machine learning. In: USENIX Symposium on Operating Systems Design and Implementation. (2016). p. 265–84.

Google Scholar

2. Paszke A, Gross S, Massa F, Lerer A, Bradbury J, Chanan G, et al.. PyTorch: an imperative style, high-performance deep learning library. In: Advances in Neural Information Processing Systems. (2019). p. 8024–35.

Google Scholar

3. Chicho BT, Sallow AB. A comprehensive survey of deep learning models based on Keras framework. J Soft Comput Data Mining. (2021) 2:49–62. doi: 10.30880/jscdm.2021.02.02.005

Crossref Full Text | Google Scholar

4. Gibson E, Li W, Sudre C, Fidon L, Shakir DI, Wang G, et al.. NiftyNet: a deep-learning platform for medical imaging. Comput Methods Programs Biomed. (2018) 158:113–22. doi: 10.1016/j.cmpb.2018.01.025

PubMed Abstract | Crossref Full Text | Google Scholar

5. Pawlowski N, Ktena SI, Lee MCH, Kainz B, Rueckert D, Glocker B, et al.. DLTK: state of the art reference implementations for deep learning on medical images. CoRR abs/1711.06853. (2017). p. 1–4.

Google Scholar

6. Wang G, Luo X, Gu R, Yang S, Qu Y, Zhai S, et al.. PyMIC: a deep learning toolkit for annotation-efficient medical image segmentation. Comput Methods Programs Biomed. (2023) 231:107398. doi: 10.1016/j.cmpb.2023.107398

PubMed Abstract | Crossref Full Text | Google Scholar

7. Cardoso MJ, Clarkson M, Modat M, Ourselin S. NiftySeg: open-source software for medical image segmentation, label fusion and cortical thickness estimation. In: ISBI Workshop on Open Source Medical Image Analysis Software (2012).

Google Scholar

8. Klein S, Staring M, Murphy K, Viergever MA, Pluim JP. Elastix: a toolbox for intensity-based medical image registration. IEEE Trans Med Imaging. (2010) 29:196–205. doi: 10.1109/TMI.2009.2035616

PubMed Abstract | Crossref Full Text | Google Scholar

9. Modat M, Ridgway GR, Taylor ZA, Lehmann M, Barnes J, Hawkes DJ, et al.. Fast free-form deformation using graphics processing units. Comput Methods Programs Biomed. (2010) 98:278–84. doi: 10.1016/j.cmpb.2009.09.002

PubMed Abstract | Crossref Full Text | Google Scholar

10. Schroeder WJ, Martin KM. The visualization toolkit. In: Visualization Handbook. (2005). p. 593–614.

Google Scholar

11. Nolden M, Zelzer S, Seitel A, Wald D, Müller M, Franz AM, et al.. The medical imaging interaction toolkit: challenges and advances. Int J Comput Assist Radiol Surg. (2013) 8:607–20. doi: 10.1007/s11548-013-0840-8

PubMed Abstract | Crossref Full Text | Google Scholar

12. Clarkson MJ, Zombori G, Thompson S, Totz J, Song Y, Espak M, et al.. The NifTK software platform for image-guided interventions: platform overview and NiftyLink messaging. Int J Comput Assist Radiol Surg. (2015) 10:301–16. doi: 10.1007/s11548-014-1124-7

PubMed Abstract | Crossref Full Text | Google Scholar

13. Pieper S, Halle M, Kikinis R. 3D slicer. In: IEEE International Symposium on Biomedical Imaging: Macro to Nano. (2004). p. 632–5.

Google Scholar

14. Benbelkacem S, Oulefki A, Agaian S, Zenati-Henda N, Trongtirakul T, Aouam D, et al.. COVI3D: automatic COVID-19 CT image-based classification and visualization platform utilizing virtual and augmented reality technologies. Diagnostics. (2022) 12. doi: 10.3390/diagnostics12030649

PubMed Abstract | Crossref Full Text | Google Scholar

15. Benbelkacem S, Oulefki A, Agaian SS, Trongtirakul T, Aouam D, Zenati-Henda N, et al.. Lung infection region quantification, recognition, and virtual reality rendering of CT scan of COVID-19. Proc SPIE. (2021) 16:11734. doi: 10.1117/12.2587757

Crossref Full Text | Google Scholar

16. Benbelkacem S, Zenati-Henda N, Zerrouki N, Oulefki A, Agaian S, Masmoudi M, et al.. Tumor lung visualization and localization through virtual reality and thermal feedback interface. Diagnostics. (2023) 13:1–13. doi: 10.3390/diagnostics13030567

Crossref Full Text | Google Scholar

17. Oulefki A, Agaian S, Trongtirakul T, Benbelkacem S, Aouam D, Zenati-Henda N, et al.. Virtual reality visualization for computerized COVID-19 lesion segmentation and interpretation. Biomed Signal Process Control. (2022) 73:103371. doi: 10.1016/j.bspc.2021.103371

PubMed Abstract | Crossref Full Text | Google Scholar

18. Li Z, Lou Y, Yan Z, Aref SA, Min JK, Axel L, et al.. Fully automatic segmentation of short-axis cardiac MRI using modified deep layer aggregation. In: International Symposium on Biomedical Imaging. (2019). p. 793–7.

Google Scholar

19. Tan C, Yan Z, Zhang S, Li K, Metaxas DN. Collaborative multi-agent learning for MR knee articular cartilage segmentation. In: Medical Image Computing and Computer-Assisted Intervention. (2019). p. 282–90.

Google Scholar

20. Li J, Hu Z, Yang S. Accurate nuclear segmentation with center vector encoding. In: Information Processing in Medical Imaging. (2019). p. 394–404.

Google Scholar

21. Qu H, Yan Z, Riedlinger GM, De S, Metaxas DN. Improving nuclei/gland instance segmentation in histopathology images by full resolution neural network and spatial constrained loss. In: Medical Image Computing and Computer-Assisted Intervention. (2019). p. 378–86.

Google Scholar

22. Yu Y, Wang N, Huang N, Liu X, Cheng J, Xu J. Tumor invasiveness and malignant prediction in GGO lung nodules using deep learning methods. In: RSNA. (2019).

Google Scholar

23. Zhang Y, Liu Y, Han H, Qu F, Huang N, Ye Z. Deep learning in discriminating atypical lung nodules: a diagnostic test. In: ECR (2019).

Google Scholar

24. Wang G, Li W, Ourselin S, Vercauteren T. Automatic brain tumor segmentation based on cascaded convolutional neural networks with uncertainty estimation. Front Comput Neurosci. (2019) 13:1–13. doi: 10.3389/fncom.2019.00056

PubMed Abstract | Crossref Full Text | Google Scholar

25. Wang G, Shapey J, Li W, Dorent R, Demitriadis A, Bisdas S, et al.. Automatic segmentation of vestibular schwannoma from T2-weighted MRI by deep spatial attention with hardness-weighted loss. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. (2019). Vol. 2. p. 264–72.

Google Scholar

26. Chen Y, Gao Y, Li K, Zhao L, Zhao J. Vertebrae identification and localization utilizing fully convolutional networks and a hidden markov model. IEEE Trans Med Imaging. (2020) 39:387–99. doi: 10.1109/TMI.2019.2927289

PubMed Abstract | Crossref Full Text | Google Scholar

27. Gao Y, Liu C, Zhao L. Multi-resolution path CNN with deep supervision for intervertebral disc localization and segmentation. In: Medical Image Computing and Computer-Assisted Intervention. (2019). p. 309–17.

Google Scholar

28. Gao Y, Huang R, Chen M, Wang Z, Deng J, Chen Y, et al.. FocusNet: imbalanced large and small organ segmentation with an end-to-end deep neural network for head and neck CT images. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. (2019). p. 829–38.

Google Scholar

29. Lei W, Wang H, Gu R, Zhang S, Zhang S, Wang G. DeepIGeoS-V2: deep interactive segmentation of multiple organs from head and neck images with lightweight CNNs. In: LABELS Workshop of MICCAI. (2019). p. 61–9.

Google Scholar

30. Shen D, Wu G, Suk H-I. Deep learning in medical image analysis. Annu Rev Biomed Eng. (2017) 19:221–48. doi: 10.1146/annurev-bioeng-071516-044442

PubMed Abstract | Crossref Full Text | Google Scholar

31. Luo X, Wang G, Song T, Zhang J, Aertsen M, Deprest J, et al.. MIDeepSeg: minimally interactive segmentation of unseen objects from medical images using deep learning. Med Image Anal. (2021) 72:102102. doi: 10.1016/j.media.2021.102102

PubMed Abstract | Crossref Full Text | Google Scholar

32. Wang G, Aertsen M, Deprest J, Ourselin S, Vercauteren T, Zhang S. Uncertainty-guided efficient interactive refinement of fetal brain segmentation from stacks of MRI slices. In: International Conference on Medical Image Computing and Computer Assisted Intervention. (2020). Vol. 4. p. 279–88.

Google Scholar

33. Sun P, Feng W, Han R, Yan S, Wen Y. Optimizing network performance for distributed DNN training on GPU clusters: ImageNet/AlexNet training in 1.5 min. CoRR abs/1902.06855 (2019).

Google Scholar

34. Luo X, Wang G, Liao W, Chen J, Song T, Chen Y, et al.. Semi-supervised medical image segmentation via uncertainty rectified pyramid consistency. Med Image Anal. (2022) 80:102517. doi: 10.1016/j.media.2022.102517

PubMed Abstract | Crossref Full Text | Google Scholar

35. Luo X, Hu M, Liao W, Zhai S, Song T, Wang G, et al.. Scribble-supervised medical image segmentation via dual-branch network and dynamically mixed pseudo labels supervision. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. Cham, Switzerland: Springer Nature (2021). p. 528–38.

Google Scholar

36. Wang G, Liu X, Li C, Xu Z, Ruan J, Zhu H, et al.. A noise-robust framework for automatic segmentation of COVID-19 pneumonia lesions from CT images. IEEE Trans Med Imaging. (2020) 39:2653–63. doi: 10.1109/TMI.2020.3000314

PubMed Abstract | Crossref Full Text | Google Scholar

37. Li J, Yang S, Huang X, Da Q, Yang X, Hu Z, et al.. Signet ring cell detection with a semi-supervised learning framework. In: Information Processing in Medical Imaging. (2019). p. 842–54.

Google Scholar

38. Cui H, Liu X, Huang N. Pulmonary vessel segmentation based on orthogonal fused U-Net++ of chest CT images. In: Medical Image Computing and Computer-Assisted Intervention. (2019). p. 293–300.

Google Scholar

39. Liu P, Kong B, Li Z, Zhang S, Fang R. CFEA: collaborative feature ensembling adaptation for domain adaptation in unsupervised optic disc and cup segmentation. In: Medical Image Computing and Computer-Assisted Intervention. (2019). p. 521–9.

Google Scholar

40. Radford A, Kim JW, Hallacy C, Ramesh A, Goh G, Agarwal S, et al.. Learning transferable visual models from natural language supervision. In: ICML. (2021). p. 1–36.

Google Scholar

41. Kirillov A, Mintun E, Ravi N, Mao H, Rolland C, Gustafson L, et al.. Segment anything. In: Proceedings of the IEEE/CVF International Conference on Computer Vision. (2023). p. 4015–26.

Google Scholar

42. Zhang S, Metaxas D. On the challenges and perspectives of foundation models for medical image analysis. Med Image Anal. (2024) 91:102996. doi: 10.1016/j.media.2023.102996

PubMed Abstract | Crossref Full Text | Google Scholar

43. Zhang G, Qi X, B BY, Wang G. IPLC: iterative pseudo label correction guided by SAM for source-free domain adaptation in medical image segmentation. In: Medical Image Computing and Computer Assisted Intervention. Switzerland: Springer Nature (2024). Vol. 1. p. 351–60.

Google Scholar

44. Zhong L, Liao X, Zhang S, Zhang X, Wang G. VLM-CPL: consensus pseudo labels from vision-language models for human annotation-free pathological image classification. arXiv:2403.15836 (2024).

Google Scholar

45. Wang G, Wu J, Luo X, Liu X, Li K, Zhang S. MIS-FM: 3D medical image segmentation using foundation models pretrained on a large-scale unannotated dataset. arXiv:2306.16925 (2023).

Google Scholar

46. Ronneberger O, Fischer P, Brox T. U-Net: convolutional networks for biomedical image segmentation. In: MICCAI. (2015). p. 234–41.

Google Scholar

47. Wang L, Xu Q, Leung S, Chung J, Chen B, Li S. Accurate automated Cobb angles estimation using multi-view extrapolation net. Med Image Anal. (2019) 58:101542. doi: 10.1016/j.media.2019.101542

PubMed Abstract | Crossref Full Text | Google Scholar

48. Luo X, Song T, Wang G, Chen J, Chen Y, Li K, et al.. SCPM-net: an anchor-free 3D lung nodule detection network using sphere representation and center points matching. Med Image Anal. (2022) 75:102287. doi: 10.1016/j.media.2021.102287

PubMed Abstract | Crossref Full Text | Google Scholar

49. Gal GL, Bounameaux H. Diagnosing pulmonary embolism: running after the decreasing prevalence of cases among suspected patients. J Thromb Haemost. (2004) 2:1244–6. doi: 10.1111/j.1538-7836.2004.00795.x

PubMed Abstract | Crossref Full Text | Google Scholar

50. Guo J, Liu X, Chen Y, Zhang S, Tao G, Yu H, et al.. AANet: artery-aware network for pulmonary embolism detection in CTPA images. In: Medical Image Computing and Computer Assisted Intervention – MICCAI 2022. (2022). p. 473–83.

Google Scholar

51. Kuo CFJ, Huang CC, Siao JJ, Hsieh CW, Huy VQ, Ko KH, et al.. Automatic lung nodule detection system using image processing techniques in computed tomography. Biomed Signal Process Control. (2020) 56:101659. doi: 10.1016/j.bspc.2019.101659

Crossref Full Text | Google Scholar

52. Zhu W, Liu C, Fan W, Xie X. DeepLung: deep 3D dual path nets for automated pulmonary nodule detection and classification. In: Proceedings – 2018 IEEE Winter Conference on Applications of Computer Vision, WACV 2018 (2018). p. 673–81.

Google Scholar

53. Ardimento P, Aversano L, Bernardi ML, Cimitile M. Deep neural networks ensemble for lung nodule detection on chest CT scans. In: Proceedings of the International Joint Conference on Neural Networks 2021-July. (2021). p. 1–8.

Google Scholar

54. Zheng S, Nie W, Pan L, Zheng B, Shen Z, Huang L, et al.. A dual-attention V-network for pulmonary lobe segmentation in CT scans. IET Image Process. (2021) 15:1644–54. doi: 10.1049/ipr2.12133

Crossref Full Text | Google Scholar

55. Xie W, Jacobs C, Charbonnier J-P, van Ginneken B. Relational modeling for robust and efficient pulmonary lobe segmentation in CT scans. IEEE Trans Med Imaging. (2020) 39:2664–75. doi: 10.1109/TMI.2020.2995108

PubMed Abstract | Crossref Full Text | Google Scholar

56. Wu R, Xin Y, Qian J, Dong Y. A multi-scale interactive U-Net for pulmonary vessel segmentation method based on transfer learning. Biomed Signal Process Control. (2023) 80:104407. doi: 10.1016/j.bspc.2022.104407

Crossref Full Text | Google Scholar

57. Wu Y, Qi S, Wang M, Zhao S, Pang H, Xu J, et al.. Transformer-based 3D U-Net for pulmonary vessel segmentation and artery-vein separation from CT images. Med Biol Eng Comput. (2023) 61:2649–63. doi: 10.1007/s11517-023-02872-5

PubMed Abstract | Crossref Full Text | Google Scholar

58. Huang SC, Kothari T, Banerjee I, Chute C, Ball RL, Borus N, et al.. PENet—a scalable deep-learning model for automated diagnosis of pulmonary embolism using volumetric CT imaging. npj Digit Med. (2020) 3:1–9. doi: 10.1038/s41746-020-0266-y

PubMed Abstract | Crossref Full Text | Google Scholar

59. Ma X, Ferguson EC, Jiang X, Savitz SI, Shams S. A multitask deep learning approach for pulmonary embolism detection and identification. Sci Rep. (2022) 12:1–11. doi: 10.1038/s41598-022-16976-9

PubMed Abstract | Crossref Full Text | Google Scholar

60. Qu H, Wu P, Huang Q, Yi J, Riedlinger GM, De S, et al.. Weakly supervised deep nuclei segmentation using points annotation in histopathology images. Proc Mach Learn Res. (2019) 12:1–12. doi: 10.1109/TMI.2020.3002244

Crossref Full Text | Google Scholar

61. Zhang X, Liu W, Dundar M, Badve S, Zhang S. Towards large-scale histopathological image analysis: hashing-based image retrieval. IEEE Trans Med Imaging. (2015) 34:496–506. doi: 10.1109/TMI.2014.2361481

PubMed Abstract | Crossref Full Text | Google Scholar

62. Li Z, Zhang X, Müller H, Zhang S. Large-scale retrieval for medical image analytics: a comprehensive review. Med Image Anal. (2018) 46:66–84. doi: 10.1016/j.media.2017.09.007

Crossref Full Text | Google Scholar

63. Zhang X, Xing F, Su H, Yang L, Zhang S. High-throughput histopathological image analysis via robust cell segmentation and hashing. Med Image Anal. (2015) 26:306–15. doi: 10.1016/j.media.2015.10.005

PubMed Abstract | Crossref Full Text | Google Scholar

64. Ba W, Wang S, Shang M, Zhang Z, Wu H, Yu C, et al.. Assessment of deep learning assistance for the pathological diagnosis of gastric cancer. Mod Pathol. (2022) 35:1262–8. doi: 10.1038/s41379-022-01073-z

PubMed Abstract | Crossref Full Text | Google Scholar

65. Huang B, Tian S, Zhan N, Ma J, Huang Z, Zhang C, et al.. Accurate diagnosis and prognosis prediction of gastric cancer using deep learning on digital pathological images: a retrospective multicentre study. EBioMedicine. (2021) 73:103631. doi: 10.1016/j.ebiom.2021.103631

PubMed Abstract | Crossref Full Text | Google Scholar

66. Ho C, Zhao Z, Chen XF, Sauer J, Saraf SA, Jialdasani R, et al.. A promising deep learning-assistive algorithm for histopathological screening of colorectal cancer. Sci Rep. (2022) 12:1–9. doi: 10.1038/s41598-022-06264-x

PubMed Abstract | Crossref Full Text | Google Scholar

67. Song Z, Yu C, Zou S, Wang W, Huang Y, Ding X, et al.. Automatic deep learning-based colorectal adenoma detection system and its similarities with pathologists. BMJ Open. (2020) 10:1–8. doi: 10.1136/bmjopen-2019-036423

Crossref Full Text | Google Scholar

68. Zhang X, Cao M, Wang S, Sun J, Fan X, Wang Q, et al.. Whole slide cervical cancer screening using graph attention network and supervised contrastive learning. In: Medical Image Computing and Computer Assisted Intervention. Switzerland: Springer Nature (2022). Vol. 1. p. 202–11.

Google Scholar

69. Cao M, Fei M, Xiong H, Zhang X, Fan X, Zhang L, et al.. Patch-to-sample reasoning for cervical cancer screening of whole slide image. IEEE Trans Artif Intell. (2024) 5:2779–89. doi: 10.1109/TAI.2023.3323637

Crossref Full Text | Google Scholar

70. Wang W, Xia Q, Hu Z, Yan Z, Li Z, Wu Y, et al.. Few-shot learning by a cascaded framework with shape-constrained pseudo label assessment for whole heart segmentation. IEEE Trans Med Imaging. (2021) 40:2629–41. doi: 10.1109/TMI.2021.3053008

PubMed Abstract | Crossref Full Text | Google Scholar

71. Tarvainen A, Valpola H. Mean teachers are better role models: weight-averaged consistency targets improve semi-supervised deep learning results. In: Advances in Neural Information Processing Systems. (2017). p. 1195–204.

Google Scholar

72. Wang W, Xia Q, Yan Z, Hu Z, Chen Y, Zheng W, et al.. AVDNet: joint coronary artery and vein segmentation with topological consistency. Med Image Anal. (2024) 91:102999. doi: 10.1016/j.media.2023.102999

PubMed Abstract | Crossref Full Text | Google Scholar

73. Van Hasselt H, Guez A, Silver D. Deep reinforcement learning with double Q-learning. In: AAAI. (2016). p. 2094–100.

Google Scholar

74. Qi CR, Su H, Mo K, Guibas LJ. PointNet: deep learning on point sets for 3D classification and segmentation. In: IEEE Conference on Computer Vision and Pattern Recognition. (2017). p. 77–85.

Google Scholar

75. Zreik M, Van Hamersvelt RW, Wolterink JM, Leiner T, Viergever MA, Išgum I. A recurrent CNN for automatic detection and classification of coronary artery plaque and stenosis in coronary CT angiography. IEEE Trans Med Imaging. (2019) 38:1588–98. doi: 10.1109/TMI.2018.2883807

PubMed Abstract | Crossref Full Text | Google Scholar

76. Zreik M, Lessmann N, van Hamersvelt RW, Wolterink JM, Voskuil M, Viergever MA, et al.. Deep learning analysis of the myocardium in coronary CT angiography for identification of patients with functionally significant coronary artery stenosis. Med Image Anal. (2018) 44:72–85. doi: 10.1016/j.media.2017.11.008

PubMed Abstract | Crossref Full Text | Google Scholar

77. Griffin WF, Choi AD, Riess JS, Marques H, Chang HJ, Choi JH, et al.. AI evaluation of stenosis on coronary CTA, comparison with quantitative coronary angiography and fractional flow reserve: a CREDENCE trial substudy. JACC: Cardiovasc Imaging. (2023) 16:193–205. doi: 10.1016/j.jcmg.2021.10.020

PubMed Abstract | Crossref Full Text | Google Scholar

78. Mu D, Bai J, Chen W, Yu H, Liang J, Yin K, et al.. Calcium scoring at coronary CT angiography using deep learning. Radiology. (2022) 302:309–16. doi: 10.1148/radiol.2021211483

PubMed Abstract | Crossref Full Text | Google Scholar

79. Hong JS, Tzeng YH, Yin WH, Wu KT, Hsu HY, Lu CF, et al.. Automated coronary artery calcium scoring using nested U-Net and focal loss. Comput Struct Biotechnol J. (2022) 20:1681–90. doi: 10.1016/j.csbj.2022.03.025

PubMed Abstract | Crossref Full Text | Google Scholar

80. Dai X, Huang L, Qian Y, Xia S, Chong W, Liu J, et al.. Deep learning for automated cerebral aneurysm detection on computed tomography images. Int J Comput Assist Radiol Surg. (2020) 15:715–23. doi: 10.1007/s11548-020-02121-2

PubMed Abstract | Crossref Full Text | Google Scholar

81. Hu B, Shi Z, Lu L, Miao Z, Wang H, Zhou Z, et al.. A deep-learning model for intracranial aneurysm detection on CT angiography images in China: a stepwise, multicentre, early-stage clinical validation study. Lancet Digit Health. (2024) 6:e261–71. doi: 10.1016/S2589-7500(23)00268-6

PubMed Abstract | Crossref Full Text | Google Scholar

82. Kim KH, Koo HW, Lee BJ, Yoon SW, Sohn MJ. Cerebral hemorrhage detection and localization with medical imaging for cerebrovascular disease diagnosis and treatment using explainable deep learning. J Korean Phys Soc. (2021) 79:321–7. doi: 10.1007/s40042-021-00202-2

Crossref Full Text | Google Scholar

83. Wang X, Shen T, Yang S, Lan J, Xu Y, Wang M, et al.. A deep learning algorithm for automatic detection and classification of acute intracranial hemorrhages in head CT scans. NeuroImage Clin. (2021) 32:102785. doi: 10.1016/j.nicl.2021.102785

PubMed Abstract | Crossref Full Text | Google Scholar

84. Wang F, Chen Q, Zhang Y, Chen Y, Zhou YZW, Liang X, et al.. CT-based radiomics for the recurrence prediction of hepatocellular carcinoma after surgical resection. J Hepatocellular Carcinoma. (2022) 23:453–65. doi: 10.2147/JHC.S362772

Crossref Full Text | Google Scholar

85. Wang S-H, Du J, Xu H, Yang D, Ye Y, Chen Y, et al.. Automatic discrimination of different sequences and phases of liver MRI using a dense feature fusion neural network: a preliminary study. Abdom Radiol. (2021) 46:4576–87. doi: 10.1007/s00261-021-03142-4

PubMed Abstract | Crossref Full Text | Google Scholar

86. Wang S-H, Han X-J, Du J, Wang Z-C, Yuan C, Chen Y, et al.. Saliency-based 3D convolutional neural network for categorising common focal liver lesions on multisequence MRI. Insights Imaging. (2021) 12:173. doi: 10.1186/s13244-021-01117-z

PubMed Abstract | Crossref Full Text | Google Scholar

87. Shaker AM, Maaz M, Rasheed H, Khan S, Yang MH, Khan FS. UNETR++: delving into efficient and accurate 3D medical image segmentation. IEEE Trans Med Imaging. (2024) 43:3377–90. doi: 10.1109/TMI.2024.3398728

PubMed Abstract | Crossref Full Text | Google Scholar

88. Chen J, Mei J, Li X, Lu Y, Yu Q, Wei Q, et al.. TransUNet: rethinking the U-Net architecture design for medical image segmentation through the lens of transformers. Med Image Anal. (2024) 97:103280. doi: 10.1016/j.media.2024.103280

PubMed Abstract | Crossref Full Text | Google Scholar

89. Zhou J, Wang W, Lei B, Ge W, Huang Y, Zhang L, et al.. Automatic detection and classification of focal liver lesions based on deep convolutional neural networks: a preliminary study. Front Oncol. (2021) 10:1–11. doi: 10.3389/fonc.2020.581210

Crossref Full Text | Google Scholar

90. Kim DW, Lee G, Kim SY, Ahn G, Lee JG, Lee SS, et al.. Deep learning—based algorithm to detect primary hepatic malignancy in multiphase CT of patients at high risk for HCC. Eur Radiol. (2021) 31:7047–57. doi: 10.1007/s00330-021-07803-2

PubMed Abstract | Crossref Full Text | Google Scholar

91. Xu M, Wang Y, Chi Y, Hua X. Training liver vessel segmentation deep neural networks on noisy labels from contrast CT imaging. In: Proceedings – International Symposium on Biomedical Imaging. (2020). p. 1552–5.

Google Scholar

92. Fu J, Liu J, Tian H, Fang Z, Lu H. Dual attention network for scene segmentation. In: CVPR. (2019). p. 3146–54.

Google Scholar

93. Rivest-Hénault D, Dowson N, Greer PB, Fripp J, Dowling JA. Robust inverse-consistent affine CT-MR registration in MRI-assisted and MRI-alone prostate radiation therapy. Med Image Anal. (2015) 23:56–69. doi: 10.1016/j.media.2015.04.014

PubMed Abstract | Crossref Full Text | Google Scholar

94. Qu Y, Li X, Yan Z, Zhao L, Zhang L, Liu C, et al.. Surgical planning of pelvic tumor using multi-view CNN with relation-context representation learning. Med Image Anal. (2021) 69:101954. doi: 10.1016/j.media.2020.101954

PubMed Abstract | Crossref Full Text | Google Scholar

95. Chen Y, Yang J, Qian J. Recurrent neural network for facial landmark detection. Neurocomputing. (2017) 219:26–38. doi: 10.1016/j.neucom.2016.09.015

Crossref Full Text | Google Scholar

96. O’Neil AQ, Kascenas A, Henry J, Wyeth D, Shepherd M, Beveridge E, et al.. Attaining human-level performance with atlas location autocontext for anatomical landmark detection in 3D CT data. In: European Conference on Computer Vision. (2018). p. 470–84.

Google Scholar

97. Kissela BM, Khoury JC, Alwell K, Moomaw CJ, Woo D, Adeoye O, et al.. Age at stroke: temporal trends in stroke incidence in a large, biracial population. Neurology. (2012) 79:1781–7. doi: 10.1212/WNL.0b013e318270401d

PubMed Abstract | Crossref Full Text | Google Scholar

98. Wang G, Song T, Dong Q, Cui M, Huang N, Zhang S. Automatic ischemic stroke lesion segmentation from computed tomography perfusion images by image synthesis and attention-based deep neural networks. Med Image Anal. (2020) 65:1–14. doi: 10.1016/j.media.2020.101787

Crossref Full Text | Google Scholar

99. Fu F, Wei J, Zhang M, Yu F, Xiao Y, Rong D, et al.. Rapid vessel segmentation and reconstruction of head and neck angiograms using 3D convolutional neural network. Nat Commun. (2020) 11:4829. doi: 10.1038/s41467-020-18606-2

PubMed Abstract | Crossref Full Text | Google Scholar

100. Fu H, Wang G, Lei W, Xu W, Zhao Q, Zhang S, et al.. HMRNet: high and multi-resolution network with bidirectional feature calibration for brain structure segmentation in radiotherapy. IEEE J Biomed Health Inform. (2022) 26:4519–29. doi: 10.1109/JBHI.2022.3181462

PubMed Abstract | Crossref Full Text | Google Scholar

101. Gao Y, Huang R, Yang Y, Zhang J, Shao K, Tao C, et al.. FocusNetv2: imbalanced large and small organ segmentation with adversarial shape constraint for head and neck CT images. Med Image Anal. (2021) 67:101831. doi: 10.1016/j.media.2020.101831

PubMed Abstract | Crossref Full Text | Google Scholar

102. Lei W, Mei H, Sun Z, Ye S, Gu R, Wang H, et al.. Automatic segmentation of organs-at-risk from head-and-neck CT using separable convolutional neural network with hard-region-weighted loss. Neurocomputing. (2021) 442:184–99. doi: 10.1016/j.neucom.2021.01.135

Crossref Full Text | Google Scholar

103. Liu Z, Wang H, Lei W, Wang G. CSAF-CNN: cross-layer spatial attention map fusion network for organ-at-risk segmentation in head and neck CT images. In: International Symposium on Biomedical Imaging. (2020). p. 1522–5.

Google Scholar

104. Liu Z, Sun C, Wang H, Li Z, Gao Y, Lei W, et al.. Automatic segmentation of organs-at-risks of nasopharynx cancer and lung cancer by cross-layer attention fusion network with TELD-Loss. Med Phys. (2021) 48:6987–7002. doi: 10.1002/mp.15260

PubMed Abstract | Crossref Full Text | Google Scholar

105. Huang R, Zheng Y, Hu Z, Zhang S, Li H. Multi-organ segmentation via co-training weight-averaged models from few-organ datasets. In: International Conference on Medical Image Computing and Computer-Assisted Intervention. (2020). p. 146–55.

Google Scholar

106. Luo X, Liao W, Xiao J, Chen J, Song T, Zhang X, et al.. WORD: a large scale dataset, benchmark and clinical applicable study for abdominal organ segmentation from CT image. Med Image Anal. (2022) 82:102642. doi: 10.1016/j.media.2022.102642

PubMed Abstract | Crossref Full Text | Google Scholar

107. Yang Y, Huang R, Lv G, Hu Z, Shan G, Zhang J, et al.. Automatic segmentation of the clinical target volume and organs at risk for rectal cancer radiotherapy using structure-contextual representations based on 3D high-resolution network. Biomed Signal Process Control. (2022) 73:103362. doi: 10.1016/j.bspc.2021.103362

Crossref Full Text | Google Scholar

108. Wang X, Zhang X, Wang G, He J, Li Z, Zhu W, et al.. OpenMEDLab: an open-source platform for multi-modality foundation models in medicine. arXiv:2402.18028 (2024).

Google Scholar

Keywords: artificial intelligence, medical image informatics, clinical research platform, data annotation, model training

Citation: Wang G, Duan Q, Shen T and Zhang S (2024) SenseCare: a research platform for medical image informatics and interactive 3D visualization. Front. Radiol. 4:1460889. doi: 10.3389/fradi.2024.1460889

Received: 7 July 2024; Accepted: 6 November 2024;
Published: 21 November 2024.

Edited by:

Yassine Himeur, University of Dubai, United Arab Emirates

Reviewed by:

Adel Oulefki, University of Sharjah, United Arab Emirates
Yassine Habchi, University Center Salhi Ahmed Naama, Algeria

Copyright: © 2024 Wang, Duan, Shen and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.

*Correspondence: Shaoting Zhang, zhangshaoting@sensetime.com

Disclaimer: All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article or claim that may be made by its manufacturer is not guaranteed or endorsed by the publisher.