- 1School of Computer Science and Engineering, University of Electronic Science and Technology of China, Chengdu, China
- 2Institute of Intelligent Manufacturing, Shenzhen Polytechnic University, Shenzhen, China
- 3Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences, Shenzhen, China
- 4Computer Engineering and Software Engineering Department, Polytechnique Montreal, Montreal, QC, Canada
- 5Islamabad Medical Complex, NESCOM Hospital, Islamabad, Pakistan
- 6Department of Computer Science, National University of Computer and Emerging Sciences, Chiniot-Faisalabad Campus, Chiniot, Pakistan
- 7Department of Computer Science and Software Technology, University of Swat, Saidu Sharif, Pakistan
- 8Department of Information Systems, College of Computer Engineering and Sciences, Prince Sattam Bin Abdulaziz University, Alkharj, Saudi Arabia
Introduction: In the evolving healthcare landscape, we aim to integrate hyperspectral imaging into Hybrid Health Care Units to advance the diagnosis of medical diseases through the effective fusion of cutting-edge technology. The scarcity of medical hyperspectral data limits the use of hyperspectral imaging in disease classification.
Methods: Our study innovatively integrates hyperspectral imaging to characterize tumor tissues across diverse body locations, employing the Sharpened Cosine Similarity framework for tumor classification and subsequent healthcare recommendation. The efficiency of the proposed model is evaluated using Cohen's kappa, overall accuracy, and f1-score metrics.
Results: The proposed model demonstrates remarkable efficiency, with kappa of 91.76%, an overall accuracy of 95.60%, and an f1-score of 96%. These metrics indicate superior performance of our proposed model over existing state-of-the-art methods, even in limited training data.
Conclusion: This study marks a milestone in hybrid healthcare informatics, improving personalized care and advancing disease classification and recommendations.
1 Introduction
Our research explores the utilization of hyperspectral imaging (HI) to revolutionize tumor tissue classification in various body regions, aiming to impact the medical field significantly. This approach promises to refine diagnostic accuracy and pave the path for more personalized treatment plans. Taking a step toward the era of highly personalized, adequate healthcare, our study aims to enhance patient care. The reason HI is utilized for disease diagnosis is grounded in the understanding that changes in tissue's optical properties, stemming from morphological and biochemical alterations during disease progression, can be detected (1). For instance, rapid cell division in malignant cells leads to increased metabolic enzyme levels and the formation of new vessels through angiogenesis to meet the demand for nutrients and oxygen (2).
HI capitalizes on these changes to identify lesions and abnormal tissue without needing histological examination, saving time and improving treatment efficacy. Biopsy samples, which are stable and easily obtained from patients, implement scanning HI feasible. Recent studies have explored correlations between HI and histological examination results to validate HI as an accurate disease diagnostic tool. Various tissues, including the breast (3), liver (4), brain (5), kidney (6), stomach (7), head and neck (8), and thyroid gland (9), have been investigated, demonstrating HI's capability for disease diagnosis. The complexity of HI is addressed by employing artificial intelligence, which exhibits comparable diagnostic accuracy compared to histology.
One notable advantage of HI-based disease diagnosis is its ability to directly examine biopsy tissue during surgery. Unlike histology, which typically takes hours, HI can analyze tissue within minutes. This rapid analysis enables real-time assessment of resection margins to check for residual tumor tissues. In a study, HI successfully identified breast cancer from excised breast tissue during surgery with an accuracy exceeding 84% (10). Additionally, HI has found application in identifying blood cells, showcasing its potential to delineate abnormal tissue without relying on biochemical techniques (11). These applications underscore the capacity of HI to support swift and accurate decision-making in clinical settings. Our research contributes significantly to the field in addressing the pressing need for more adaptable and precise tumor classification in healthcare diagnostics. The following points outline the key contributions made in this study:
• Versatile tumor classification: introduces a hyperspectral imaging-based classifier offering location-independent and adaptable tumor classification, surpassing the limitations of existing methods.
• Sharpened Cosine Similarity (SCS): SCS is proposed as an innovative technique within the hyperspectral imaging classification framework, demonstrating superior precision and efficiency for tumor classification, especially under limited training data.
• Empirical evaluation: provides a rigorous empirical evaluation of the proposed model, substantiating its superior performance through metrics like Cohen's kappa, overall accuracy, and f1-score.
• Hybrid Health Care (HHC) integration: applies hyperspectral imaging classification within HHC Units, contributing to personalized and effective medical care solutions with broader implications for healthcare informatics.
2 Literature review
Traditional imaging techniques like Magnetic Resonance Imaging (MRI) (12), Computed Tomography (CT) Scans (13), Positron Emission Tomography (PET) Scans (14), Functional MRI (fMRI) (15), and Magnetic Resonance Spectroscopy (MRS) (16) have their own set of challenges in tumor detection (17). While these methods are indispensable, their specificity to specific tumor types hinders widespread application. Furthermore, implementing advanced deep learning algorithms presents scalability and real-time processing issues in clinical environments (18). Addressing these limitations, our approach offers a more versatile and computationally efficient alternative, enhancing its potential for clinical integration.
Elaborating on existing imaging modalities, MRI stands out for its high sensitivity (90%–95%) in brain tumor detection but grapples with the risk of false results and limitations in pinpointing specific tumor types or smaller lesions (19). CT Scans, utilizing X-rays, exhibit a sensitivity range of 60%–90% and a specificity of ~90%. Still, the method is constrained by radiation risks and less detailed soft tissue imaging (20). PET Scans employing ionizing radiation show varying sensitivity (70%–90%) and reasonable specificity (80%–90%), yet are subject to sensitivity limitations due to tumor characteristics and tracer use (21). fMRI, indicating brain activity through blood flow, offers high sensitivity (80%–90%) and specificity in identifying key brain areas but is susceptible to motion artifacts and variable interpretation (22). MRS provides a window into the biochemical makeup of tissues, yielding crucial data on tumor metabolism and types (23). Each modality contributes uniquely to tumor diagnosis, balancing specific advantages and inherent challenges.
Tumors, formed when cells behave abnormally, exhibit a range of sizes and can emerge anywhere in the body. Genes mutation, whether inherited, acquired gradually, or induced by substances like alcohol and tobacco, transform cells into cancerous ones (24). Growing tumors can invade neighboring tissues, displace normal cells, and produce enzymes breaking down surrounding tissues. Local invasion occurs when tumors grow larger, and metastasis happens when cancer cells spread to other body parts through blood or lymphatics (25). Classification involves categorizing tumors broadly by tissue, organ, or system, specifically by type, grading based on cellular and structural features using the World Health Organization (WHO) system, and staging using the Tumor Node Metastasis (TNM) system (26). Solid neoplasms, including carcinomas, sarcomas, and lymphomas, are classified based on type. The WHO Classification of Tumors provides detailed insights into tumor histotypes across various organ systems (27). According to the WHO system, tumor grading assigns a numerical grade (1–3) based on cellular differentiation. Staging relies on the TNM system, considering the presence of distant metastases (M), lymph node involvement (N), and the size or extension of the primary tumor (T) (28).
The most common cause of cancer death among children under the age of 15 and the second fastest-growing cause of cancer death among those over the age of 65 are brain tumors, which originate in brain cells and may be benign or malignant (29). Gene defects, exposure to certain chemicals, and radiation therapy to the head increase the risk of these tumors (30). Gliomas, the most common type, form from neural cells, including astrocytomas and ependymomas (31). Other types, such as brain stem gliomas, optic nerve gliomas, primitive neuroectodermal tumors (PNET), medulloblastomas, craniopharyngiomas, and pineal region tumors, pose distinct challenges in terms of location and characteristics (32). Understanding these variations is crucial for tailored treatment approaches and underscores the complexity of brain tumor classification and detection.
Moreover, Lung carcinoma, or lung cancer (33), results from genetic mutations in airway cells triggered by factors like smoking (34). It manifests as non-small-cell lung cancer (85%) and small-cell lung cancer (15%) (35). Breast cancer originates from mutated breast cells, often spreading invasively, with common types being lobular, ductal carcinoma in situ (DCIS), and invasive ductal carcinoma (IDC) (36, 37). Meningiomas, arising from brain membranes, may compress nearby tissues, and their slow growth lacks a defined cause (38). HI stands poised to revolutionize tumor classification and identification by capturing unique optical properties associated with different tumor types (39). HI offers a non-invasive and potentially rapid method for precise diagnosis, contributing to improved treatment strategies and patient outcomes in lung, breast, and meningiomas.
3 Materials and methods
3.1 Dataset
The dataset we used to conduct experiments was initially collected and published by the In-vivo HS Human Brain database (40) comprising 36 in-vivo brain surface images from 22 unique patients. This labeled dataset includes tumor and normal tissue, blood vessels, and other irrelevant materials within the surgical scene (referred to as background). Tumor types are differentiated in the dataset, encompassing primary (grade IV glioblastoma and grade III and II anaplastic oligodendrogliomas) and secondary tumors (lung and breast). Additionally, RGB representations of hyperspectral cubes within the in-vivo hyperspectral human brain image database are presented in Figure 1. The dataset designates the approximate tumor area using a yellow line, aiding in identifying the rubber ring marker corresponding to pathological analyses of the tumor tissue. Patient ID and Image ID details in Table 1 offer a comprehensive overview, including image characteristics and the pathological diagnosis of each image. The total number of labeled pixels for each class and image is specified, addressing cases where certain images were diagnosed as specific tumor types without labeled tumor samples due to procedural challenges.
Figure 1. RGB representations of dataset images with PatientID-ImageID codes, delineating approximate tumor areas guided by neurosurgeon expertise and the IGS system.
Table 1. Patient-specific image data and label distribution where “N” refers to “normal” tissue, “T” signifies “tumor tissue,” “BV” represents “blood vessels,” and “B” denotes the “background.”
The authors (40) mention the inherent challenges in acquiring in-vivo HI during neurosurgical procedures; the dataset primarily captures common tumor types over two years. The customized hyperspectral acquisition system, a preliminary demonstrator, is designed to capture tumor images on the surface or in easily focused deeper layers. The authors utilize a push broom camera for spatial scanning; the system's limitations include increased acquisition time and potential spatial coherence issues due to patient brain movement and procedural artifacts. As snapshot cameras offer real-time image acquisition but have fewer spectral bands than push-broom cameras, future investigations using high spectral resolution push-broom cameras are warranted. The dataset creation process by authors (40) addresses challenges from limited patient availability, presenting a preliminary database for exploring HI applications in tissue and tumor identification, tumor boundary delineation, and providing pertinent information for neurosurgeons. Their methodology leverages spectral characteristics guided by intraoperative MRI, surgeon expertise, and pathological analysis results. Subsequent data acquisition efforts are anticipated to broaden the database, encompassing more tumor types with detailed pathological descriptions.
3.2 HHC: AI tumor diagnostics
Our innovative methodology for tumor tissue classification within an HHC Unit unfolds with the patient's arrival at the facility. The initial phase involves a hyperspectral sensor scan, capturing intricate details of the patient's internal composition. This technology provides a comprehensive overview, laying the foundation for precise diagnosis. Following the hyperspectral scan, the acquired data undergoes processing through Factor Analysis. This step is crucial for dimension reduction, ensuring that the hyperspectral cube retains only relevant features essential for accurate classification. The processed data then traverses through the layers of our SCS model. As a breakthrough in tumor classification, the SCS model enhances precision, even when trained with limited data. This stage is pivotal for predicting and classifying tumor tissues, contributing to superior performance compared to existing models.
Once the classification is complete, the results are securely stored within the hospital's private records, ensuring data confidentiality. This stored information becomes a valuable resource for future reference and analysis. Integrated into the HHC Unit is a seamless access mechanism through Healthcare APIs. Healthcare professionals can leverage these APIs to access detailed reports and results related to tumor tissues. This integration streamlines the diagnostic process, providing a user-friendly interface for medical interpretation.
The final phases of our methodology involve the medical interpreter within the Healthcare API, aiding healthcare professionals in interpreting results and making informed recommendations. These recommendations extend to surgical interventions and ongoing medical care, all tailored to the specific classification of tumor tissues and their respective locations. Figure 2 presents a comprehensive and patient-centric approach to tumor tissue classification within the HHC Unit. By seamlessly integrating hyperspectral imaging, Factor Analysis, and the innovative SCS model, we aim to revolutionize healthcare diagnostics and enhance the overall patient experience.
Figure 2. Streamlined tumor diagnosis in hybrid healthcare: a patient-centric approach from initial scan to tailored treatment.
3.3 Proposed Sharpened Cosine Similarity method
HI represented as X∈ℝ(M×N) × B, where the dimensions (M×N) correspond to a specific area on the tissue surface and B denotes the total number of spectral bands in the HI. Each pixel within X, indicated as xij where i = 1, 2, …, M and j = 1, 2, …, N, is grouped into C unique tissue types, collectively expressed as Y = (y1, y2, …, yn). Moreover, every xij∈X describes a tissue pixel through a spectral vector xij = [xi, j, 1, xi, j, 2, …, xi, j, B]∈X, containing a series of B spectral data points.
In the initial processing phase, spatial characteristics are emphasized by implementing a patch extraction method. This preliminary step involves the creation of a hyperspectral cube, , encapsulating the area surrounding the focal pixel (i, j) over a region of dimensions s×s. This approach is instrumental in enhancing the model's ability to distinguish between different features by integrating spectral and spatial attributes. As such, the spectral-spatial cubes xi, j, drawn from the primary data and conforming to the dimensionality ℝ(s×s) × D, are consolidated into the dataset X in preparation for subsequent feature extraction processes. The concluding step involves the selection of training and testing samples across each distinct class.
In neural networks, the convolution operation involves a sliding dot product operation, symbolized as w·xij, between an image patch xij and a filter w, which might miss crucial information due to its basic similarity measure. Enhancing this with normalization transforms the operation into cosine similarity, defined as . This is similar to calculating the cosine of the angle between vectors, utilizing Euclidean distance.
To address these limitations, Strided Cosine Similarity (SCS) was developed as expressed in Equation (1). It operates similarly to convolution but includes key differences. In standard convolution, the operation is a dot product w·xij, while SCS involves normalizing the vectors. The normalization in SCS ensures the magnitude of vectors is unity before the dot product, leading to an expression like , where q is a small value to avoid numerical instability.
The similarity measure in SCS ranges between –1 and 1, indicating complete opposition or perfect alignment of the kernel and image patch, respectively. To mitigate the issue of small magnitudes, which can lead to noise inclusion, additional parameters are introduced in SCS, formulated as;
Similar to conventional convolution in deep learning, SCS is a striding operation that extracts features from an image patch. However, it includes an additional step of magnitude normalization before the dot product, leading to what some literature refers to as Sharpened Cosine normalization. The effectiveness of SCS surpasses traditional convolutional processes in terms of speed due to fewer required parameters and the absence of normalization or activation functions.
In contrast to standard pooling, absolute max-pooling is employed in SCS for backpropagation filter updates, selecting the highest magnitude irrespective of the sign. The overall model with SCS is trained over 50 epochs, a batch size of 256, and a learning rate of 0.001. The learning rate significantly influences the model's learning rate, while momentum aids accuracy and speed. An root mean square prop and momentum-based optimizer, specifically the Adam optimizer, is utilized for its efficiency and computational advantages.
4 Experiment analysis
This section presents an overview of the evaluation metrics, baselines SOTA and implementation details.
4.1 Evaluation metrics
The results presented in this study are evaluated using the following metrics:
Kappa statistic: This statistical measure assesses the level of agreement between predicted classifications and ground-truth maps, as defined by Equation (2). In this equation, Ao represents the observed agreement, calculated using Equation (3), while Ae denotes the expected agreement, computed using Equation (4).
where,
and,
Here, TP and FP denote true positives and false positives, respectively, while TN and FN represent true negatives and false negatives.
Average accuracy (AA): AA signifies the average classification performance across different classes, as depicted in Equation (5).
Overall accuracy (OA): OA is computed as the ratio of correctly classified examples to the total number of test examples, as defined by Equation (6).
In the equations above, TP represents true positives, FP represents false positives, TN represents true negatives, and FN represents false negatives.
4.2 Baseline models
4.2.1 Recurrent Neural Networks
The Recurrent Neural Networks (RNN) architecture (41) presents a blend of convolutional and fully connected layers within a Sequential model. Beginning with a Conv2D layer employing a 3 × 3 kernel and ReLU activation, the subsequent MaxPooling2D layer downsamples the spatial dimensions. Flattening the output precede a fully connected layer of 100 neurons, integrated with Batch Normalization and ReLU activation for regularization. With softmax activation, the final layer tailors the output to fit the specified number of classes. This design reflects a hybrid approach, incorporating convolutional operations followed by dense layers, offering flexibility for various applications in classification tasks.
4.2.2 2-Dimensional Convolution Neural Network
The 2-Dimensional Convolution Neural Network (2D CNN) architecture (42) is structured within a Sequential model, featuring a Conv2D layer with a 3 × 3 kernel and ReLU activation, applied to input data of shape (window size, window size, kernel size). Subsequently, a MaxPooling2D layer down-samples spatial dimensions with a pooling size adjustment option. The flattened output leads to a fully connected layer with 100 neurons, supplemented by Batch Normalization and ReLU activation for regularization. The final layer, employing softmax activation, tailors the output to match the specified number of classes. This design reflects a standard 2D convolutional neural network suitable for diverse classification tasks with image data. Adjustments to the pooling size provide adaptability based on specific requirements.
4.2.3 LeNet
The LeNet architecture, a seminal convolutional neural network devised by Yann LeCun in the 1990s, marked a breakthrough in computer vision (43). Comprising two convolutional layers with 5 × 5 filters and ReLU activation, each succeeded by average pooling; the network captures hierarchical features in the input. The subsequent dense layers, with 120 and 84 neurons, distill high-level representations. The final layer, employing softmax activation, tailors the output to the number of classes. LeNet's simplicity and efficacy laid the foundation for modern CNNs, influencing subsequent developments in image classification.
4.2.4 Xception
The Xception architecture (44) is a 2D variant of the Xception neural network, known for its depth-wise separable convolutions and exceptional performance in image classification tasks. The model begins with an entry flow featuring a series of convolutional layers with batch normalization and ReLU activation. The residual block 1 introduces separable convolutions, preserving spatial information efficiently. The middle flow comprises eight repeated blocks, each containing three separable convolutional layers, facilitating feature extraction. The exit flow further refines features with a combination of separable convolutions and residual connections. The model concludes with a global average pooling layer and a dense layer with softmax activation, tailoring the output to the specified number of classes. The Xception architecture is designed to capture complex hierarchical features in image data, making it suitable for various image classification tasks. Adjustments to the number of filters and other parameters can be made based on specific requirements.
4.3 Implementation details
For our empirical assessment, we utilized in-vivo HS Human Brain database which is already discussed in Section 3.1 and accessible on request on this https://hsibraindatabase.iuma.ulpgc.es/. This experiment used a Jupyter notebook running on an Intel 11th Gen processor and 32 GB of RAM. For all experiments, the training, validation, and test samples distribution was set at 15%, 15%, and 70%, respectively. To ensure an equitable comparison, all models, including the RNN, 2D CNN, LeeNet, Xception, and proposed SCS models, were executed simultaneously with a single, randomly chosen set of samples. The reported results were achieved using a patch size of 3 × 3, and the three most informative bands were identified through Factor Analysis (FA). Regarding training parameters, the models began with randomly initialized weights, which were subsequently optimized via backpropagation using the Adam optimizer and a softmax loss function. Figure 3 present a detailed analysis of the validation loss and accuracy for all models under consideration. In this study, we adhere to this principle by keeping these parameters uniform across all compared methods, including our SCS pipeline, within a single execution run.
5 Discussion
In this section, we conduct a twofold comparative analysis to evaluate the performance of our SCS pipeline for the Hybrid Healthcare Unit. Firstly, at the patient level, we assess the system's efficacy in providing personalized tumor tissue classifications and treatment recommendations. Subsequently, at the same tumor class level, we analyze the system's precision in distinguishing minute variations within specific classes. These comparative experiments aim to comprehensively understand the Hybrid Healthcare Unit's capabilities, addressing individual patient needs and the challenges within distinct tumor classes.
5.1 Comparative experiment—Class level
We present a comprehensive performance analysis based on the F1-Score, comparing SOTA, including 2D CNN, RNN, LeeNet, Xception, and our proposed SCS across different patients. The objective of this comparative experiment class level is to evaluate and compare the performance of these models in accurately classifying different tissue types in HI as results discussed in Table 2. Across different patients, our SCS consistently achieves high accuracy in predicting tissue classes, as presented in the Table 3. Notably, for Patient IDs 005, 008, 022, 028, and 029, SCS achieves exceptional accuracy close to or at 100% in classifying normal tissue, hypervascularized tissue, and background classes. This demonstrates the model's robustness in handling diverse cases. In cases where tumor tissue is present, the SCS model also successfully achieves accurate predictions comparison with other models (2D CNN, RNN, LeeNet, and Xception 2D). The model's effectiveness in leveraging SCS-enhanced features for accurate tissue classification irrespective of body location are shown in classification map outputs (Figure 4). These sub-figures correspond to different tumor tissue types: Normal Brain, Renal Carcinoma, Lung Carcinoma, Meningioma, and Lung Adenocarcinoma. The model's ability to handle various tissue classes and consistent accuracy across different patients and images highlight its potential as a valuable tool in medical diagnostics, particularly for tumor tissue classification.
Table 2. Performance analysis of the SOTA models on each predicted class for different patient IDs and image IDs.
Table 3. Comparative performance analysis of SOTA at patient and image level for each predicted class.
5.2 Comparative experiment—Patient level
A detailed analysis of the performance of SOTA models on key metrics, including Kappa Accuracy, Overall Accuracy, Average Accuracy, F1-Score, Training Time, Testing Time, and Memory Consumption. Table 3 summarizes the performance metrics for each patient and their corresponding image IDs across various tissue classes. From the patient-level experiment, the SCS model consistently outperformed both models across multiple performance metrics. For instance, in Patient ID 004, the SCS model achieved a Kappa Accuracy of 80.53, surpassing 2D CNN (49.93), RNN (47.66), LeeNet (71.67) and Xception (30.02). Similar trends were observed regarding Overall Accuracy, Average Accuracy, and F1 score, where the SCS model consistently demonstrated superior performance across all patient IDs. Notably, in Patient ID 021, the SCS model achieved a Kappa Accuracy of 98.55, significantly surpassing 2D CNN (85.12), RNN (45.74), LeeNet (97.83), and Xception (38.61). SCS model's ability to consistently attain high accuracy, coupled with efficient training times and memory consumption, underscores its potential for accurate tissue classification in HI data, highlighting its value in practical medical applications. Although other models such as 2D CNN, RNN, and LeeNet have less training time, their accuracy is low compared to the SCS model; as we know, in deep learning, there is a trade-off between speed and accuracy. Figure 5 shows results underscore the superior performance of the SCS model across various metrics, indicating its efficacy in accurately classifying tissue types in HI data. The consistent out performance of the SCS model reaffirms its potential to enhance medical diagnostics and contribute to real-world applications.
6 Conclusion
Our research highlights the pivotal role of HI integrated with AI in advancing tumor tissue classification with the new Hybrid Health Care Units landscape. The innovative application of the Sharpened Cosine Similarity framework has proven highly effective, achieving remarkable performance metrics of 91.76% Cohen's kappa, 95.60% overall accuracy, and 94.29% f1-score. These results, surpassing current SOTA research even under limited training data, affirm our proposed model's robustness and potential clinical impact. The scarcity of specific hyperspectral medical data has been acknowledged as a challenge, emphasizing the need for ongoing efforts to expand and diversify datasets for further validation and generalization of our approach. However, the demonstrated superiority of our model in tumor classification positions it as a valuable tool for enhancing diagnostic capabilities in medical imaging. Future research could extend the proposed model by diversifying and expanding hyperspectral medical datasets for broader validation. Exploring real-time implementation in clinical settings and investigating additional AI techniques could enhance predictive capabilities. Furthermore, exploring broader applications beyond tumor classification, such as skin conditions, could maximize the model's utility. These efforts would advance healthcare informatics, improving diagnostic accuracy within Hybrid Health Care Units.
Data availability statement
The datasets presented in this study can be found in online repositories. The names of the repository/repositories and accession number(s) can be found at: https://hsibraindatabase.iuma.ulpgc.es.
Author contributions
MB: Visualization, Methodology, Conceptualization, Writing—review & editing, Writing—original draft. JL: Writing—review & editing, Supervision. JJ: Writing—review & editing, Supervision. WR: Writing—original draft, Funding acquisition. NA: Writing—original draft, Methodology, Formal analysis. FB: Writing—original draft, Validation, Data curation. MA: Writing—review & editing, Methodology, Conceptualization. AS: Writing—original draft, Visualization. AA: Writing—review & editing, Formal analysis, Validation. MU: Writing—review & editing, Supervision, Validation.
Funding
The author(s) declare that financial support was received for the research, authorship, and/or publication of this article. The research conducted for this paper has been made possible through the generous support of several funding sources. Specifically, we acknowledge the stable support project of Shenzhen colleges and universities (20220811224429001), the Post-doctoral Later-stage Foundation Project of Shenzhen Polytechnic (6021271014K), and the Shenzhen Polytechnic Scientific Research Start-up Project (6022312028K). Furthermore, our work has been supported by the National Science Foundation of Guangdong Province (2022A1515011724) and the Featured Innovation Project of Guangdong Provincial Department of Education (2021KTSCX275). These funding sources have played a crucial role in enabling our research endeavors and advancing our understanding in the field of healthcare informatics.
Acknowledgments
The authors would like to express their sincere appreciation to FF Butt, a dedicated medical professional at NESCOM Hospital in Islamabad, Pakistan. FF Butt played a pivotal role in providing invaluable insights into the domain knowledge concerning various types of tumor tissues within the human body. Her expertise extended to a comprehensive understanding of the existing medical imaging modalities utilized for tumor tissue classification, along with a thorough grasp of the historical and background aspects. The authors acknowledge FF Butt's significant contributions, which greatly enhanced the depth and quality of this research.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Publisher's note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
1. Bashkatov AN, Genina EA, Tuchin VV. Optical properties of skin, subcutaneous, and muscle tissues: a review. J Innov Opt Health Sci. (2011) 4:9–38. doi: 10.1142/S1793545811001319
2. Kroemer G, Pouyssegur J. Tumor cell metabolism: cancer's Achilles' heel. Cancer Cell. (2008) 13:472–82. doi: 10.1016/j.ccr.2008.05.005
3. Aboughaleb IH, Aref MH, El-Sharkawy YH. Hyperspectral imaging for diagnosis and detection of ex-vivo breast cancer. Photodiagnosis Photodyn Ther. (2020) 31:101922. doi: 10.1016/j.pdpdt.2020.101922
4. Zhang Y, Yu S, Zhu X, Ning X, Liu W, Wang C, et al. Explainable liver tumor delineation in surgical specimens using hyperspectral imaging and deep learning. Biomed Opt Express. (2021) 12:4510–29. doi: 10.1364/BOE.432654
5. Ortega S, Halicek M, Fabelo H, Camacho R. Plaza MdlL, Godtliebsen F, et al. Hyperspectral imaging for the detection of glioblastoma tumor cells in H&E slides using convolutional neural networks. Sensors. (2020) 20:1911. doi: 10.3390/s20071911
6. Lv M, Chen T, Yang Y, Tu T, Zhang N, Li W, et al. Membranous nephropathy classification using microscopic hyperspectral imaging and tensor patch-based discriminative linear regression. Biomed Opt Express. (2021) 12:2968–78. doi: 10.1364/BOE.421345
7. Liu N, Guo Y, Jiang H, Yi W. Gastric cancer diagnosis using hyperspectral imaging with principal component analysis and spectral angle mapper. J Biomed Opt. (2020) 25:066005. doi: 10.1117/1.JBO.25.6.066005
8. Halicek M, Dormer JD, Little JV, Chen AY, Myers L, Sumer BD, et al. Hyperspectral imaging of head and neck squamous cell carcinoma for cancer margin detection in surgical specimens from 102 patients using deep learning. Cancers. (2019) 11:1367. doi: 10.3390/cancers11091367
9. Halicek M, Dormer JD, Little JV, Chen AY, Fei B. Tumor detection of the thyroid and salivary glands using hyperspectral imaging and deep learning. Biomed Opt Express. (2020) 11:1383–400. doi: 10.1364/BOE.381257
10. Kho E. Hyperspectral Imaging for Resection Margin Assessment during Breast Cancer Surgery [PhD Thesis - Research external, graduation UT]. University of Twente, Enschede (2020).
11. Huang Q, Li W, Zhang B, Li Q, Tao R, Lovell NH. Blood cell classification based on hyperspectral imaging with modulated Gabor and CNN. IEEE J Biomed Health Inform. (2019) 24:160–70. doi: 10.1109/JBHI.2019.2905623
12. Van der Leest M, Israel B, Cornel EB, Zamecnik P, Schoots IG, van der Lelij H, et al. High diagnostic performance of short magnetic resonance imaging protocols for prostate cancer detection in biopsy-naïve men: the next step in magnetic resonance imaging accessibility. Eur Urol. (2019) 76:574–81. doi: 10.1016/j.eururo.2019.05.029
13. Huang S, Yang J, Shen N, Xu Q, Zhao Q. Artificial intelligence in lung cancer diagnosis and prognosis: current application and future perspective. Semin Cancer Biol. (2023) 89:30–7. doi: 10.1016/j.semcancer.2023.01.006
14. Pulumati A, Pulumati A, Dwarakanath BS, Verma A, Papineni RV. Technological advancements in cancer diagnostics: improvements and limitations. Cancer Rep. (2023) 6:e1764. doi: 10.1002/cnr2.1764
15. King AD, Thoeny HC. Functional MRI for the prediction of treatment response in head and neck squamous cell carcinoma: potential and limitations. Cancer Imaging. (2016) 16:1–8. doi: 10.1186/s40644-016-0080-6
16. Faghihi R, Zeinali-Rafsanjani B, Mosleh-Shirazi MA, Saeedi-Moghadam M, Lotfi M, Jalli R, et al. Magnetic resonance spectroscopy and its clinical applications: a review. J Med Imaging Radiat Sci. (2017) 48:233–53. doi: 10.1016/j.jmir.2017.06.004
17. Karim S, Qadir A, Farooq U, Shakir M, Laghari AA. Hyperspectral imaging: a review and trends towards medical imaging. Curr Med Imaging. (2023) 19:417–27. doi: 10.2174/1573405618666220519144358
18. Tobore I, Li J, Yuhang L, Al-Handarish Y, Kandwal A, Nie Z, et al. Deep learning intervention for health care challenges: some biomedical domain considerations. JMIR mHealth uHealth. (2019) 7:e11966. doi: 10.2196/11966
19. Gull S, Akbar S. Artificial intelligence in brain tumor detection through MRI scans: advancements and challenges. In:Mohan Goyal L, Saba T, Rehman A, Larabi-Marie-Sainte S, , editors. Artificial Intelligence and Internet of Things. Boca Raton, FL: CRC Press (2021), p. 241–76. doi: 10.1201/9781003097204-10
20. Demehri S, Baffour FI, Klein JG, Ghotbi E, Ibad HA, Moradi K, et al. Musculoskeletal CT imaging: state-of-the-art advancements and future directions. Radiology. (2023) 308:e230344. doi: 10.1148/radiol.230344
21. Islam MZ, Naqvi RA, Haider A, Kim HS. Deep learning for automatic tumor lesions delineation and prognostic assessment in multi-modality PET/CT: a prospective survey. Eng Appl Artif Intell. (2023) 123:106276. doi: 10.1016/j.engappai.2023.106276
22. Pasquini L, Peck KK, Jenabi M, Holodny A. Functional MRI in neuro-oncology: state of the art and future directions. Radiology. (2023) 308:e222028. doi: 10.1148/radiol.222028
23. Seelen LW, van den Wildenberg L, van der Kemp WJ, Mohamed Hoesein FA, Mohammad NH, Molenaar IQ, et al. Prospective of 31P MR spectroscopy in hepatopancreatobiliary cancer: a systematic review of the literature. J Magn Reson Imaging. (2023) 57:1144–55. doi: 10.1002/jmri.28372
24. Pecorino L. Molecular Biology of Cancer: Mechanisms, Targets, and Therapeutics. Oxford: Oxford University Press. (2021). doi: 10.1093/hesc/9780198833024.001.0001
25. Pérez-González A, Bévant K, Blanpain C. Cancer cell plasticity during tumor progression, metastasis and response to therapy. Nat Cancer. (2023) 4:1063–82. doi: 10.1038/s43018-023-00595-y
26. Carbone A. Cancer classification at the crossroads. Cancers. (2020) 12:980. doi: 10.3390/cancers12040980
27. Choi JH, Ro JY. The 2020 WHO classification of tumors of bone: an updated review. Adv Anat Pathol. (2021) 28:119–38. doi: 10.1097/PAP.0000000000000293
28. Gress DM, Edge SB, Greene FL, Washington MK, Asare EA, Brierley JD, et al. Principles of cancer staging. In:Amin MB, Edge S, Greene F, Byrd DR, Brook-land RK, Washington MK, et al., , editors. AJCC Cancer Staging Manual, 8th edition. New York, NY: Springer (2017), p. 3–30.
29. Thierheimer M, Cioffi G, Waite KA, Kruchko C, Ostrom QT, Barnholtz-Sloan JS. Mortality trends in primary malignant brain and central nervous system tumors vary by histopathology, age, race, and sex. J Neurooncol. (2023) 162:167–77. doi: 10.1007/s11060-023-04279-6
30. Tariq MR, Ali SW, Fatima N, Jabeen A, Qazi AS, Hameed A, et al. Radiation therapies in cancer. In:Saleem Qazi A, Tariq K, , editors. Therapeutic Approaches in Cancer Treatment. New York, NY: Springer (2023), 59–77. doi: 10.1007/978-3-031-27156-4_4
31. Triulzi FM, editor. Gliomas, glioneuronal tumors and neuronal tumors: ependymaltumors. choroid plexus tumors. In: Neuroradiology of Brain Tumors: Practical Guide based on the 5th Edition of WHO Classification. New York, NY: Springer; 2023. p. 125-137. doi: 10.1007/978-3-031-38153-9_5
32. Gillespie C, Pettorini B. Central nervous system tumors. In:Losty PD, La Quaglia M, Sarnacki S, Fuchs J, Taguchi T, , editors. Pediatric Surgical Oncology. Boca Raton, FL: CRC Press (2022), p. 272–312. doi: 10.1201/9781351166126-28
33. Saboor A, Li JP, Ul Haq A, Haq A, Rehman SU. Lung cancer accurate detection employing machine learning stacking approach and relief algorithm for feature selection using clinical data. In: 2023 20th International Computer Conference on Wavelet Active Media Technology and Information Processing (ICCWAMTIP). Chengdu (2023), p. 1–10. doi: 10.1109/ICCWAMTIP60502.2023.10387125
34. Schuller HM. The impact of smoking and the influence of other factors on lung cancer. Expert Rev Respir Med. (2019) 13:761–9. doi: 10.1080/17476348.2019.1645010
35. Siringo M, Baena J, Bote de, Cabo H, Torres-Jiménez J, Zurera M, Zugazagoitia J, et al. Future perspectives in the second line therapeutic setting for non-oncogene addicted non-small cell lung cancer. Cancers. (2023) 15:5505. doi: 10.3390/cancers15235505
36. Yeong J, Thike AA, Tan PH, Iqbal J. Identifying progression predictors of breast ductal carcinoma in situ. J Clin Pathol. (2017) 70:102–8. doi: 10.1136/jclinpath-2016-204154
37. Agbley BLY, Li JP, Haq AU, Bankas EK, Mawuli CB, Ahmad S, et al. Federated fusion of magnified histopathological images for breast tumor classification in the internet of medical things. IEEE J Biomed Health Inform. (2023) 1–12. doi: 10.1109/JBHI.2023.3256974
38. Solomon DA, Pekmezci M. Pathology of meningiomas. In:McDermott MW, , editor. Handbook of Clinical Neurology, Vol. 169. Amsterdam: Elsevier (2020), p. 87–99. doi: 10.1016/B978-0-12-804280-9.00005-6
39. Zhang Y, Wu X, He L, Meng C, Du S, Bao J, et al. Applications of hyperspectral imaging in the detection and diagnosis of solid tumors. Transl Cancer Res. (2020) 9:1265. doi: 10.21037/tcr.2019.12.53
40. Fabelo H, Ortega S, Szolna A, Bulters D, Piñeiro JF, Kabwama S, et al. In-vivo hyperspectral human brain image database for brain cancer detection. IEEE Access. (2019) 7:39098–116. doi: 10.1109/ACCESS.2019.2904788
41. Mou L, Ghamisi P, Zhu XX. Deep recurrent neural networks for hyperspectral image classification. IEEE Trans Geosci Remote Sens. (2017) 55:3639–55. doi: 10.1109/TGRS.2016.2636241
42. Cao X, Yao J, Xu Z, Meng D. Hyperspectral image classification with convolutional neural network and active learning. IEEE Trans Geosci Remote Sens. (2020) 58:4604–16. doi: 10.1109/TGRS.2020.2964627
43. Verdhan V, Verdhan V. Image classification using LeNet. In: Computer vision Using Deep learning: Neural Network Architectures with Python and Keras. Berkeley, CA: Apress (2021), p. 67–101. doi: 10.1007/978-1-4842-6616-8_3
Keywords: hyperspectral imaging classification, Sharpened Cosine Similarity, deep learning, tumor tissues, Hybrid Health Care
Citation: Butt MHF, Li JP, Ji J, Riaz W, Anwar N, Butt FF, Ahmad M, Saboor A, Ali A and Uddin MY (2024) Intelligent tumor tissue classification for Hybrid Health Care Units. Front. Med. 11:1385524. doi: 10.3389/fmed.2024.1385524
Received: 13 February 2024; Accepted: 03 April 2024;
Published: 26 June 2024.
Edited by:
Hikmat A. M. Abdeljaber, Applied Science Private University, JordanReviewed by:
Aminreza Karamoozian, Shaoxing University, ChinaSaeed Ahmed, Lund University, Sweden
Hira Munir, Coventry University, United Kingdom
Copyright © 2024 Butt, Li, Ji, Riaz, Anwar, Butt, Ahmad, Saboor, Ali and Uddin. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Jiancheng (Charles) Ji, jcji20@szpu.edu.cn