- 1School of Astronautics, Beihang University, Beijing, China
- 2Aerospace Information Research Institute, Chinese Academy of Sciences, Beijing, China
- 3Center of Remote Sensing, Nanjing Tech University, Nanjing, China
Hyperspectral image classification in remote sensing often encounters challenges due to limited annotated data. Semi-supervised learning methods present a promising solution. However, their performance is heavily influenced by the quality of pseudo labels. This limitation is particularly pronounced during the early stages of training, when the model lacks adequate prior knowledge. In this paper, we propose an Iterative Pseudo Label Generation (IPG) framework based on the Segment Anything Model (SAM) to harness structural prior information for semi-supervised hyperspectral image classification. We begin by using a small number of annotated labels as SAM point prompts to generate initial segmentation masks. Next, we introduce a spectral voting strategy that aggregates segmentation masks from multiple spectral bands into a unified mask. To ensure the reliability of pseudo labels, we design a spatial-information-consistency-driven loss function that optimizes IPG to adaptively select the most dependable pseudo labels from the unified mask. These selected pseudo labels serve as iterative point prompts for SAM. Following a suitable number of iterations, the resultant pseudo labels can be employed to enrich the training data for the classification model. Experiments conducted on the Indian Pines and Pavia University datasets demonstrate that even a simple 2D CNN based classification model trained with our generated pseudo labels significantly outperforms eight state-of-the-art hyperspectral image classification methods.
1 Introduction
With the continuous advancement of remote sensing technology, hyperspectral imagery has garnered increasing attention due to its rich spectral information. For instance, Mérida-García et al. (2024) analyzed wheat gene dissection using hyperspectral images, and Nagy et al. (2024) employed machine learning algorithms to estimate maize chlorophyll content. Hyperspectral image classification presents critical challenges and remains a fundamental component for the effective application of hyperspectral technology. Despite significant progress in deep learning, hyperspectral image classification techniques often rely heavily on extensive pixel-level annotations, which are both time-consuming and labor-intensive to obtain. To address these challenges, recent studies have proposed various methods to enhance the quality of hyperspectral data. For example, Han et al. (2024) addressed the issue of incomplete spectral coverage by utilizing spectral libraries to improve spectral resolution, effectively enriching spectral information from low-resolution or incomplete data. Additionally, Wu et al. (2024) introduced a novel network for super-resolution tasks, focusing on multi-scale background feature enhancement, enabling the effective recovery of high-resolution remote sensing images from low-resolution inputs. These advances have significantly improved the accuracy and reliability of hyperspectral image classification. However, they still fail to address the inherent challenge of limited labeled data. To tackle this issue, semi-supervised learning has emerged as a powerful approach in hyperspectral image classification, leveraging the combination of a small set of labeled data and a large amount of unlabeled data to optimize model’s performance.
Current semi-supervised learning methods for hyperspectral image classification typically generate pseudo labels from unlabeled pixels during training, which are then integrated into the classification network as additional training samples. A prevalent approach is the self-training scheme (Marconcini et al., 2009), which generates highly confident predictions to augment the available labeled data. Furthermore, Zhang et al. (2014) adapted a co-training process that incorporates both original spectral signatures and 2D Gabor features for classification in scenarios with extremely limited labeled samples. Haut et al. (2018) introduced a Bayesian CNN framework assisted by active learning for semi-supervised hyperspectral image classification, which iteratively strengthens the small set of labeled samples by selecting and annotating the most informative unlabeled data. Although semi-supervised learning approaches can increase the classification accuracy through generating pseudo labels, the credibility of these labels is often compromised due to the lack of prior knowledge in the early stages, leading to noise that negatively degrades the classification network’s performance.
In recent years, large language models have made remarkable advancements in natural language processing. Notably, models like GPT-3 (Patel et al., 2023), with billions of parameters, have demonstrated impressive capabilities in zero/few-shot learning. In the realm of computer vision, pre-trained visionlanguage models such as LLaVa (Liu et al., 2024) have exhibited exceptional zero-shot generalization performance across various visual tasks. Additionally, the Segment Anything Model (SAM) (Kirillov et al., 2023) showcases the ability to perform category-agnostic segmentation by utilizing visual cues such as boxes, points, or masks. Recent research has focused on distillation (Julka and Granitzer, 2023) techniques tailored for specific task scenarios and developing adapters (Huang et al., 2024), underscoring SAM’s adaptability and potential for customization. However, due to the spectral band specificity inherent in hyperspectral images, these large vision models often struggle to achieve satisfactory results when directly applied to hyperspectral image classification.
In this paper, we propose an Iterative Pseudo Label Generation (IPG) framework based on SAM to produce high-confidence pseudo labels. The iterative refinement of mask predictions for a given category is illustrated in Figure 1, demonstrating a notable increase in the confidence of predicted labels as input constraints become more accurate. Specifically, the spectral bands are divided into multiple groups, with every three adjacent bands forming a group, which are then input into the SAM model. A small number of annotated labels are used as point prompts for SAM to generate initial segmentation masks. To enhance accuracy, we introduce a spectral voting strategy to merge segmentation masks generated from multiple groups of spectral bands into a unified mask. Furthermore, a spatial-information-consistency-driven loss function is employed to optimize the IPG framework, enabling the dynamic generation of reliable pseudo labels from the unified mask. These pseudo labels iteratively serve as point prompts for SAM, with the final pseudo labels utilized for hyperspectral image classification.
Figure 1. Illustration of the iterative generation process of pseudo labels by using point prompts to fine-tune SAM. (A) Hyperspectral image overlaid with the ground-truth label for a single object category, (B) Ground-truth labels, (C) Initial mask generated by SAM, (D) Mask generated by SAM after sufficient rounds of optimization with iteratively updated point prompts. A comparison between (C, D) demonstrates that refining the pseudo labels significantly enhances SAM’s prediction accuracy.
The main contributions of this study are outlined as follows:
● We propose a SAM-based Iterative Pseudo Label Generation (IPG) framework for semi-supervised hyperspectral image classification, utilizing the structural prior knowledge of large models to generate reliable pseudo labels.
● To align the spectral channels of hyperspectral images with the input requirements of the SAM model, we divide the spectral bands into multiple groups, each comprising three adjacent bands. A spectral voting strategy is then employed to merge the segmentation masks generated from these groups into a unified representation, facilitating precise pixel-level classification.
● To enhance the reliability of pseudo labels derived from SAM segmentation masks, we develop a spatial-information-consistency-driven loss function. This function minimizes the feature distance between the generated pseudo labels and annotated labels in the spatial dimension, ensuring higher consistency and accuracy.
2 Proposed method
The proposed Iterative Pseudo Label Generation (IPG) framework leverages the Segment Anything Model (SAM) to iteratively refine pseudo labels for hyperspectral image classification. The process begins by decomposing hyperspectral images into a series of three-channel spectral bands. A limited set of annotated labels is then used to create initial point prompts for SAM, which generates segmentation masks for these spectral bands. These masks are aggregated using a spectral voting strategy to improve the pseudo labels’ reliability. Subsequently, a spatial-information-consistency-driven loss function is applied to identify high-confidence pseudo labels. These refined pseudo labels are iteratively fed back as new point prompts, enabling SAM to produce increasingly accurate labels with each iteration.
For clarity, an overview of the IPG framework is presented in Figure 2. In the context of hyperspectral image classification, the classification model’s performance can be significantly enhanced through data augmentation using the pseudo labels generated by our IPG framework. Specifically, we concatenate these pseudo labels with the hyperspectral images, effectively expanding the training dataset and enriching the supervision information available for model training.
Figure 2. Illustration of the structure of the IPG framework. It introduces spectral voting strategy and spatial-information-consistency-driven loss function, freezing the SAM mask encoder and decoder, and updating prompt encoder, iteratively optimizing the generated pseudo labels.
The following subsections provide an in-depth exploration of the IPG framework’s three core components (i.e., pseudo label generation, spectral voting, and consistency-driven loss function), and describe its application in hyperspectral image classification.
2.1 SAM-based iterative pseudo label generation
At the beginning, the input hyperspectral image is decomposed into a series of three-channel images along the channel dimension, enabling the integration of this hyperspectral image into the SAM image encoder module for segmentation purposes. To be specific, denote the input image by , which is decomposed along the channel dimension, establishing a sequence of three-channel images . That is, every three adjacent spectral channels in M compose a three-channel image, and thus we will obtain (C − 2) images.
Afterwards, each three-channel image in the sequence is matched with labels for each category in the training set. Then, the matched labels are initialized as the point prompts of SAM. Following a category-wise processing manner, the current category label serves as the foreground, while the remaining category labels are considered as background for predicting the mask of the current category. Setting a confidence threshold, we designate labels with confidence scores above the threshold as foreground and continue to designate the remaining labels as background iteratively optimizing the generated labels. Taking point a as an example, the threshold is set as follows:
where represents the predicted label for in category , and denotes the confidence score of the prediction. The parameters and are the confidence thresholds. According to the equation, a value of indicates the foreground, while denotes the background.
2.2 Spectral voting
When splitting the hyperspectral image to along the spectral dimension and cyclically predicting masks, category confusion may arise. To better leverage spectral information and mitigate this issue, we propose a spectral voting strategy. Assuming there are K categories of objects in the hyperspectral image, the prediction head of SAM generates an output tensor comprising (C − 2) × K masks. For each pixel a and category i, let N (a,i) denote the number of binary classification predictions corresponding to category i.
Specifically, if the prediction for the jth image in the current category is positive, N (a,i) increases by 1. Otherwise, it remains 0. Subsequently, compute the total number of predictions for that pixel across all categories T(a) by
After k rounds of processing, choose the category with the highest total prediction count as the final category for that pixel:
where means the final category for pixel a. By employing this process, each pixel undergoes (C − 2) predictions for every category and is ultimately assigned to the category with the highest total prediction count.
2.3 Spatial-information-consistency-driven loss function
To optimize the point prompts, we introduce a spatial-information-consistency-driven loss function. Specifically, in the context of the label generation process for the ith category, the feature set of the point prompts is represented as Fi = {fi1,…,fij,…,fin}. The feature of a sample a is denoted by fa, and the similarity Saj between a and a reference prompt point can be calculated as:
where Saj represents the similarity between fa and fij, with fij ∈ Fi being the feature corresponding to the ith category label. The confidence of sample a is calculated as the average of its similarity with all labels, and is expressed as:
During the training phase of the Prompt Encoder in SAM, we freeze most of the parameters of the feature decoder and update only the parameters of the first-layer decoder. A spatial-information-consistency-driven loss function is constructed by calculating the feature similarity between the samples of pseudo labels and those of ground-truth labels. This loss is then used to update the Prompt Encoder through backpropagation. In each iteration, IPG selects high-confidence pseudo labels as positive samples to optimize the generation of segmentation maps. The incoming “point prompts” are based on the most recently selected highconfidence pseudo labels, which are treated as new positive samples and simultaneously serve as prompts to SAM, ensuring continuous optimization and updating of the segmentation masks.
Specifically, the loss function for training the Prompt Encoder is designed to minimize the distance between samples with generated labels and those with labels from the same category, while maximizing the separation from other categories. To achieve this purpose, we employ a binary classification loss function to optimize the pseudo label generation framework:
where indicates whether pseudo labels are used for training the network, and can be calculated as:
where quantifies the uncertainty of the pseudo labels and can be calculated using Equation 8. The parameter represents the threshold for the uncertainty of positive samples. When the uncertainty falls below this threshold, indicating that the pseudo label is considered reliable, we set , thereby allowing the label to be used in subsequent classification task.
2.4 IPG-based hyperspectral image classification
In hyperspectral image classification, the IPG framework plays a crucial role in enhancing the training process. After a sufficient number of iterations, the high-confidence pseudo labels generated by IPG are combined with the limited labeled samples, thereby enriching the training data for the 2D Convolutional Neural Network (2DCNN)-based classification model, as shown in Figure 3. This approach effectively utilizes both the scarce labeled data and the pseudo labels to enable more efficient feature learning. The learned features are then passed through the classification block of the model, which generates the final segmentation map by predicting category labels for each pixel in the hyperspectral image.
Figure 3. Illustration of the application of our IPG framework in an hyperspectral image classification network. Specifically, the pseudo labels generated by IPG are combined with the ground-truth labels to train the classification network, thereby improving its performance.
3 Experimental results and discussion
3.1 Experimental settings
To evaluate the efficacy of the proposed framework, we conducted experiments on two publicly available datasets: the Indian Pines dataset (Zare and Gader, 2008) and the Pavia University dataset (Huang and Zhang, 2009). During the training sample generation process, we employed a random sampling strategy. This approach ensures a diverse representation of samples, contributing to the robustness and generalizability of the model.
We compared our proposed method with several representative classification approaches, including Support Vector Machine (SVM) (Melgani and Bruzzone, 2004), Contextual Deep CNN (CDCNN) (Lee and Kwon, 2017), Spectral-Spatial Residual Network (SSRN) (Zhong et al., 2018), Double-Branch MultiAttention Mechanism Network (DBMA) (Ma et al., 2019), Adaptive Spectral-Spatial Kernel ResNet (A2S2K) (Roy et al., 2020), Discriminative Co-alignment (DCA) (Zhang et al., 2021), Dual-layer Deep Spatial Manifold Representation (SMR-EG) (Wang et al., 2022), and the Semi-Supervised Long-Tailed Learning Framework with Spatial Neighborhood Information (SLN-SNI) (Feng et al., 2023).
To ensure a comprehensive and fair comparison among the methods, we employed three additional metrics in addition to single-category classification accuracy: overall accuracy (OA) (Li et al., 2021), average accuracy (AA) (Li et al., 2022), and Kappa coefficient (Kappa) (Li et al., 2023, 2024). These metrics provide a more nuanced understanding of model performance by capturing different aspects of classification effectiveness. The calculations for these three metrics are detailed below:
where N is the number of categories, TPi represents the number of true positives for the ith category, and Totali refers to the total number of samples in the ith category, including both correctly and incorrectly classified samples. Additionally, Pe is calculated as follows:
where FPi represents the number of false positives for the ith category.
3.2 Implementation details
In our proposed method, we initialized an equal number of training samples, each of size 9 × 9, with five samples per category. Using SAM, we actively selected one sample per category per epoch for iterative optimization in each subsequent round. In the IPG model, we set the parameters as follows: κh = 0.2, τh = 0.8, and τl = 0.5. Given that our framework operates iteratively, and based on the ablation results in Section 3.5, we set the number of iterations to 50 to balance efficiency and effectiveness. After several rounds of iterative optimization, the generated pseudo samples were combined with ground-truth samples to train the 2DCNN, aiming for high precision. The network was trained using an SGD optimizer with the following default parameters: learning rate = 0.05, momentum = 0.7, and weight decay = 0.0001. All implementations were based on the PyTorch backend and executed on a desktop equipped with a single NVIDIA A100 GPU.
Throughout the experiments, we ensured consistency with the control method in terms of training samples, test samples, and learning rate, thus establishing a fair comparison between all models evaluated.
3.3 Qualitative evaluation results
Figure 4 displays the classification map generated by various methods on the Indian Pines dataset, providing a comprehensive assessment of model performance. It is clear that our method significantly outperforms competing algorithms in terms of visual quality, showcasing superior integrity in overall segmentation. The classification map produced by our proposed method exhibits sharper boundaries, reduced noise, and more distinct category separations, thereby enhancing the interpretability and reliability of hyperspectral image classification.
Figure 4. Visual comparison of classification results on the Indian Pines dataset. (A) Ground-truth map, (B) SVM, (C) CDCNN, (D) SSRN, (E) DBMA, (F) A2S2K, (G) DCA, (H) SMR-EG, (I) SLN-SNI, and (J) Ours.
In contrast, Figure 5 presents the classification results of the comparison methods on the Pavia University dataset. Especially in regions corresponding to categories such as Asphalt and Trees, the classification results are clear with well-defined boundaries, and there are fewer classification errors. In contrast, traditional methods like SVM and CDCNN exhibit significant noise in certain land cover areas, particularly in complex regions such as Buildings, where the classification results are more disordered. In comparison, our method not only effectively reduces noise but also demonstrates more accurate boundary recognition.
Figure 5. Visual comparison of classification results on the Pavia University dataset. (A) Ground-truth map, (B) SVM, (C) CDCNN, (D) SSRN, (E) DBMA, (F) A2S2K, (G) DCA, (H) SMR-EG, (I) SLN-SNI, and (J) Ours.
3.4 Quantitative evaluation results
Tables 1 and 2 summarize the overall accuracy (OA), average accuracy (AA), and Kappa coefficient (Kappa) values for various methods evaluated on two real datasets. The results indicate that our proposed method achieves substantial improvements over comparative techniques across all evaluation metrics, demonstrating its robustness and effectiveness. The detailed results and discussions are presented below.
In Table 1, our method attains 100% accuracy in predicting five categories, with only three categories showing accuracy below 90%. This highlights the model’s robustness in handling diverse categories. However, performance of our method in the third and fifth categories is comparatively lower, likely due to the similar spectral signatures of these categories, which complicates their distinction. The semi-supervised nature of our model, though powerful, may struggle to fully capture the nuances of these categories, especially as the iterative pseudo-labeling process can introduce noisy labels that degrade subsequent classification accuracy.
Further supporting this, Table 2 demonstrates that our approach significantly outperforms previous models in two specific categories, underscoring the framework’s capability to enhance classification accuracy in hyperspectral image analysis. These quantitative findings validate the effectiveness of our method in advancing hyperspectral classification performance across various challenging scenarios.
3.5 Ablation study
3.5.1 Ablation of different module combinations
Table 3 presents the contributions of various modules within our proposed method to classification accuracy. In this table, “SAM” denotes the framework based on the Segment Anything Model, while “SCC” refers to the spatial-information-consistency-driven loss function, and “SV” indicates the spectral voting strategy. If Spectral Voting strategy absent, it indicates that hyperspectral images are reduced to three-channel images for processing. The results demonstrate that incorporating the IPG framework significantly improves classification accuracy. This improvement can be attributed to the use of unsupervised large models in the semi-supervised generation of pseudo labels. Additionally, the spatial-informationconsistency-driven loss function and the spectral voting strategy further enhance classification accuracy within the IPG framework.
3.5.2 Ablation for different iteration numbers in spatial-information-consistency-driven loss function
To evaluate the effect of iteration numbers in the spatial-information-consistency-driven loss function on classification performance, we conducted experiments using the Pavia University dataset. As shown in Table 4, increasing the number of iterations initially results in a consistent improvement in classification accuracy, with only a slight increase in running time. However, when the iteration number surpasses 50, accuracy begins to decline, and the running time increases sharply. This can be attributed to the increasing number of pseudo labels fed into SAM in each iteration, leading to increased inference time. Meanwhile, as the number of iterations grows, the consistency-constrained pseudo labels tend to converge too closely to the labeled data, reducing the inclusion of useful spatial information and ultimately causing a drop in accuracy. These ablation results demonstrate the importance of selecting an optimal number of iterations to balance classification accuracy and computational efficiency.
Table 4. Ablation study of different iteration numbers in spatialinformation-consistency-driven loss function on the Pavia University dataset.
4 Conclusion
This paper introduces an iterative pseudo label generation (IPG) framework for hyperspectral image classification. The proposed approach integrates the Segment Anything Model (SAM) with a spectral voting strategy, effectively leveraging the rich spectral information in hyperspectral images for label estimation. Experimental results confirm that the IPG framework significantly improves classification performance, even with limited annotations. Despite its promising results, this study has some limitations. First, the performance of the IPG framework can be affected by the method used to group spectral bands, which may result in variability in outcomes across different datasets. Second, the model’s dependence on the quality of initial annotated labels may restrict its effectiveness in scenarios with insufficient or low-quality labeled data. In future work, we will further enhance the proposed method and evaluate its performance on a broader range of datasets to reinforce its robustness and demonstrate its generalizability.
Data availability statement
The datasets used in this study is available at: https://www.ehu.eus/ccwintco/index.php/Hyperspectral_Remote_Sensing_Scenes.
Author contributions
ZZ: Conceptualization, Formal analysis, Methodology, Software, Writing – original draft. GZhou: Conceptualization, Formal analysis, Writing – original draft. QW: Conceptualization, Formal analysis, Writing – original draft. JF: Conceptualization, Formal analysis, Writing – original draft. HJ: Conceptualization, Formal analysis, Writing – original draft. GZhang: Writing – review & editing. YZ: Funding acquisition, Supervision, Writing – review & editing.
Funding
The author(s) declare financial support was received for the research, authorship, and/or publication of this article. This work was supported in part by the National Key Research and Development Program of China under grant 2022ZD0117400, in part by the National Natural Science Foundation of China under Grant 62132002, and in part by the Fundamental Research Funds for the Central Universities under Grants 501QYJC2024115010 and 502GWXM2024115007.
Acknowledgments
We would like to express our sincere gratitude to the editors and reviewers for their constructive comments and suggestions, which greatly contributed to improving the quality of this paper.
Conflict of interest
The authors declare that the research was conducted in the absence of any commercial or financial relationships that could be construed as a potential conflict of interest.
Generative AI statement
The author(s) declare that no Generative AI was used in the creation of this manuscript.
Publisher’s note
All claims expressed in this article are solely those of the authors and do not necessarily represent those of their affiliated organizations, or those of the publisher, the editors and the reviewers. Any product that may be evaluated in this article, or claim that may be made by its manufacturer, is not guaranteed or endorsed by the publisher.
References
Feng, Y., Song, R., Ni, W., Zhu, J., Wang, X. (2023). A novel semi-supervised long-tailed learning framework with spatial neighborhood information for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 20, 1–5. doi: 10.1109/LGRS.2023.3241340
Han, X., Leng, W., Zhang, H., Wang, W., Xu, Q., Sun, W. (2024). Spectral library based spectral super-resolution under incomplete spectral coverage conditions. IEEE Trans. Geosci. Remote Sens. 62, 5516312. doi: 10.1109/TGRS.2024.3392606
Haut, J. M., Paoletti, M. E., Plaza, J., Li, J., Plaza, A. (2018). Active learning with convolutional neural networks for hyperspectral image classification using a new bayesian approach. IEEE Trans. Geosci. Remote Sens. 56, 6440–6461. doi: 10.1109/TGRS.2018.2838665
Huang, X., Zhang, L. (2009). A comparative study of spatial approaches for urban mapping using hyperspectral rosis images over pavia city, northern Italy. Int. J. Remote Sens. 30, 3205–3221. doi: 10.1080/01431160802559046
Huang, Y., Yang, X., Liu, L., Zhou, H., Chang, A., Zhou, X., et al. (2024). Segment anything model for medical images. Med. Image Anal. 92, 103061. doi: 10.1016/j.media.2023.103061
Julka, S., Granitzer, M. (2023). “Knowledge distillation with segment anything (SAM) model for planetary geological mapping,” in Machine Learning, Optimization, and Data Science. (Switzerland: Springer Nature), 68–77.
Kirillov, A., Mintun, E., Ravi, N., Mao, H., Rolland, C., Gustafson, L., et al. (2023). “Segment anything,” in 2023 IEEE/CVF International Conference on Computer Vision (ICCV),Paris, France, October 1-6, 2023. (IEEE), 3992–4003.
Lee, H., Kwon, H. (2017). Going deeper with contextual CNN for hyperspectral image classification. IEEE Trans. Image Process. 26, 4843–4855. doi: 10.1109/TIP.2017.2725580
Li, L., Ma, H., Jia, Z. (2021). Change detection from sar images based on convolutional neural networks guided by saliency enhancement. Remote Sens. 13, 3697. doi: 10.3390/rs13183697
Li, L., Ma, H., Jia, Z. (2022). Multiscale geometric analysis fusion-based unsupervised change detection in remote sensing images via flicm model. Entropy 24, 291. doi: 10.3390/e24020291
Li, L., Ma, H., Jia, Z. (2023). Gamma correction-based automatic unsupervised change detection in sar images via flicm model. J. Indian Soc. Remote Sens. 51, 1077–1088. doi: 10.1007/s12524-023-01674-4
Li, L., Ma, H., Zhang, X., Zhao, X., Lv, M., Jia, Z. (2024). Synthetic aperture radar image change detection based on principal component analysis and two-level clustering. Remote Sens. 16, 1861. doi: 10.3390/rs16111861
Liu, H., Li, C., Wu, Q., Lee, Y. J. (2024). Visual instruction tuning. Adv. Neural Inf. Process. Syst. 36, 34892–34916.
Ma, W., Yang, Q., Wu, Y., Zhao, W., Zhang, X. (2019). Double-branch multi-attention mechanism network for hyperspectral image classification. Remote Sens. 11, 1307. doi: 10.3390/rs11111307
Marconcini, M., Camps-Valls, G., Bruzzone, L. (2009). A composite semisupervised SVM for classification of hyperspectral images. IEEE Geosci. Remote Sens. Lett. 6, 234–238. doi: 10.1109/LGRS.2008.2009324
Melgani, F., Bruzzone, L. (2004). Classification of hyperspectral remote sensing images with support vector machines. IEEE Trans. Geosci. Remote Sens. 42, 1778–1790. doi: 10.1109/TGRS.2004.831865
Mérida-García, R., Gálvez Rojas, S., Solís, I., Martinez, F., Camino, C., Soriano, J. M., et al. (2024). High-throughput phenotyping using hyperspectral indicators supports the genetic dissection of yield in durum wheat grown under heat and drought stress. Front. Plant Sci. 15, 1470520. doi: 10.3389/fpls.2024.1470520
Nagy, A., Szabó, A., Elbeltagi, A., Nxumalo, G. S., Bódi, E. B., Tamás, J. (2024). Hyperspectral indices data fusion-based machine learning enhanced by mrmr algorithm for estimating maize chlorophyll content. Front. Plant Sci. 15, 1419316. doi: 10.3389/fpls.2024.1419316
Patel, A., Li, B., Rasooli, M. S., Constant, N., Raffel, C., Callison-Burch, C. (2023). “Bidirectional language models are also few-shot learners,” in The Eleventh International Conference on Learning Representations, ICLR 2023, Kigali, Rwanda, May 1-5, 2023 (OpenReview.net).
Roy, S. K., Manna, S., Song, T., Bruzzone, L. (2020). Attention-based adaptive spectral–spatial kernel resnet for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 59, 7831–7843. doi: 10.1109/TGRS.2020.3043267
Wang, C., Zhang, L., Wei, W., Zhang, Y. (2022). Toward effective hyperspectral image classification using dual-level deep spatial manifold representation. IEEE Trans. Geosci. Remote Sens. 60, 1–14. doi: 10.1109/TGRS.2021.3073932
Wu, T., Zhao, R., Lv, M., Jia, Z., Li, L., Wang, Z., et al. (2024). Lightweight remote sensing image super-resolution via background-based multi-scale feature enhancement network. IEEE Geosci. Remote Sens. Lett. 21, 7509405. doi: 10.1109/LGRS.2024.3481645
Zare, A., Gader, P. (2008). Hyperspectral band selection and endmember detection using sparsity promoting priors. IEEE Geosci. Remote Sens. Lett. 5, 256–260. doi: 10.1109/LGRS.2008.915934
Zhang, Y., Li, W., Tao, R., Peng, J., Du, Q., Cai, Z. (2021). Cross-scene hyperspectral image classification with discriminative cooperative alignment. IEEE Trans. Geosci. Remote Sens. 59, 9646–9660. doi: 10.1109/TGRS.2020.3046756
Zhang, X., Song, Q., Liu, R., Wang, W., Jiao, L. (2014). Modified co-training with spectral and spatial views for semisupervised hyperspectral image classification. Remote Sens. 7, 2044–2055. doi: 10.1109/JSTARS.4609443
Keywords: hyperspectral image classification, remote sensing, semi-supervised learning, Segment Anything Model, pseudo label generation
Citation: Zhao Z, Zhou G, Wang Q, Feng J, Jiang H, Zhang G and Zhang Y (2024) An Iterative Pseudo Label Generation framework for semi-supervised hyperspectral image classification using the Segment Anything Model. Front. Plant Sci. 15:1515403. doi: 10.3389/fpls.2024.1515403
Received: 22 October 2024; Accepted: 27 November 2024;
Published: 23 December 2024.
Edited by:
Liangliang Li, Beijing Institute of Technology, ChinaCopyright © 2024 Zhao, Zhou, Wang, Feng, Jiang, Zhang and Zhang. This is an open-access article distributed under the terms of the Creative Commons Attribution License (CC BY). The use, distribution or reproduction in other forums is permitted, provided the original author(s) and the copyright owner(s) are credited and that the original publication in this journal is cited, in accordance with accepted academic practice. No use, distribution or reproduction is permitted which does not comply with these terms.
*Correspondence: Yu Zhang, dXplZnVsQDE2My5jb20=