Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
End-to-End Detail-Enhanced Dehazing Network for Remote Sensing Images
Next Article in Special Issue
Hyperspectral Image Classification on Large-Scale Agricultural Crops: The Heilongjiang Benchmark Dataset, Validation Procedure, and Baseline Results
Previous Article in Journal
The ESA Permanent Facility for Altimetry Calibration in Crete: Advanced Services and the Latest Cal/Val Results
Previous Article in Special Issue
Pan-Sharpening Network of Multi-Spectral Remote Sensing Images Using Two-Stream Attention Feature Extractor and Multi-Detail Injection (TAMINet)
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Self-Supervised Deep Multi-Level Representation Learning Fusion-Based Maximum Entropy Subspace Clustering for Hyperspectral Band Selection

Information Science and Technology College, Dalian Maritime University, Dalian 116026, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2024, 16(2), 224; https://doi.org/10.3390/rs16020224
Submission received: 4 December 2023 / Revised: 31 December 2023 / Accepted: 2 January 2024 / Published: 5 January 2024
Figure 1
<p>Overall flowchart of the proposed MLRLFMESC framework.</p> ">
Figure 2
<p>The IP dataset with (<b>a</b>) a pseudo-color map of IP, and (<b>b</b>) true image class distribution with labels.</p> ">
Figure 3
<p>The PU dataset, with (<b>a</b>) a pseudo-color map of PU, and (<b>b</b>) true image class distribution with labels.</p> ">
Figure 4
<p>The SA dataset, with (<b>a</b>) a pseudo-color map of SA, and (<b>b</b>) true image class distribution with labels.</p> ">
Figure 5
<p>Box plot of the OA for different BS methods on three hyperspectral datasets. (<b>a</b>) Indian Pines, (<b>b</b>) Pavia University, and (<b>c</b>) Salinas.</p> ">
Figure 6
<p>Classification performance with a different number of selected bands on IP, (<b>a</b>) OA, (<b>b</b>) AA, (<b>c</b>) Kappa.</p> ">
Figure 7
<p>Distribution of bands selected using various BS algorithms on IP data.</p> ">
Figure 8
<p>Classification result maps, (<b>a</b>) labeled image, and (<b>b</b>–<b>h</b>) classification result maps on IP-selected 30 bands using UBS, E-FDPC, ISSC, ASPS_MN, DSC, MLRLFMESC, and all bands, respectively.</p> ">
Figure A1
<p>Classification performance of PU dataset with a different number of selected bands, (<b>a</b>) OA, (<b>b</b>) AA, (<b>c</b>) Kappa.</p> ">
Figure A2
<p>Distribution of bands selected using various BS algorithms for PU dataset.</p> ">
Figure A3
<p>Classification result maps with (<b>a</b>) a labeled image of PU, and (<b>b</b>–<b>h</b>) classification result maps with 20 bands using UBS, E-FDPC, ISSC, ASPS_MN, DSC, MLRLFMESC, and all bands.</p> ">
Figure A4
<p>Classification performance of SA dataset with a different number of selected bands, (<b>a</b>) OA, (<b>b</b>) AA, (<b>c</b>) Kappa.</p> ">
Figure A5
<p>Distribution of bands selected using various BS algorithms for SA dataset.</p> ">
Figure A6
<p>Classification result maps with a (<b>a</b>) labeled image of SA, and (<b>b</b>–<b>h</b>) classification result maps with 30 bands using UBS, E-FDPC, ISSC, ASPS_MN, DSC, MLRLFMESC, and all bands.</p> ">
Versions Notes

Abstract

:
As one of the most important techniques for hyperspectral image dimensionality reduction, band selection has received considerable attention, whereas self-representation subspace clustering-based band selection algorithms have received quite a lot of attention with good effect. However, many of them lack the self-supervision of representations and ignore the multi-level spectral–spatial information of HSI and the connectivity of subspaces. To this end, this paper proposes a novel self-supervised multi-level representation learning fusion-based maximum entropy subspace clustering (MLRLFMESC) method for hyperspectral band selection. Firstly, to learn multi-level spectral–spatial information, self-representation subspace clustering is embedded between the encoder layers of the deep-stacked convolutional autoencoder and its corresponding decoder layers, respectively, as multiple fully connected layers to achieve multi-level representation learning (MLRL). A new auxiliary task is constructed for multi-level representation learning and multi-level self-supervised training to improve its capability of representation. Then, a fusion model is designed to fuse the multi-level spectral–spatial information to obtain a more distinctive coefficient matrix for self-expression, where the maximum entropy regularization (MER) method is employed to promote connectivity and the uniform dense distribution of band elements in each subspace. Finally, subspace clustering is conducted to obtain the final band subset. Experiments have been conducted on three hyperspectral datasets, and the corresponding results show that the proposed MLRLFMESC algorithm significantly outperforms several other band selection methods in classification performance.

1. Introduction

A hyperspectral image (HSI) is a unified image of interest targets captured by a specific wavelength of an optical sensor, consisting of hundreds of continuous bands with a fine resolution [1,2]. These bands are rich in available spectral and spatial information, allowing the identification of subtle differences between features. In recent years, due to the higher and higher spectral resolution of HSI, it has been successfully used in numerous applications, including military monitoring [3], food safety [4], medical diagnosis [5], etc. However, the strong inter-band correlation and high-dimensional hyperspectral data lead to a large amount of information redundancy, heavy computational and storage burdens, and dimensional catastrophe problems, which pose a great deal of difficulty and challenges to the development of HSI processing methods. Hyperspectral dimensionality reduction is of great significance as a form of preprocessing to address the above challenges.
The most commonly used methods for HSI dimensionality reduction to reduce spectral band redundancy include band selection (BS) [6,7] and feature extraction (FE) [8,9], where BS-based methods are based on the criteria of selecting the most representative band subset directly from the original HSI data without any transformation, and the obtained sub-bands are informative, distinguishable, and beneficial for subsequent tasks. Compared with FE, BS can reduce data dimensionality while preserving the physical meaning, inherent properties, and spectral characteristics of the original data, which is beneficial for interpreting the selected subset of bands in subsequent analysis and has been widely used in practical applications [10,11]. According to whether labeled information is used or not, the existing BS methods include three types as follows: supervised [12,13,14], semi-supervised [15,16,17], and unsupervised [18,19,20,21]. Owing to the fact that labeled HSI data are difficult to obtain, this paper focuses on unsupervised BS methods with the advantage of being more flexible, feasible, and effective in practice.
As a popular technique for unsupervised band selection, self-representation is implemented using the self-expression properties of data and different regularization constraints, where the most representative algorithms include fast and robust self-representation (FRSR) [22], robust dual graph self-representation (RDGSR) [23], and self-marginalized graph self-representation (MGSR) [24]. In addition, self-representation-based subspace clustering (SSC) has achieved a large number of successful results in unsupervised BS [25,26,27], where under the representation-based framework, the clustering structure of the spectral band can be learned in the low-dimensional subspace with the robustness of noise and outliers, and can effectively cluster high-dimensional data. However, in practical applications, the HSI data to be processed typically have great spectral variability and are located in nonlinear subspace, where traditional SSC-based BS methods with linear characteristics are not applicable for the nonlinear relationships of HSI, failing to achieve satisfactory performance.
Recently, deep neural networks have demonstrated superiority in handling high-dimensional data because of their remarkable ability to extract complex nonlinear relationships of features in an end-to-end learnable way. Deep learning models for HSI band selection have achieved tremendous success [28,29,30,31]. Although these methods are able to learn the complex nonlinear structure information of HSI, they ignore the correlation between bands, resulting in a large number of band redundancies. In view of this issue, deep learning-based clustering methods are introduced into band selection [32,33,34,35], considering the spatial information inherent in band images. However, they also have certain limitations. On the one hand, the representation learning of these models is embedded in the deep convolutional autoencoder, leading to a lack of an effective self-supervised representation ability; on the other hand, without considering the subspace clustering representation of the low-level and high-level information of the input HSI, these models ignore the meaningful multi-scale information embedded in different layers of deep convolutional autoencoders, imposing a waste of information that is conducive to clustering. In addition, due to the ignorance of the connectivity within the subspace, the existing models are blocked from further improving the clustering performance.
To solve the above problems, a self-supervised deep multi-level representation learning fusion-applying maximum entropy subspace clustering (MLRLFMESC) algorithm is proposed for BS in this paper, with the main contributions as follows:
(1)
Considering the multi-level spectral–spatial information of hyperspectral data, self-representation-based subspace clustering, comprising multiple fully connected layers, is respectively inserted between the encoder layers of the deep stacked convolutional autoencoder and its corresponding decoder layers, respectively, to realize multi-level representation learning (MLRL), which can fully extract low-level and high-level information and obtain more informative and discriminative multi-level representations.
(2)
Self-supervised information is provided to further enhance the representation capability of the MLRL, and a new auxiliary task is constructed for MLRL to perform multi-level self-supervised learning (MLSL). Furthermore, a fusion module is designed to fuse the multi-level spectral–spatial information extracted by the proposed MLRL to obtain a more informative subspace representation matrix.
(3)
To enhance the connectivity within the same subspace, the MER method is applied to ensure that the elements within the same subspace are uniformly and densely distributed, which is beneficial for subsequent spectral clustering.
The remainder of this paper is organized as follows. Section 2 gives a detailed description of the proposed MLRLFMESC algorithm. Section 3 presents the experiments and corresponding analysis of the proposed algorithm with other state-of-the-art BS methods. Finally, the conclusions and discussions are drawn in Section 4.

2. Proposed Method

This section describes in detail the proposed MLRLFMESC method for BS. The flowchart of this method is shown in Figure 1. The main steps are as follows: firstly, considering the low-level and high-level spatial information of hyperspectral data, the proposed method inserts multiple fully connected layers between the encoder layers and their corresponding decoder layers to achieve multi-level representation learning (MLRL), thus generating multiple sets of self-expression coefficient matrices at different levels of the encoder layers to obtain more informative and discriminative subspace clustering representations. Secondly, a new auxiliary task is constructed based on the MLRL, which provides multi-level self-supervised information to further enhance the representation ability of the model, termed multi-level self-supervision learning (MLSL). Finally, a fusion module is designed to integrate the multi-scale information extracted from different layers of representation learning to obtain a more differentiated self-expression coefficient matrix, where maximum entropy regularization (MER) is introduced to ensure that elements of the same subspace are evenly and densely distributed, thereby enhancing connectivity within each subspace and facilitating subsequent spectral clustering to determine the most representative band subset.

2.1. Multi-Level Representation Learning (MLRL)

The proposed MLRLFMESC method exploits a deep stacked convolutional autoencoder (SCAE) constructed by the structure of a symmetrical encoder–decoder as the core network for feature extraction to sufficiently extract the spectral–spatial information of HSI data. The original HSI data cube X = x i x i W × H i = 1 B contains W × H spatial dimensions and B spectral band dimensions. To ensure that the input of HSI samples X can be constructed by the deep SCAE, the definition of the loss of function is as follows:
L S C A E = 1 2 X X ^ F 2 ,
where the Frobenius norm is expressed as · F , and the reconstructed HSI samples are denoted as X ^ .
Self-representation-based subspace clustering assumes that all data points belong to a combination of linear or affinity subspaces, which is typically represented as a self-representative model. Let the spectral bands x i x i W × H i = 1 B come from the union of n different subspaces j = 1 n S j with the dimension d j j = 1 n in B . In order to extract spatial information, considering the nonlinear relationships of HSI, the self-representative model is embedded in the latent space of the deep SCAE to implement self-representation properties and obtain representations of subspace clustering; then, the potential clusters are recovered with spectral clustering.
Inspired by the fact that the encoder with different layers can learn more complicated feature representations of the input HSI data, a self-representation model, comprising multiple fully connected layers, is inserted between the encoder layers of the deep stacked convolutional autoencoder and its corresponding decoder layers, respectively, to achieve MLRL; thus, the multi-level spectral–spatial information can be extracted. In order to capture the shared information between encoders and generate unique information for each layer, the consistency matrix C C B × B and the discrimination matrix D l l = 1 L B × B are defined, respectively. In view of the above-mentioned consideration, MLRL can be performed by the following loss function:
L e x p = l = 1 L Z l Z l C C + D l F 2   s . t .   diag C C + D l = 0 , l 1 , , L ,
where Z l = z 1 l , z 2 l , , z B l T m × B represents the latent representation matrix and m is the dimension of the deep spatial feature.
The loss of self-expression L e x p is used to promote the learning of self-expression feature representations at different encoder levels. As for the discrimination matrix D l l = 1 L , the Frobenius norm is employed; thus, the connectivity of subspace representations related to each fully connected layer can be ensured. Meanwhile, to generate the sparse representation of the HSI, the l1-norm is used in the consistency matrix C C . Accordingly, the regular terms added to the model are shown as follows:
L C C = C C 1 ,
L D = l = 1 L D l F 2 ,
The multi-level spectral–spatial information of HSI is obtained via MLRL to facilitate the feature learning process, thereby obtaining multiple sets of information representations accordingly.

2.2. Multi-Level Self-Supervised Learning (MLSL)

Aiming at the further improvement of the representation ability of the proposed model, MLSL is used as a self-supervised method to better learn self-expression feature representations by constructing auxiliary tasks for MLRL.
To perform MLSL, the auxiliary tasks are constructed as follows: Firstly, positive and negative sample pairs are formulated for the inputs and outputs of the MLRL. For a given input Z i l and its corresponding set of outputs G j l = Z l C C j + D j l j = 1 B at layer l of MLRL, Z i l and G j l are matched as a positive pair, while Z i l and G j l j = 1 , j i B are treated as a negative pair. Subsequently, the MLSL is implemented by formulating a self-supervised loss function, expressed as follows:
L S = l = 1 L log exp Z i l T G i l / σ j = 1 B exp Z i l T G j l / σ ,
where σ is a temperature parameter that controls the distribution of the concentration level. Z i l and G i l are normalizations of Z i l and G i l , respectively.
An l2-normalization layer is used to satisfy Z i l = 1 and G i l = 1 . The classifier of B-way softmax is exploited to classify Z i l as G i l in the loss function of MLRL.

2.3. Fusion Module with Maximum Entropy Regularization (MER)

Considering that the coefficient matrices of the information representations obtained from MLRL have multiple information of the input HSI data, it is preferable to fuse these matrices into a more discriminative and informative coefficient subspace representation matrix as the input of the subspace clustering.
The matrices C C and D l l = 1 L learned using MLRL are fused through the fusion module. Stacking C C and D l l = 1 L along the channel dimension can help acquire the stacked matrix C S . Then, the channels of C S are merged using a convolutional kernel k to realize the fusion. Finally, a more informative subspace representation matrix C F is obtained via channel fusion learning, expressed as follows:
C F B × B = k C S ,
where is the convolution operation.
By using an appropriate kernel size k, C F is able to capture more local information on C C and each D l l = 1 L with the block diagonal structure.
Entropy is a measure of uncertain information contained within a random variable, where for a discrete random variable X, the entropy can be calculated as H X = y Y p x log p x with p(X) as the probability distribution function of X. The similarity between hyperspectral data samples i and j in the subspace representation matrix C F can be expressed as C F i , j , and the MER method is applied to the subspace representation matrix. According to the fact that max H C = min H C , the following loss function for C F can be obtained as follows:
L C F = i = 1 B j = 1 B C F i , j ln C F i , j   s . t .   C F i , j 0 ,
where C F i , j = 0 satisfies C F i , j ln C F i , j = 0 . The MER forces the equal strength of connections between elements from the same subspace. Simultaneously, it ensures a uniform dense distribution of elements belonging to the same subspace.

2.4. Implementation Details

The final loss function of the proposed MLRLFMESC method is expressed as follows:
L θ , θ ^ , C F = L S C A E + λ 1 L exp + λ 2 L C C + λ 3 L D + λ 4 L S + λ 5 L C F ,
where λ 1 , λ 2 , λ 3 , λ 4 , λ 5 are parameters that balance the trade-off between the earlier-mentioned different losses. θ and θ ^ are the parameters updated by the standard backpropagation of the network trained with the Adam gradient.
Once the network has been trained to obtain the matrix C F , a symmetric affinity matrix can be created for spectral clustering,
A = 1 2 C F + C F T ,
Matrix A shows the pairwise relationship between the bands. Given the above, the spectral clustering algorithms can be utilized to recover the underlying subspaces and cluster the samples into respective subspaces to obtain the clustering results. The clustering centers can be obtained via the average of the spectrum in each cluster. Then, the distances between the cluster center and each band in the same cluster can be calculated to find the closest band as the selected representative band of the cluster, and the final subset of bands can be obtained.

3. Experiments and Results

3.1. Hyperspectral Datasets

Comparative experiments are conducted on three publicly available HSIs with different scenarios to prove the effectiveness of the proposed algorithm, including the Indian Pines (IP) dataset, the Pavia University (PU) dataset, and the Salinas (SA) dataset. Detailed descriptions of the three hyperspectral datasets are given in Table 1.

3.1.1. Indian Pines (IP) Dataset

The IP dataset was captured using the AVIRIS sensor in Northwestern Indiana on 12 June 1992, containing 220 spectral bands with wavelengths of 0.4 to 2.5 μm and containing 145 × 145 pixels with a spatial resolution of 20 m. After the removal of 20 water absorption bands, 200 bands with 16 classes of crops are left for experiments. Figure 2 shows the IP dataset’s pseudo-color map as well as the true image feature class distribution.

3.1.2. Pavia University (PU) Dataset

The PU dataset was collected via the ROSIS sensor for the city of Pavia in North Italy during a flying activity in 2003, consisting of 115 bands with wavelengths of 0.43 to 0.86 μm and containing 610 × 340 pixels with a spatial resolution of 1.3 m. After the removal of 12 noise bands, there are 9 types of objects available in the remaining 103 bands (the types of objects are shown in the label of Figure 3). The pseudo-color map and real image feature class distribution of the PU dataset are shown in Figure 3.

3.1.3. Salinas (SA) Dataset

The SA dataset was gathered using the AVIRIS sensor in Salinas Valley, California, in 1998, which comprises 224 bands with wavelengths of 0.36 to 2.5 µm and contains 512 × 217 pixels with spatial resolution of 3.7 m. After deleting 20 water absorption bands, the remaining 204 bands containing 16 classes are utilized for experiments. The pseudo-color map and real image feature class distribution of the SA dataset are shown in Figure 4.

3.2. Experimental Setup

In order to prove the effectiveness of the proposed BS algorithm, five existing BS methods are used for comparison. Considering search-based, clustering-based, and ranking-based, as well as the comparison between traditional methods and deep learning methods, the selected comparable algorithms are UBS [18], E-FDPC [19], ISSC [25], ASPS_MN [5] and DSC [32], as these algorithms have open-source implementations provided by their respective authors. Performance analysis is conducted on a subset of bands obtained from various BS approaches using the same SVM classifier in [36] with an open-source code.
To quantitatively assess the quality of the selected band subsets, indicators of classification accuracy are used in this section, including overall accuracy (OA), average accuracy (AA), and the kappa coefficient (Kappa). For a fair comparative purpose, identical training and testing data subsets within each round are utilized when evaluating different BS algorithms. Specifically, 10% of samples from each class are randomly chosen as the training set, with the remaining samples allocated to the testing set. The experimental results are averaged via ten independent runs to reduce the randomness.
The deep SCAE network of the proposed method is composed of three symmetric layers of the stacked encoder and decoder with the following parameter settings: the stacked encoder consists of 10, 20, and 30 filters with corresponding kernel sizes of 5 × 5, 3 × 3, and 3 × 3, respectively. The network learning rate is 1.0 × 10 4 . The trade-off parameters are λ 1 = 1.0 × 10 1 , λ 2 = 1.0 × 10 2 , λ 3 = 1.0 × 10 2 , λ 4 = 1.0 × 10 2 , and λ 5 = 1.0 × 10 2 . The kernel size k is 3 × 3.

3.3. Randomness Validation by Random Selection of Training and Testing Sets

As for classification results, the training and testing sets have a significant influence on the classification performance. Therefore, before the comparison of various algorithms, this section gives the average and variance of OA across 10 distinct runs, allowing the assessment of disparities between experiment runs where each run involves alterations to the training and testing sets.
Figure 5a–c illustrates the box plots depicting the OA results for the six algorithms utilizing 35 bands across the three HSI datasets in 10 separate runs. The experiments entail the repeated random selection of training and testing datasets. To ensure an equitable comparison, the training and testing sets for all six algorithms within the same round remain identical. Sub-figures demonstrate that the proposed MLRLFMESC obtains the optimal mean OA with a perfect criteria bias, especially for IP and PU datasets, as shown in Figure 5a–c.

3.4. Ablation Study of the Proposed MLRLFMESC Method

To separately verify the effectiveness of maximum entropy regularization and self-supervised learning in the proposed MLRLFMESC method, ablation studies are conducted in this section. As can be shown in Equation (8) in Section 2, the proposed method adopts partial loss functions from the DSC method, namely L S C A E , L exp , L C c , and L D , collectively referred to as L D S C in this paper. Building upon this foundation, the proposed method introduces the following two innovative techniques: the multi-level self-supervised learning model (MLSL) and maximum entropy regularization (MER). In MLSL, the L S is introduced to obtain improved self-supervised features. In MER, the L C F is introduced to promote connectivity within the same subspace while ensuring a uniform and dense distribution of elements within the subspace, which is beneficial for subsequent spectral clustering. As a result, the ablation study should be implemented with/without the L S and the L C F loss in Table 2.
The following conclusions are given from Table 2: Firstly, when adding either the MER into the model (as shown in the second line in Table with the form of L C F ) or the multi-level self-supervised learning model (as shown in the third line in the Table with the form of L S ), the OA performance can be better than the DSC method. The classification performance of the proposed MLRLFMESC with both MER and MLSL (as shown in the fourth line in Table with the form of L S and L C F ) has the best OA performance with bold fonts under the selection with a different number of selected band subsets. These experiments demonstrate the effectiveness of maximum entropy regularization and self-supervised learning in the proposed MLRLFMESC approach.

3.5. Classification Results Analysis for Different BS Algorithms

In this section, comparative experiments are conducted on three publicly available HSIs with different scenarios to prove the effectiveness of the proposed algorithm, including the Indian Pines (IP) dataset, the Pavia University (PU) dataset, and the Salinas (SA) dataset. However, limited by the number of pages and considering the reproducibility of conclusions, only the results of the IP dataset are shown in this section, with the results of the PU and SA datasets in Appendix A and Appendix B.

3.5.1. BS Results with Different Number of Selected Bands

To evaluate the classification accuracy of the proposed MLRLFMESC algorithm while comparing it with the existing BS methods, the quantity n of selected bands in different BS methods is varied in the region of [5,35] with a step of five. The reason why we chose 35 as the maximum is that when using virtual dimension (VD) analysis, which is a widely used technique for selecting the number of bands, the VD is typically less than 35.
In Figure 6, the suggested MLRLFMESC technique clearly outperforms the other five BS approaches in terms of classification accuracy for OA, AA, and Kappa. In Figure 6a of the IP dataset, the proposed MLRLFMESC has the highest and most stable OA with a significant improvement over the other five comparable BS methods, especially when n = 5, 30, 35, and the proposed MLRLFMESC has better accuracies of 3.42%, 3.56%, and 2.83% compared to the suboptimal approach. The AA accuracy is shown in Figure 6b, and the MLRLFMESC technique exhibits the “Hughes” phenomena that grows and then drops as the number of bands increases. In terms of the Kappa given in Figure 6c, the MLRLFMESC approach has a growing advantage over the suboptimal approach as the number of chosen bands grows, notably when n = 30, 35, MLRLFMESC has 3.56% and 2.83% greater accuracy than the suboptimal approach, respectively. The other five BS approaches exhibit some “Hughes” phenomena in the OA, AA, and Kappa curves, indicating the necessity for band selection.

3.5.2. Classification Performance Analysis by Band Subsets Using Various BS Algorithms

To further assess the efficacy of the proposed MLRLFMESC approach for the analysis, Figure 7 shows the index distribution of the 30 bands selected from the IP dataset via various band selection approaches. It is usually regarded as a bad strategy if the selected bands are scattered across a relatively short range.
As can be seen in Figure 7, most band selection approaches pick bands that span a broad range of all bands.
Figure 8 presents the labeled image, the classification result maps for the various band selection methods, and all bands, which may be used to visually analyze the classification performance of various band selection methods on the IP dataset. There is some visual deviance, especially regarding the fact that the proposed algorithm has the best performance of the corn-min class with the orange color in the lower left corner and the better performance of the wood class with the brown color in the lower right corner. Further quantitative analysis is conducted for a fair comparison.
In order for a quantitative comparison, Table 3 displays the average classification results of 10 runs with OA, AA, and Kappa, using various band selection methods (UBS, E-FDPC, ISSC, ASPS_MN, DSC, and the proposed MLRLFMESC, respectively) and all bands, where the best classification results are blacked out and the second-best results are underlined.
It can be easily seen from Table 3 that the proposed method outperforms all previous band selection methods except all bands in terms of OA, AA, and Kappa, as well as classification accuracy in most classes. The proposed MLRLFMESC approach provides the highest classification accuracy in classes 8, 12, and 15 when compared to existing approaches.

3.6. Time Consuming for Different BS Algorithms

In this section, a comparison of computing times across different band selection algorithms is conducted to discern their respective computational complexity. Table 4 lists the average computation time for ten runs of various BS approaches on the IP dataset, where the ISSC has the lowest running time with bold font. Since MLRLFMESC and DSC use deep neural networks, the running time is longer compared to traditional BS methods. However, the runtime of MLRLFMESC is significantly less than that of DSC, efficiently obtaining the desired bands within an acceptable timeframe while achieving a superior classification performance.

4. Conclusions and Discussion

This paper presents a novel MLRLFMESC framework for unsupervised hyperspectral band selection. Self-representation subspace clustering is applied in deep SCAE to enable the learning of hyperspectral nonlinear spectra–spatial relationships in a trainable deep network.
(1)
From the results in Section 3, it can be seen that the proposed MLSL model retains good band subsets with multi-level spectral–spatial information and multi-level discriminative information representations.
(2)
A fusion module is employed to fuse the multi-level discriminative information representations, where the MER method is applied to enhance the objectiveness of the bands in each subspace while ensuring the uniform and dense distribution of bands in the same subspace, which was shown to be successful in the ablation study.
(3)
Comparable experiments indicate that the proposed MLRLFMESC approach performs better than the other five state-of-the-art BS methods on three real HSI datasets for classification performance.
Although this work, and even other existing BS methods, have been extensively studied and achieved excellent performance in classification tasks, there has been limited research on integrating BS algorithms into tasks such as hyperspectral image target detection, target tracking, unmixing, etc. An important research direction in the future is to apply band selection algorithms to different task requirements while considering high performance. Furthermore, this paper does not pay attention to the problem of the uneven distribution of samples in classification-oriented tasks, resulting in the better performance of large categories and the poor performance of small categories. Therefore, the expansion of small category samples and the design or improvement of targeted models are the direction of focus in further research.

Author Contributions

Conceptualization, Y.W.; methodology, Y.W. and H.M.; software, H.M. and Y.Y.; validation, E.Z., M.S. and C.Y.; formal analysis, Y.W.; investigation, H.M.; data curation, H.M. and Y.Y.; writing—original draft preparation, Y.W. and H.M.; writing—review and editing, Y.W. All authors have read and agreed to the published version of the manuscript.

Funding

This research was supported in part by the National Nature Science Foundation of China under Grant 42271355 and Grant 61801075, in part by the Natural Science Foundation of Liaoning Province under Grant 2022-MS-160, in part by the China Postdoctoral Science Foundation under Grant 2020M670723, and in part by the Fundamental Research Funds for the Central Universities under Grant 3132023238.

Data Availability Statement

Data are contained within this article. The source code of the proposed algorithm will be available at https://github.com/YuleiWang1.

Acknowledgments

We would like to thank the authors of the compared algorithms for their open-source codes.

Conflicts of Interest

The authors declare no conflicts of interest.

Appendix A. BS Results and Analysis for PU

In this appendix, the performance of different BS algorithms is compared using the PU dataset.
As shown in Figure A1, the OA, AA, and Kappa curves generated via various BS approaches with varying numbers of bands selected on the PU dataset are computed and recorded. The proposed MLRLFMESC method’s performance for OA, AA, and Kappa initially and gradually improves as selected bands grow, then shows a downward trend from n = 20 to n = 25, followed by an upward trend; this is the behavior known as the “Hughes” phenomenon. The proposed MLRLFMESC algorithm wins over the other five comparable BS methods with a significant classification priority in terms of OA, AA, and Kappa performance on the PU dataset, as illustrated in Figure A1, especially when n = 5, 10, 15, 20, and the OA, AA, and Kappa values have a clear advantage compared to the sub-optimal approach.
Figure A1. Classification performance of PU dataset with a different number of selected bands, (a) OA, (b) AA, (c) Kappa.
Figure A1. Classification performance of PU dataset with a different number of selected bands, (a) OA, (b) AA, (c) Kappa.
Remotesensing 16 00224 g0a1
The index distribution of the six band selection algorithms for selecting 20 bands of the PU dataset is shown in Figure A2. The selected bands are distributed widely and discretely in the proposed MLRLFMESC algorithm.
Figure A2. Distribution of bands selected using various BS algorithms for PU dataset.
Figure A2. Distribution of bands selected using various BS algorithms for PU dataset.
Remotesensing 16 00224 g0a2
Figure A3 shows the real label image, the visual classification maps for six BS methods and 20 selected bands, and full bands on the PU dataset, illustrating that the classification maps contain many misclassified samples, and the proposed MLRLFMESC method has a slightly better classification result map than the other five BS methods.
Figure A3. Classification result maps with (a) a labeled image of PU, and (bh) classification result maps with 20 bands using UBS, E-FDPC, ISSC, ASPS_MN, DSC, MLRLFMESC, and all bands.
Figure A3. Classification result maps with (a) a labeled image of PU, and (bh) classification result maps with 20 bands using UBS, E-FDPC, ISSC, ASPS_MN, DSC, MLRLFMESC, and all bands.
Remotesensing 16 00224 g0a3
The same comparison is conducted for the PU dataset to quantitatively compare the performance of different BS algorithms, where the average classification results for all bands and all 20 BS algorithm-selected 20 bands are listed in Table A1 using OA, AA, and Kappa performance. The best values of OA, AA, and Kappa in each category are blacked out, and the second-best values are underlined. It can be seen that the proposed MLRLFMESC algorithm achieves the highest classification accuracy in most categories, except for all bands. Compared with the DSC algorithm, the MLRLFMESC algorithm exhibits 1.43%, 1.47%, and 2.13% improvements in OA, AA, and Kappa, respectively.
Table A1. Comparison of classification results from various BS approaches on the PU dataset.
Table A1. Comparison of classification results from various BS approaches on the PU dataset.
AlgorithmsUBSE-FDPCISSCASPS_MNDSCMLRLFMESCAll Bands
OA89.88 ± 0.5089.38 ± 0.9490.88 ± 0.4992.23 ± 0.3692.63 ± 0.2194.06 ± 0.3895.71 ± 0.31
AA86.39 ± 0.7185.88 ± 1.1387.17 ± 0.7488.33 ± 0.5489.01 ± 0.4390.48 ± 0.5692.95 ± 0.82
Kappa84.94 ± 0.6684.12 ± 1.3286.41 ± 0.7388.42 ± 0.5788.99 ± 0.3791.12 ± 0.5593.57 ± 0.43
196.10 ± 0.5695.85 ± 1.0095.88 ± 0.7296.16 ± 0.9096.23 ± 0.8196.70 ± 0.7497.23 ± 0.63
295.16 ± 0.4594.40 ± 0.3696.01 ± 0.6097.15 ± 0.3897.19 ± 0.5598.09 ± 0.5598.82 ± 0.23
377.67 ± 4.2675.45 ± 4.2278.62 ± 2.4680.28 ± 3.5080.86 ± 2.7482.85 ± 2.3687.37 ± 3.99
482.13 ± 4.7485.88 ± 2.9585.77 ± 3.8290.16 ± 3.6390.81 ± 2.4191.85 ± 3.0394.10 ± 2.87
599.81 ± 0.5399.79 ± 0.5399.79 ± 0.3699.82 ± 0.3599.80 ± 0.7099.77 ± 0.5499.87 ± 0.19
670.20 ± 4.4968.49 ± 7.4673.34 ± 4.8277.17 ± 3.7778.36 ± 2.0883.61 ± 2.0688.14 ± 1.43
776.76 ± 4.1874.60 ± 4.4774.86 ± 5.7273.81 ± 3.1276.65 ± 2.8979.09 ± 5.3585.05 ± 4.85
879.70 ± 2.7878.50 ± 1.9380.32 ± 2.6780.47 ± 2.0281.19 ± 2.3482.47 ± 1.5386.00 ± 2.60
9100.00 ± 0.00100.00 ± 0.0099.97 ± 0.2599.92 ± 0.25100.00 ± 0.0099.90 ± 0.5099.95 ± 0.25

Appendix B. BS Results and Analysis for SA

In this appendix, the performance of different BS algorithms is compared using the SA dataset.
Figure A4 compares the classification curves for the proposed MLRLFMESC method and alternative BS methods selecting different numbers of bands, as well as the classification performance of all bands. For the SA dataset, the proposed MLRLFMESC algorithm shows the stable and best OA and Kappa accuracy performance when compared to the other five comparable BS methods, as seen in Figure A4; as n increases, the proposed MLRLFMESC method gradually increases in terms of the OA and Kappa performance. Notably, when n = 15, E-FDPC outperforms MLRLFMESC in terms of AA, and MLRLFMESC outperforms ASPS_MN in terms of AA when n = 20. However, the MLRLFMESC method outperforms the other five methods in terms of the AA performance for the majority of selected bands.
Figure A4. Classification performance of SA dataset with a different number of selected bands, (a) OA, (b) AA, (c) Kappa.
Figure A4. Classification performance of SA dataset with a different number of selected bands, (a) OA, (b) AA, (c) Kappa.
Remotesensing 16 00224 g0a4
Figure A5 provides the index distribution of the six band selection methods with 30 bands selected, where similar conclusions can be given as IP and PU datasets.
Figure A5. Distribution of bands selected using various BS algorithms for SA dataset.
Figure A5. Distribution of bands selected using various BS algorithms for SA dataset.
Remotesensing 16 00224 g0a5
Figure A6 shows the true label image and the classification result maps for six BS methods, selecting 30 bands and all bands on the SA dataset. The visual map of the classification results shows that all bands and the other six methods have several clear misclassifications. All bands maintain better regional consistency and provide a visualization closer to the ground truth.
The quantitative average classification results of all BS algorithms for 30 bands and all bands are shown in Table A2, with the best results blacked out and the second-best results underlined. Compared with DSC, MLRLFMESC has a slightly higher OA, AA, and Kappa accuracy. For each category of accuracy, all BS methods achieved a performance of 98% or above except for categories 8 and 15, and the proposed MLRLFMESC approach achieved the best category accuracy in categories 8 and 15 compared to the other five approaches.
Figure A6. Classification result maps with a (a) labeled image of SA, and (bh) classification result maps with 30 bands using UBS, E-FDPC, ISSC, ASPS_MN, DSC, MLRLFMESC, and all bands.
Figure A6. Classification result maps with a (a) labeled image of SA, and (bh) classification result maps with 30 bands using UBS, E-FDPC, ISSC, ASPS_MN, DSC, MLRLFMESC, and all bands.
Remotesensing 16 00224 g0a6
Table A2. Comparison of classification results from various BS approaches on SA dataset.
Table A2. Comparison of classification results from various BS approaches on SA dataset.
AlgorithmsUBSE-FDPCISSCASPS_MNDSCMLRLFMESCAll Bands
OA92.05 ± 0.8592.18 ± 0.5091.80 ± 0.6292.12 ± 0.5392.11 ± 0.5692.49 ± 0.7394.02 ± 0.35
AA97.10 ± 0.4497.27 ± 0.3297.08 ± 0.3997.30 ± 0.3397.21 ± 0.2397.40 ± 0.3297.79 ± 0.21
Kappa90.83 ± 0.9890.97 ± 0.5990.53 ± 0.6990.90 ± 0.6290.89 ± 0.6491.32 ± 0.8493.10 ± 0.41
1100.00 ± 0.0099.92 ± 0.1299.98 ± 0.1299.95 ± 0.1299.97 ± 0.2399.99 ± 0.1299.95 ± 0.12
299.92 ± 0.2499.92 ± 0.1999.88 ± 0.3499.92 ± 0.2499.92 ± 0.2999.95 ± 0.1999.95 ± 0.24
399.04 ± 1.2799.80 ± 0.7099.68 ± 0.4799.76 ± 0.5799.61 ± 0.8099.78 ± 0.8099.87 ± 0.34
499.36 ± 0.9999.44 ± 0.9999.45 ± 0.9799.43 ± 1.3099.38 ± 0.8399.49 ± 0.9899.38 ± 1.14
599.01 ± 1.5399.50 ± 0.8699.48 ± 0.7799.48 ± 1.0299.46 ± 1.2899.50 ± 1.1199.53 ± 1.03
699.97 ± 0.1399.96 ± 0.1399.96 ± 0.0999.94 ± 0.1899.96 ± 0.1399.96 ± 0.0999.96 ± 0.13
799.97 ± 0.0599.98 ± 0.0599.99 ± 0.0599.94 ± 0.2199.96 ± 0.1099.99 ± 0.1099.97 ± 0.15
884.46 ± 2.6784.34 ± 2.5983.56 ± 1.6783.95 ± 2.4984.48 ± 2.4084.73 ± 2.5989.96 ± 1.57
999.67 ± 0.1499.67 ± 0.1999.62 ± 0.2699.65 ± 0.1499.68 ± 0.1799.73 ± 0.1999.73 ± 0.21
1098.07 ± 1.2298.71 ± 0.8898.38 ± 1.4898.48 ± 1.5598.52 ± 1.1198.58 ± 0.6598.37 ± 1.29
1198.86 ± 2.0899.23 ± 0.8598.56 ± 1.9199.94 ± 0.2299.36 ± 1.6999.21 ± 1.0699.62 ± 1.48
1299.71 ± 0.7099.61 ± 0.8199.78 ± 0.7099.64 ± 0.5999.62 ± 0.8399.62 ± 1.1699.69 ± 0.47
1399.67 ± 0.7899.80 ± 0.5199.75 ± 0.7899.90 ± 0.5299.75 ± 0.7499.95 ± 0.2699.80 ± 0.51
1498.84 ± 2.6798.73 ± 2.8698.69 ± 2.1298.92 ± 2.6798.57 ± 3.0499.03 ± 1.8698.52 ± 2.23
1577.39 ± 2.7977.88 ± 1.5476.95 ± 3.7178.14 ± 2.2277.36 ± 2.3079.08 ± 2.5980.80 ± 1.78
1699.68 ± 0.8799.77 ± 0.6399.62 ± 0.6399.75 ± 0.5199.75 ± 0.6499.75 ± 0.7799.47 ± 1.25

References

  1. Wang, J.; Liu, J.; Cui, J.; Luan, J.; Fu, Y. Multiscale fusion network based on global weighting for hyperspectral feature selection. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2023, 16, 2977–2991. [Google Scholar] [CrossRef]
  2. Yu, C.; Zhou, S.; Song, M.; Gong, B.; Zhao, E.; Chang, C.-I. Unsupervised hyperspectral band selection via hybrid graph convolutional network. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5530515. [Google Scholar] [CrossRef]
  3. Zhong, Y.; Wang, X.; Xu, Y.; Wang, S.; Jia, T.; Hu, X.; Zhao, J.; Wei, L.; Zhang, L. Mini-UAV-borne hyperspectral remote sensing: From observation and processing to applications. IEEE Geosci. Remote Sens. Mag. 2018, 6, 46–62. [Google Scholar] [CrossRef]
  4. Ghamisi, P.; Yokoya, N.; Li, J.; Liao, W.; Liu, S.; Plaza, J.; Rasti, B.; Plaza, A. Advances in hyperspectral image and signal processing: A comprehensive overview of the state of the art. IEEE Geosci. Remote Sens. Mag. 2017, 5, 37–78. [Google Scholar] [CrossRef]
  5. Wang, Q.; Li, Q.; Li, X. Hyperspectral band selection via adaptive subspace partition strategy. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2019, 12, 4940–4950. [Google Scholar] [CrossRef]
  6. Li, S.; Liu, Z.; Fang, L.; Li, Q. Block diagonal representation learning for hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5509213. [Google Scholar] [CrossRef]
  7. Ma, M.; Mei, S.; Li, F.; Ge, Y.; Du, Q. Spectral correlation-based diverse band selection for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5508013. [Google Scholar] [CrossRef]
  8. Deng, Y.-J.; Li, H.-C.; Tan, S.-Q.; Hou, J.; Du, Q.; Plaza, A. t-Linear tensor subspace learning for robust feature extraction of hyperspectral images. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5501015. [Google Scholar] [CrossRef]
  9. Yu, W.; Huang, H.; Shen, G. Deep spectral–spatial feature fusion-based multiscale adaptable attention network for hyperspectral feature extraction. IEEE Trans. Instrum. Meas. 2023, 72, 5500813. [Google Scholar] [CrossRef]
  10. Ou, X.; Wu, M.; Tu, B.; Zhang, G.; Li, W. Multi-objective unsupervised band selection method for hyperspectral images classification. IEEE Trans. Image Process. 2023, 32, 1952–1965. [Google Scholar] [CrossRef]
  11. Li, S.; Peng, B.; Fang, L.; Zhang, Q.; Cheng, L.; Li, Q. Hyperspectral band selection via difference between intergroups. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5503310. [Google Scholar] [CrossRef]
  12. Habermann, M.; Fremont, V.; Shiguemori, E.H. Supervised band selection in hyperspectral images using single-layer neural networks. Int. J. Remote Sens. 2019, 40, 3900–3926. [Google Scholar] [CrossRef]
  13. Esmaeili, M.; Abbasi-Moghadam, D.; Sharifi, A.; Tariq, A.; Li, Q. Hyperspectral image band selection based on CNN embedded GA (CNNeGA). IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2023, 16, 1927–1950. [Google Scholar] [CrossRef]
  14. Sun, W.; Yang, G.; Peng, J.; Du, Q. Hyperspectral band selection using weighted kernel regularization. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2019, 12, 3665–3676. [Google Scholar] [CrossRef]
  15. Sellami, A.; Farah, I.R. A spatial hypergraph based semi-supervised band selection method for hyperspectral imagery semantic interpretation. Int. J. Comput. Inf. Eng. 2016, 10, 1839–1846. [Google Scholar]
  16. He, F.; Nie, F.; Wang, R.; Jia, W.; Zhang, F.; Li, X. Semisupervised band selection with graph optimization for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2021, 59, 10298–10311. [Google Scholar] [CrossRef]
  17. Cao, X.; Wei, C.; Ge, Y.; Feng, J.; Zhao, J.; Jiao, L. Semi-supervised hyperspectral band selection based on dynamic classifier selection. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2019, 12, 1289–1298. [Google Scholar] [CrossRef]
  18. Chein, I.C.; Su, W. Constrained band selection for hyperspectral imagery. IEEE Trans. Geosci. Remote Sens. 2006, 44, 1575–1585. [Google Scholar] [CrossRef]
  19. Jia, S.; Tang, G.; Zhu, J.; Li, Q. A novel ranking-based clustering approach for hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2016, 54, 88–102. [Google Scholar] [CrossRef]
  20. Fu, H.; Zhang, A.; Sun, G.; Ren, J.; Jia, X.; Pan, Z.; Ma, H. A novel band selection and spatial noise reduction method for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5535713. [Google Scholar] [CrossRef]
  21. Ji, H.; Zuo, Z.; Han, Q.L. A divisive hierarchical clustering approach to hyperspectral band selection. IEEE Trans. Instrum. Meas. 2022, 71, 5014312. [Google Scholar] [CrossRef]
  22. Sun, W.; Tian, L.; Xu, Y.; Zhang, D.; Du, Q. Fast and robust self-representation method for hyperspectral band selection. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2017, 10, 5087–5098. [Google Scholar] [CrossRef]
  23. Zhang, Y.; Wang, X.; Jiang, X.; Zhou, Y. Robust dual graph self-representation for unsupervised hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5538513. [Google Scholar] [CrossRef]
  24. Zhang, Y.; Wang, X.; Jiang, X.; Zhou, Y. Marginalized graph self-representation for unsupervised hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5516712. [Google Scholar] [CrossRef]
  25. Sun, W.; Zhang, L.; Du, B.; Li, W.; Lai, Y.M. Band selection using improved sparse subspace clustering for hyperspectral imagery classification. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2015, 8, 2784–2797. [Google Scholar] [CrossRef]
  26. Zhai, H.; Zhang, H.; Zhang, L.; Li, P. Laplacian-regularized low-rank subspace clustering for hyperspectral image band selection. IEEE Trans. Geosci. Remote Sens. 2019, 57, 1723–1740. [Google Scholar] [CrossRef]
  27. Huang, S.; Zhang, H.; Pižurica, A. A structural subspace clustering approach for hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5509515. [Google Scholar] [CrossRef]
  28. Sun, H.; Ren, J.; Zhao, H.; Yuen, P.; Tschannerl, J. Novel gumbel-softmax trick enabled concrete autoencoder with entropy constraints for unsupervised hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5506413. [Google Scholar] [CrossRef]
  29. Zhang, H.; Sun, X.; Zhu, Y.; Xu, F.; Fu, X. A global-local spectral weight network based on attention for hyperspectral band selection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6004905. [Google Scholar] [CrossRef]
  30. Zhang, X.; Xie, W.; Li, Y.; Lei, J.; Du, Q.; Yang, G. Rank-aware generative adversarial network for hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2022, 60, 5521812. [Google Scholar] [CrossRef]
  31. Feng, J.; Bai, G.; Li, D.; Zhang, X.; Shang, R.; Jiao, L. MR-selection: A meta-reinforcement learning approach for zero-shot hyperspectral band selection. IEEE Trans. Geosci. Remote Sens. 2023, 61, 5500320. [Google Scholar] [CrossRef]
  32. Zeng, M.; Cai, Y.; Cai, Z.; Liu, X.; Hu, P.; Ku, J. Unsupervised hyperspectral image band selection based on deep subspace clustering. IEEE Geosci. Remote Sens. Lett. 2019, 16, 1889–1893. [Google Scholar] [CrossRef]
  33. Das, S.; Pratiher, S.; Kyal, C.; Ghamisi, P. Sparsity regularized deep subspace clustering for multicriterion-based hyperspectral band selection. IEEE J. Sel. Top. Appl. Earth Observ. Remote Sens. 2022, 15, 4264–4278. [Google Scholar] [CrossRef]
  34. Dou, Z.; Gao, K.; Zhang, X.; Wang, H.; Han, L. Band selection of hyperspectral images using attention-based autoencoders. IEEE Geosci. Remote Sens. Lett. 2021, 18, 147–151. [Google Scholar] [CrossRef]
  35. Goel, A.; Majumdar, A. K-means embedded deep transform learning for hyperspectral band selection. IEEE Geosci. Remote Sens. Lett. 2022, 19, 6008705. [Google Scholar] [CrossRef]
  36. Kang, X.D.; Li, S.T.; Benediktsson, J.A. Spectral—Spatial Hyperspectral Image Classification with Edge-Preserving Filtering. IEEE Trans. Geosci. Remote Sens. 2014, 52, 2666–2677. [Google Scholar] [CrossRef]
Figure 1. Overall flowchart of the proposed MLRLFMESC framework.
Figure 1. Overall flowchart of the proposed MLRLFMESC framework.
Remotesensing 16 00224 g001
Figure 2. The IP dataset with (a) a pseudo-color map of IP, and (b) true image class distribution with labels.
Figure 2. The IP dataset with (a) a pseudo-color map of IP, and (b) true image class distribution with labels.
Remotesensing 16 00224 g002
Figure 3. The PU dataset, with (a) a pseudo-color map of PU, and (b) true image class distribution with labels.
Figure 3. The PU dataset, with (a) a pseudo-color map of PU, and (b) true image class distribution with labels.
Remotesensing 16 00224 g003
Figure 4. The SA dataset, with (a) a pseudo-color map of SA, and (b) true image class distribution with labels.
Figure 4. The SA dataset, with (a) a pseudo-color map of SA, and (b) true image class distribution with labels.
Remotesensing 16 00224 g004
Figure 5. Box plot of the OA for different BS methods on three hyperspectral datasets. (a) Indian Pines, (b) Pavia University, and (c) Salinas.
Figure 5. Box plot of the OA for different BS methods on three hyperspectral datasets. (a) Indian Pines, (b) Pavia University, and (c) Salinas.
Remotesensing 16 00224 g005
Figure 6. Classification performance with a different number of selected bands on IP, (a) OA, (b) AA, (c) Kappa.
Figure 6. Classification performance with a different number of selected bands on IP, (a) OA, (b) AA, (c) Kappa.
Remotesensing 16 00224 g006
Figure 7. Distribution of bands selected using various BS algorithms on IP data.
Figure 7. Distribution of bands selected using various BS algorithms on IP data.
Remotesensing 16 00224 g007
Figure 8. Classification result maps, (a) labeled image, and (bh) classification result maps on IP-selected 30 bands using UBS, E-FDPC, ISSC, ASPS_MN, DSC, MLRLFMESC, and all bands, respectively.
Figure 8. Classification result maps, (a) labeled image, and (bh) classification result maps on IP-selected 30 bands using UBS, E-FDPC, ISSC, ASPS_MN, DSC, MLRLFMESC, and all bands, respectively.
Remotesensing 16 00224 g008
Table 1. Descriptions of three HIS datasets.
Table 1. Descriptions of three HIS datasets.
DescriptionsIPPUSA
Size145 × 145610 × 340512 × 227
Bands200103204
Classes16916
Samples10,24942,77654,129
Resolution20 m1.3 m3.7 m
SensorsAVIRISROSISAVIRIS
Wavelength0.4–2.5 μm0.43–0.86 μm0.36–2.5 μm
Table 2. Ablation study for maximum entropy regularization and self-supervised learning.
Table 2. Ablation study for maximum entropy regularization and self-supervised learning.
Ablation ComponentsNumber of Selected Bands (OA Values)
L D S C ( L S C A E , L exp , L C c , L D ) L C F L S 510152025
××70.67 ± 1.4379.16 ± 1.2881.20 ± 0.6582.98 ± 0.4283.88 ± 0.52
×73.02 ± 1.4081.19 ± 1.6482.77 ± 1.9584.41 ± 2.5485.09 ± 1.33
×73.16 ± 1.4880.94 ± 0.9383.44 ± 0.0884.72 ± 0.6384.95 ± 0.86
74.15 ± 1.5982.90 ± 1.5885.05 ± 1.2585.77 ± 1.6186.00 ± 1.79
Table 3. Comparison of classification results from various BS approaches on the IP dataset.
Table 3. Comparison of classification results from various BS approaches on the IP dataset.
AlgorithmsUBSE-FDPCISSCASPS_MNDSCMLRLFMESCAll Bands
OA82.98 ± 1.3783.85 ± 1.5381.52 ± 1.1583.39 ± 1.2783.72 ± 1.5186.95 ± 1.0988.43 ± 1.24
AA81.86 ± 2.6783.48 ± 1.7381.55 ± 3.8082.65 ± 2.1882.11 ± 3.9786.43 ± 2.8786.69 ± 3.23
Kappa80.06 ± 1.5881.09 ± 1.6778.35 ± 1.3680.53 ± 1.4980.91 ± 1.8084.68 ± 1.2786.42 ± 1.47
176.30 ± 25.7485.47 ± 15.5885.26 ± 34.6281.20 ± 24.4478.16 ± 36.8486.28 ± 23.9091.89 ± 8.11
276.49 ± 4.9677.40 ± 5.6576.78 ± 4.1775.85 ± 3.2278.58 ± 3.3182.45 ± 3.2384.83 ± 1.95
373.75 ± 7.1577.19 ± 6.3271.97 ± 4.1475.46 ± 7.0171.73 ± 10.1680.19 ± 4.8083.54 ± 2.35
461.55 ± 10.4161.71 ± 13.5260.35 ± 7.0060.51 ± 8.6456.46 ± 10.9668.03 ± 13.1266.93 ± 3.82
579.61 ± 8.4087.84 ± 5.7081.81 ± 8.3786.96 ± 6.4088.57 ± 7.6889.94 ± 7.3592.48 ± 3.06
695.61 ± 3.3393.65 ± 2.7891.36 ± 5.3494.05 ± 2.7595.58 ± 2.7996.47 ± 3.3196.10 ± 1.42
792.80 ± 18.7586.72 ± 27.2793.18 ± 20.0093.27 ± 14.2987.56 ± 25.0093.24 ± 14.2987.73 ± 5.13
896.98 ± 2.6498.29 ± 2.9294.76 ± 3.5197.42 ± 1.7797.43 ± 3.9598.53 ± 2.1297.19 ± 1.83
963.60 ± 38.9677.04 ± 34.4476.29 ± 50.0063.45 ± 40.0067.75 ± 44.6475.43 ± 34.3471.39 ± 18.61
1073.09 ± 4.8373.26 ± 6.6869.52 ± 2.8474.68 ± 2.3374.67 ± 4.8777.00 ± 4.6180.28 ± 1.91
1186.69 ± 3.0987.79 ± 4.4185.33 ± 2.8386.28 ± 2.2587.46 ± 3.6289.38 ± 3.0891.93 ± 1.18
1278.99 ± 9.3178.34 ± 8.0375.92 ± 11.7381.29 ± 9.2175.12 ± 7.6383.69 ± 9.5881.97 ± 3.31
1398.10 ± 3.6696.52 ± 6.9394.52 ± 5.7595.62 ± 4.9296.97 ± 8.0797.91 ± 4.4997.45 ± 2.55
1496.86 ± 1.7796.15 ± 1.8396.48 ± 1.6496.43 ± 1.7296.89 ± 1.7597.11 ± 1.3397.50 ± 0.94
1563.32 ± 18.3363.06 ± 15.9557.77 ± 10.6364.37 ± 16.3866.52 ± 13.7071.87 ± 10.0968.07 ± 6.64
1695.95 ± 7.6995.27 ± 5.2593.49 ± 9.2095.53 ± 12.2094.26 ± 12.2095.34 ± 9.3097.77 ± 2.23
Table 4. Comparison of time consumption from various BS approaches on the IP dataset.
Table 4. Comparison of time consumption from various BS approaches on the IP dataset.
AlgorithmsMLRLFMESCDSCASPS_MNISSCE-FDPCUBS
Time (s)26.0739.810.540.491.54N/A
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wang, Y.; Ma, H.; Yang, Y.; Zhao, E.; Song, M.; Yu, C. Self-Supervised Deep Multi-Level Representation Learning Fusion-Based Maximum Entropy Subspace Clustering for Hyperspectral Band Selection. Remote Sens. 2024, 16, 224. https://doi.org/10.3390/rs16020224

AMA Style

Wang Y, Ma H, Yang Y, Zhao E, Song M, Yu C. Self-Supervised Deep Multi-Level Representation Learning Fusion-Based Maximum Entropy Subspace Clustering for Hyperspectral Band Selection. Remote Sensing. 2024; 16(2):224. https://doi.org/10.3390/rs16020224

Chicago/Turabian Style

Wang, Yulei, Haipeng Ma, Yuchao Yang, Enyu Zhao, Meiping Song, and Chunyan Yu. 2024. "Self-Supervised Deep Multi-Level Representation Learning Fusion-Based Maximum Entropy Subspace Clustering for Hyperspectral Band Selection" Remote Sensing 16, no. 2: 224. https://doi.org/10.3390/rs16020224

APA Style

Wang, Y., Ma, H., Yang, Y., Zhao, E., Song, M., & Yu, C. (2024). Self-Supervised Deep Multi-Level Representation Learning Fusion-Based Maximum Entropy Subspace Clustering for Hyperspectral Band Selection. Remote Sensing, 16(2), 224. https://doi.org/10.3390/rs16020224

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop