Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Perception in the Dark; Development of a ToF Visual Inertial Odometry System
Previous Article in Journal
Designing a Streaming Algorithm for Outlier Detection in Data Mining—An Incrementa Approach
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

Sparse Feature Learning of Hyperspectral Imagery via Multiobjective-Based Extreme Learning Machine

1
The Department of Computer Science, China University of Geosciences, Wuhan 430074, China
2
The Beibu Gulf Big Data Resources Utilization Laboratory, Beibu Gulf University, Qinzhou 535011, China
3
The Guangxi Key Laboratory of Marine Disaster in the Beibu Gulf, Beibu Gulf University, Qinzhou 535011, China
*
Authors to whom correspondence should be addressed.
Sensors 2020, 20(5), 1262; https://doi.org/10.3390/s20051262
Submission received: 1 January 2020 / Revised: 22 February 2020 / Accepted: 23 February 2020 / Published: 26 February 2020
(This article belongs to the Section Remote Sensors)
Figure 1
<p>(<b>a</b>) ELM-AE includes NPR and LR. Using <math display="inline"><semantics> <msup> <mi mathvariant="bold-italic">β</mi> <mi>T</mi> </msup> </semantics></math> as the transformation matrix to transform features. (<b>b</b>) AE consists of an encoder (red rhomboid box) and a decoder (green rhomboid box). The outputs of encoder represent the learned features.</p> ">
Figure 2
<p>Structure of EMO-ELM which consists of an encoder (red rhomboid box) and decoder (green rhomboid box). Where the red neurons in hidden layer denote they are activated while the orange neurons represent they are limited.</p> ">
Figure 3
<p>Pseudo-color image (<b>a</b>) and ground truth (<b>b</b>) of Salinas-A data set.</p> ">
Figure 4
<p>Pseudo-color image (<b>a</b>) and ground truth (<b>b</b>) of KSC data set.</p> ">
Figure 5
<p>Normalized Pareto front and solution selection of (<b>a</b>) SalinasA and (<b>b</b>) KSC data sets. The curvatures are normalized for plotting it in a same coordinate. The best compromise is denoted as the top three points of closing to the maximum curvature.</p> ">
Figure 6
<p>The 2-dimensional visualization of Iris dataset of (<b>a</b>) NRP, (<b>b</b>) SPCA, (<b>c</b>) ELM-AE, (<b>d</b>) SELM-AE, (<b>e</b>) AE, (<b>f</b>) SAE, (<b>g</b>) EMO-ELM(<math display="inline"><semantics> <msub> <mi>f</mi> <mn>1</mn> </msub> </semantics></math>), (<b>h</b>) EMO-ELM(<math display="inline"><semantics> <msub> <mi>f</mi> <mn>2</mn> </msub> </semantics></math>) and (<b>i</b>) EMO-ELM(best).</p> ">
Figure 7
<p>The sparsity of different algorithms of (<b>a</b>) SalinasA and (<b>b</b>) KSC data set.</p> ">
Figure 8
<p>Box plot of SalinasA and KSC data sets. (<b>a</b>–<b>c</b>) denotes the box plot of the SalinasA data set in terms of OA, AA, and Kappa, respectively; (<b>d</b>–<b>f</b>) represents the box plot of the KSC data set with respect to OA, AA, and Kappa, respectively. The edges of boxes are the 25th and 75th percentiles and the middle lines indicate the median line. Whiskers extend to the maximum and minimum points. Abnormal outliers are shown as “∘”s.</p> ">
Review Reports Versions Notes

Abstract

:
Hyperspectral image (HSI) consists of hundreds of narrow spectral band components with rich spectral and spatial information. Extreme Learning Machine (ELM) has been widely used for HSI analysis. However, the classical ELM is difficult to use for sparse feature leaning due to its randomly generated hidden layer. In this paper, we propose a novel unsupervised sparse feature learning approach, called Evolutionary Multiobjective-based ELM (EMO-ELM), and apply it to HSI feature extraction. Specifically, we represent the task of constructing the ELM Autoencoder (ELM-AE) as a multiobjective optimization problem that takes the sparsity of hidden layer outputs and the reconstruction error as two conflicting objectives. Then, we adopt an Evolutionary Multiobjective Optimization (EMO) method to solve the two objectives, simultaneously. To find the best solution from the Pareto solution set and construct the best trade-off feature extractor, a curvature-based method is proposed to focus on the knee area of the Pareto solutions. Benefited from the EMO, the proposed EMO-ELM is less prone to fall into a local minimum and has fewer trainable parameters than gradient-based AEs. Experiments on two real HSIs demonstrate that the features learned by EMO-ELM not only preserve better sparsity but also achieve superior separability than many existing feature learning methods.

1. Introduction

Hyperspectral Imagery (HSI), which is obtained by remote sensing systems, contains high-resolution spectral information over a wide range of the electromagnetic spectrum with hundreds of observed spectral bands [1]. The detailed spectral and spatial information provides the power of accurately differentiating or recognizing materials of interest [2]. In recent years, HSI has been applied in a wide variety of applications, including agriculture, surveillance, astronomy, and biomedical imaging, among others [3]. However, a great number of redundancies between spectral bands bring heavy computation burdens in HSI data analysis [4]. Furthermore, high dimensionality even rises “Hughes” problem, namely the so-called curse of dimensionality [5,6].
As an effective solution, feature learning overcomes these issues and guarantees good classification accuracy [7,8,9,10,11,12,13]. The conventional feature learning methods, such as Principal Component Analysis (PCA) [9] and its variants [14,15,16] are widely applied in HSI. However, PCA is a linear combination of all the original data, thus it is often difficult to interpret the results. To improve PCA’s performance, Sparse PCA (SPCA) [11] was proposed using the lasso ( elastic net) to produce modified principal components with sparse loadings. Besides that, Autoencoder (AE) [17] is a usual recent feature learning method. AE is a well-known unsupervised neural network which is composed of an encoder and a decoder. AE aims at learning an identity transformation for raw inputs [18]. The learned features are generally represented as outputs of the encoder (the hidden layer). In many applications, labels of hyperspectral imagery are expensive and difficult to obtain, thus AE can work as an effective unsupervised feature learning approach and be widely applied in the hyperspectral image [19,20,21,22]. However, AE is faced with its instinct limitations that need to be solved. On the one hand, AE has to simultaneously update the parameters for encoder and decoder, that means the high cost of parametric learning. On the other hand, the vanishing gradient problem exists in the training of AE that bases on the backpropagation algorithm [23]. To solve it, sparse AE(SAE) was proposed by imposing sparse constraints on the encoder to guarantee the sparsity of coding results [24]. In this case, new hyperparameters could be introduced so that it is difficult to determine the best hyperparameters. However, parameters of AE and SAE need to be adjusted frequently, resulting in high cost of learning time.
Extreme Learning Machine( ELM) was originally proposed for generalized single-hidden layer feedforward neural networks [25,26,27]. Due to its fast learning speed, good generalization ability, and ease of implementation, ELM has become a popular research topic and been widely used for supervised learning of HSI [28,29,30,31,32]. Although ELM is successfully used for supervised learning in hyperspectral image classification, it is also difficult to avoid a large number of labeled samples. However, the number of labeled samples is so small because of expensively and difficultly labeling for samples in the hyperspectral imagery. Thus, an unsupervised feature learning approach is strongly desired to solve it. But unsupervised feature learning based on ELM has not drawn much attention, the reason for which is mainly that ELM is limited by its hidden layer parameters that are generated randomly. Actually, in terms of AE, ELM can be viewed as a Nonlinear Random Projection (NRP) combined with a Linear Regression (LR). The non-optimal NRP weights could result in its outputs including noise and redundancies. Compared with AE, ELM can learn quickly without adjusting parameters. However, ELM is difficult for unsupervised learning. Based on the above consideration, ELM-AE [33] is proposed by taking into account the merits of AE and ELM, and applied to unsupervised feature learning in hyperspectral imagery. Compared with AE, ELM-AE only deals with the parameters of an encoder, instead of decoder, which means the complexity of the problem is decreased. Meanwhile, it preserves the strong capability to extract features. Besides, it can handle the unsupervised feature learning for hyperspectral imagery that is what the basic ELM can’t have. According to the advantages mentioned above, ELM-AE is often used for unsupervised learning and constructing multilayer ELM [34,35,36]. In [37], ELM-AE has been proven effective in dimension reduction. To increase the sparsity of parameters, Sparse Random Projection (SRP) [38,39] was introduced into ELM-AE and provided better generalization ability. ELM-AE and sparse ELM-AE(SELM-AE) essentially are a linear transformation of initial features using LR’s weights as a transform matrix. However, according to the above-mentioned methods, the parameters of the hidden layer are found out in the matrix transformation way, which is not suitable for the fact that a large number of nonlinear data exist in the hyperspectral image. In the global data domain, Evolutionary Multiobjective Optimization(EMO) can obtain the nonlinear solution in the form of parallel. Evolutionary Multiobjective-based ELM(EMO-ELM) is proposed to solve the nonlinear problem of ELM-AE and SELM-AE.
The motivation of this paper is to propose a novel sparse unsupervised feature learning approach, and then deal with nonlinear data evolutionarily. Based on the consideration above, our paper proposes a novel AE framework for sparse feature learning by treating NRP and LR in ELM-AE as encoder and decoder. Different from ELM-AE, the proposed method provides the optimal weights and bias by EMO. Similar to the conventional AE, we use the outputs of encoder to represent the learned feature. The decoder’s weights are directly calculated through the ELM theories, thus we only need to update the parameters in the NRP. Our goal is to reduce the parameters involved in AE and improve ELM-AE. Due to that the gradient-based methods are disabled in our method, we design a multiobjective model composed of two conflicting objectives, including the sparsity of learned features and reconstruction error. In order to solve this two-objective model, EMO is effectively employed to handle this type of optimization problem. Comparing with AE and SAE, the number of parameters have been reduced to half; comparing with ELM-AE and SELM-AE, this approach learns features by a nonlinear manner, moreover, the using of optimization method avoids the influence of NRP. Through constructing two conflicting objectives composed of the sparsity and reconstruction error, and extracting features from the selected solutions based on optimal Pareto Front (PF) obtained by EMO, EMO-ELM improves the classification accuracy while ensuring the sparsity. Experimental results verify that the purposed method outperforms NRP, SPCA, AE, SAE, ELM-AE and SELM-AE in terms of the sparsity and reconstruction error.
The main contributions of this paper can be summarized as follows: (1) A novel unsupervised feature learning approach is proposed; (2) We combine the advantages of basic ELM and conventional AE; (3) A hybrid feature learning framework is presented which uses EMO for training.
The remainder of this paper is organized as follows. Section 2 reviews the related work, including ELM-AE and AE. Section 3 describes the proposed EMO-ELM method in detail. The experimental results and discussions are shown in Section 4. Finally, the conclusions are represented in Section 5.

2. Related Work

2.1. ELM-AE

ELM-AE can be divided into NRP and LR, and its structure is shown in Figure 1a. Let us consider N distinct original samples X = x i i = 1 N , where in x i R n (n-dimensional feature space). For example, ELM-AE is used to extract features of SalinasA data set for comparison in this paper, then N = 86 × 83 and n = 204 . Supposing the dimension of the extracted feature X is 100, X is the preprocessed data of N = 86 × 83 samples that have the same number of samples as the original data X. The input and output layers have the same number of neurons and the input and output vectors have the same dimension, which is n. The input layer is equivalent to the encoding part of the automatic encoder, while the output layer to the decoding part. Further supposing that the number of hidden neurons L is less than the number of hidden neurons n, the weight matrix is orthogonal, we see that the first part of ELM-AE aims at mapping X into an L-dimensional space by the following equation:
h i = g W T x i + b , i = 1 , , N W T W = I b T b = 1
where h i = h i 1 , h i 2 , , h i L T R L , x i = x i 1 , x i 2 , , x i n T R n , W = w 1 , w 2 , , w L R n × L and b = b 1 , b 2 , , b L T R L denotes the output vector of a hidden layer for the i -th input sample, the vector of the i -th input sample, NRP’s orthogonal weights matrix connecting the neurons of the input layer with the neurons of the hidden layer, and the orthogonal bias vector in the hidden layer respectively. The output weights β of ELM-AE are responsible for the transformation from the feature space to input data and satisfy:
Minimize β | | H β X | | 2
According to ELM-AE’s theories, the output weights β of ELM-AE are:
β = W T V V T
where V is the eigenvectors of covariance matrix X T X .
The new feature can be achieved in ELM-AE by projecting the data X along the decoder stage weight β as:
X = X β T
X is the new feature and can be applied to hyperspectral image classification.
When the number of input neurons in ELM-AE is smaller than the number of hidden neurons ( n < L ), the weight matrix is sparse. According to the ELM theories, W and b can be randomly generated. g · is any nonlinear activation function, such as Sigmoid function. To further increase the model’s sparsity, W and b , which are orthogonal, can be generated using the following equation [37]:
w i j = b j = 1 / L × + 3 p = 1 / 6 0 p = 2 / 3 3 p = 1 / 6
where w i j indicates the weight connecting the i-th neuron in the input layer with the j-th neuron in the hidden layer, b j is the bias of the j-th neuron in the hidden layer, and p represents the ratio of elements in the sparse random matrix W . w i j and b j are generated randomly according to the p-value, similar to the wheel algorithm. It is assumed that there is a random probability r before every matrix element is generated, and the value range of r is between 0 and 1. The elements in the matrix change according to the random generation probability r. When the probability r is less than 1/3, the values of w i j and b j are 3 L ; when the random probability r is more than 1/3 and less than 5/6, the values of w i j and b j are 0; when the random probability r is greater than 5/6 and less than 1, the values of w i j and b j are 3 L .
The output weights β of SELM-AE are calculated as:
Minimize β | | H β X | | 2
According to ELM-AE’s theories, the output weight β of SELM-AE are:
β = W T V V T
where V is the eigenvectors of covariance matrix X T X .
Dimension reduction is achieved in SELM-AE by projecting the data X along the decoder stage weights β as:
X = X β T
The second part of ELM-AE is a well-known ridge regression or regularized least squares, which aims to solve the output weights matrix β of which β j i indicates the weight connecting the j-th neuron in the hidden layer with the i-th neuron in the output layer by minimizing the following learning problem:
Minimize β : β t σ 1 + λ H β X q σ 2
The first term in the objective function is the output weights matrix, and the second term represents the reconstruction errors. λ is a regularization term that controls the complexity of the learned model. Here, according to the ELM theorem [37,40], σ 1 > 0 , σ 2 > 0 , t, and q can be any matrix norm (e.g., t , q = 0 , 1 2 , 1 , 2 , , + ). To obtain the analytical solution, σ 1 , σ 2 , t and q are often set to 2, which is also known as a ridge regression problem. However, t , q = 0 (i.e., L0-norm) are possible. Under the circumstances, β 0 denotes the number of non-zero weights in the hidden layer, and H β X 0 indicates that the amount of non-zero reconstruction errors for the original data X .
When the penalty coefficient λ is infinitely large in Equation (9), β can be calculated as
β = H X
where H indicates the Moore–Penrose generalized inverse of H . According to the ELM learning theory, ELM with a minimum norm of output weights has better generalization performance and a more robust solution [41].
Different output weights β can be obtained based on the concerns on the efficiency in different size of training data sets. When the number of training samples is not huge, the output weight β can be expressed as
β = H T H H T + I λ 1 T
The output function of ELM classifier is
f ( x ) = h ( x ) β = h ( x ) H T H H T + I λ 1 T
When the number of training samples is very large, the output weight β can be represented as
β = H T H + I λ 1 H T X
And the output function of ELM classifier is
f ( x ) = h ( x ) β = h ( x ) H T H + I λ 1 H T T
Subsequently, the new feature X can be represented as:
X = X β T

2.2. AE

AE has a symmetrical encoder-to-decoder structure as seen in Figure 1b. For the sake of simplicity, we represent AE’s cost function as
J Θ e , Θ d = L X F d F e X ; Θ e ; Θ d + λ R F e
where L is the loss term such as Mean Squared Error (MSE); R is the regularization term with the regularization coefficient λ , in SAE, R can be written as Θ e p or F e X ; Θ e p ; F e and F d denote encoder and decoder’s transformation with hyperparameters Θ e and Θ d , respectively. The training of AE can be finished using the gradient backpropagation algorithms.
Θ = Θ μ J Θ , Θ = [ Θ e , Θ d ]
Here μ denotes the learning rate, and J Θ indicates the gradient of Θ . Unlike ELM-AE, AE represents the learned features using the outputs of an encoder and it can be written as
X = F e X ; Θ e
Because F e is generally nonlinear, AE can process complicated data. For example, AE has been successfully used for HSI feature learning and classification in [24].

3. EMO-ELM

The structure of the proposed approach is given in Figure 2. We still keep ELM-AE’s structure and parameter settings but use parametric learning steps by encoder-to-decoder like AE. In our approach, the gradient-based tuning method is difficult to work, therefore a multiobjective-based method is taken into consideration. In this section, we first introduce the multiobjective construction; next, we give the Non-dominated Sorting Genetic Algorithm II(NSGA-II)-based learning steps; then we describe the solution selection procedure; finally, we use the proposed approach to represent the sparse feature.

3.1. Constructing a Multiobjective Model

Our multiobjective model contains two conflicting objective functions with respect to the decision variable vector α , where α = [ w 11 , , w 1 L , , w n 1 , , w n L , b 1 , , b L ] T , and we indicate them as f 1 and f 2 , respectively. The detailed definitions of f 1 and f 2 are given below.
Firstly, we define f 1 as the sparseness of the encoded results. Suppose the activation function of the hidden neurons to be the Sigmoid function, then the activation values in the hidden layer will be limited in 0 , 1 . Informally, we think of a hidden neuron as being “active” (or as “firing”) if its output value is close to 1, or as being “inactive” if its output value is close to 0. The lower the probability of the activation function among the hidden layer, the sparser the encoded results, and the smaller f 1 is. Our goal is to constrain the hidden neurons to be inactive most of the time. We first define the average activation of the j-th hidden neuron as
ρ ^ j = 1 N i = 1 N g i j x i = 1 N i = 1 N h i j
Further, let ρ be the desired activation, our goal is forcing ρ ^ j to close to ρ . To measure the difference between ρ and ρ ^ j , the Kullback-Leibler(KL)-divergence is takeninto consideration, which can be expressed as follows
KL ρ ρ ^ j = ρ log ρ ρ ^ j + 1 ρ log 1 ρ 1 ρ ^ j
Considering all the hidden neurons, f 1 is represented as the sum of the KL-divergence of all hidden neurons.
f 1 = j = 1 L KL ρ ρ ^ j
Secondly, we define f 2 as the Root Mean Squared Error (RMSE) of the raw input and the decoded results. However, the decoder determines parameters mathematically, we adopt K-fold crossover validation to calculate RMSE to further avoid overfitting. Thus, f 2 can be denoted as
f 2 = 1 K k = 1 K i = 1 n k x i y x i 2 2 n k × n
where K indicates a K-fold crossover validation is used, n k is the number of validation samples in k-th fold subjected to k = 1 K n k = N , and y x i is the decoded results for x i . Ultimately, we aim to simultaneously minimize the two functions. For clarity, the overall multiobjective model is expressed as
arg min α F α = arg min α f 1 , f 2

3.2. Solving a Multiobjective Model

To effectively solve the optimization problem given in Equation (23), we employ EMO algorithms. Specifically, NSGA-II [42] is used in this paper due to its fast solvingspeed, excellent convergence, and popular applications. All kinds of evolutionary multiobjective algorithms such as Multi-Objective Evolutionary Algorithm based on Decomposition (MOEA/D) [43] and Multi-Objective Particle Swarm Optimization(MOPSO) [44], are practicable in our framework. The mainsteps using NSGA-II to solve our multiobjective model are given in Algorithm 1.
Algorithm 1: NSGA-II-based Solving
   Input: N P , g e n , other evolutionary parameters
   Output: Pareto optimal solution set
1 Initialize population;
2 while Termination criteria not met do
3 Sensors 20 01262 i001Elitist selection technique;
4Generic operations;
5Objectives evaluation;
6Fast nondominated sorting;
7Crowding distance assignment;
8 end

3.3. Selecting Solution

By evolving, we get a set of solutions, which is called Pareto optimal solutions. Generally, the corresponding objective values can be plotted as a curve named PF curve. Unlike single-objective optimization method, we have to choose from the Pareto optimal set, because these solutions are deemed to be equally important. In this paper, the following three solutions are considered as alternatives.
  • the solution getting a minimum value of f 1 ;
  • the solution getting a minimum value of f 2 ;
  • the solution locating at the knee area.
The first two solutions are easy to determine. However, focusing on the knee area is commonly difficult, because the PF could be not smooth and the objective functions involved in the model usually have greatly different magnitudes. To observe the PF, we find a fact that the knee usually occurs concurrently with a maximum curvature of PF curve. Thus, this problem is transformed to find out the maximum curvature of the PF curve. To do this, the following steps are carried out. Firstly, normalize the Pareto solution to overcome the different magnitudes. Next, smooth the PF by interpolating the PF using B-splines. And then evenly resample from the smooth spline. Finally, estimate the curvature according to the derivatives of the B-spline curve and select the solution which is closest to the maximum curvature.

3.4. Sparse Feature learning Using EMO-ELM

We denote the selected solution as α F . Subsequently, we use α F . to regenerate the parameters of the encoder, which is represented as W F and b F . Depending on this, the learned features can be obtained from the following processing.
X = g X W F + B F
where X is the learned features, and B F = [ b F T , , b F T ] N × L T . As seen in Equation (24), this procedure is a nonlinear transformation. The complete pseudocode of using EMO-ELM to extract sparse features is given in Algorithm 2.
Algorithm 2: EMO-ELM for sparse feature learning
   Input: X , ρ , L
   Output: Learned feature
1 Optimize Equation (23) according to Algorithm 1, and obtain the Pareto optimal solution set;
2 Select α F from the obtained Pareto optimal solution set according to selection criteria;
3 Regenerate W F and b F ;
4 Extract features according to Equation (24);
5 Return extracted features X ;

4. Experiments

4.1. Data Description and Experiment Design

4.1.1. SalinasA Data Set

This image was collected by the 224-band Airborne Visible/Infrared Imaging Spectrometer(AVIRIS) sensor over SalinasValley, California, and was characterized by high spatial resolution (3.7-meter pixels). The area covered comprises 86 × 83 pixels and includes six classes, of which the class information is given in Table 1. The pseudo-color image and its ground truth are shown in Figure 3. This scene has reduced the number of bands to 204 by removing 20 bands covering the region of water absorption: 108–112, 154–167, 224.

4.1.2. Kennedy Space Center (KSC) Data Set

This scene was acquired by NASA AVIRIS (Airborne Visible/Infrared Imaging Spectrometer) instrument over the Kennedy Space Center (KSC), Florida, on March 23, 1996. This data is with 512 × 614 pixels and has a spatial resolution of 18 m, which was shown in Figure 4. Regardless of discarded water absorption and low signal-to-noise ratio (SNR) bands, 176 spectral bands are used for classification. 13 different land cover classes available in the original dataset are displayed in Table 2.

4.2. Experiment Settings

Experiments will be organized into three parts for the convergence, sparsity, and separability. The first one aims at analyzing the convergence of EMO-ELM, and further illustrating the solution selection. In the second experiment, we quantitatively compare the sparsity of EMO-ELMs (EMO-ELM( f 1 ), EMO-ELM( f 2 ) and EMO-ELM(best) represent EMO-ELM with different strategies of solution selection respectively) with NRP (namely unoptimized ELM), SPCA [11], ELM-AE [37], SELM-AE [37], AE, and SAE. Moreover, Finally, we use the basic ELM with 500 hidden neurons as the classifier to estimate the classification capabilities of the learned features in terms of Overall Accuracy (OA), Average Accuracy (AA), and Kappa coefficient. In this part, all experiments are executed through three-fold cross-validation and repeated ten times and take the mean as the performance criteria. 10 percent of samples are used for training and the rest is for testing. For AE and SAE, we use the implementation available from the Keras website [45]. For EMO-ELM, the well-known NSGA-II (https://github.com/Project-Platypus/Platypus) is applied, in which we uniformly set N P = 50 and g e n = 5000 . Additionally, all experiments are carried out using Python 2.7 on an Intel i5 Core(TM) 3.2-GHz machine with 8 GB of RAM.

4.3. Convergence and Solution Selection

The optimizing of Multi-Objective Problems(MOP) requires the objectives that need to be solved are conflicted to each other. Thus, we can determine whether the algorithm is convergent by observing the PF. We consider algorithm is convergent if all these solutions are non-dominated. At this point, the PF looks like a ’smooth’ curve. In Figure 5, we illustrate the normalized PFs obtained by optimizing 10-hidden-neurons ELM with 5000 generations on SalinasA and KSC data sets. As we can see, after 5000-generation evolution, the final results are completely convergent. Furthermore, it was revealed that two objectives we constructed are conflicted. Observing the curvature curves, the values will be enlarged dramatically in the knee area, therefore we can accurately find out it.

4.4. Visual Investigation of Features Learned by Different Algorithms

Iris data set, which contains 150 samples with three classes, is suitable for visualization. Figure 6a–i show the distribution of the features of Iris data projected into two-dimensional space. Observed from Figure 6a, the original nonlinear random projection in ELM disorders the data distribution, which indicates that there are included noise in the original ELM. As seen in Figure 6b–i, the features of Iris are effectively distributed in the feature space after feature learning. By performing EMO-ELM, these noises have been dramatically reduced. From the cluster point of view, EMO-ELM decreased the within-class distance and increased the between-class distance. This phenomenon is especially evident in Figure 6i which uses the proposed knee-based solution selection strategy.

4.5. Measuring Sparsity of the Learned Features

In order to quantify the sparsity, we use the L 2 / L 1 sparsity measure, which is applied in [37,46]. L 2 / L 1 measure indicates sparsity at an abstract level. A higher L 2 / L 1 measure demonstrates that there are few large feature values and more small feature values. On the contrary, there are more large feature values and a few small feature values. In other words, the higher the L 2 / L 1 gets, the sparser the learned feature is. In Figure 7, we give the L 2 / L 1 sparsity measure of different algorithms under the various number of features. It is certain that the features learned by EMO-ELMs are sparser than all the competitors. Comparing with NRP, EMO-ELM has significantly increased the sparsity.

4.6. Comparison of Classification Ability

For feature learning, an important criterion for evaluating the effectiveness of learned features is classification ability. In this experiment, we compare EMO-ELM’s performance with the above-mentioned six feature learning approaches under 10-dimensional features. The results of SalinasA and KSC are given in Table 3 and Table 4. The best results are bold. As seen from Table 3 and Table 4, the EMO-ELMs, especially EMO-ELM( f 2 ), outperforms other methods, which means that the performance could be more excellent in optimization way. To visually present the differences between competitors, we also plot the statistical evaluations of OA, AA, and Kappa in Figure 8a–c for SalinasA and (d–f) for KSC. As shown in Figure 8, EMO-ELM( f 2 ) obtains the best results in comparison with other strategies in terms of all aspects, which is because that EMO-ELM( f 2 ) acquires the least reconstruction error. The detailed conclusions are given as follows:
  • EMO-ELM( f 1 ) v.s. EMO-ELM( f 2 ) v.s EMO-ELM(best): Generally, EMO-ELM( f 2 ) yields higher accuracy than other competitors in terms of mean performance for SalinasA and KSC. The reason is that because EMO-ELM( f 2 ) guarantees the model to achieve the smallest reconstruction error, whereas EMO-ELM( f 1 ), although, obtains the smallest sparsity, the feature reconstruction is limited. Reviewing Figure 7a,b, EMO-ELM( f 2 ) also maintains good sparsity. Hence, the solution selection strategy based on f 2 can be considered best in our experiments. EMO-ELM(best) also plays a trade-off role between sparsity and reconstruction error, thus we view it as the second choice.
  • NRP v.s. EMO-ELM: As shown in Figure 8a–f, the original nonlinear random projection (NRP) is effective in feature mapping, but EMO-ELM has shown that the NRP’s performance can be further improved after optimizing.
  • SPCA v.s EMO-ELM: The features learned by SPCA maintain the remarkable sparsity, however, the classification ability is damaged. Furthermore, as a dimension-reduction method, SPCA cannot work when the learned dimension is larger than the original dimension. On the contrary, EMO-ELM outperforms SPCA in respects of many tested performances.
  • ELM-AE and SELM-AE v.s EMO-ELM: As we known, ELM-AE and SELM-AE learn features linearly. In this experiment, SELM-AE performs better than ELM-AE due to the sparse matrix is used. Whereas, EMO-ELM learns features nonlinearly and has significantly enhanced the classification capacity of the learned features.
  • AE and SAE v.s EMO-ELM: The learning procedure of EMO-ELM is similar to AE and SAE, but EMO-ELM becomes more competitive in both respects of classification ability and sparsity after the same times of updating. Especially, EMO-ELM optimizes only the hidden layer, whereas AE and SAE have to simultaneously optimize the hidden layer and the output layer.

4.7. Discussion

Based on the above experimental results, we provide a meaningful discussion on this article:
  • The proposed approach can be regarded as a general framework that is composed of a nonlinear encoder and a linear decoder. For the optimization of this framework, we only need to focus on the encoder (or the hidden layer) since the decoder can be represented as a closed-form solution, which is very different from neural networks. Thus, compared with SAE and AE, the number of EMO-ELM’s parameters can be reduced to half.
  • In addition to the objectives used in this paper, various objectives, such as classification error and matrix norm constraints are considered in this framework. More importantly, the optimizer is replaceable and flexible. Therefore, EMO-ELM can be used as an alternative to unsupervised feature learning.
  • It is well known that the evolutionary operating is time-consuming, this is the main challenge faced in EMO-ELM. Thus, EMO-ELM is difficult to directly handle the big data. Fortunately, evolutionary algorithms are easy to implement in parallel. There is an issue that how to use EMO-ELM to do deep representation learning is worth studying in the future works.

5. Conclusions

This paper proposed an EMO-ELM approach for sparse feature learning of hyperspectral image. The main idea of EMO-ELM is that, firstly, using an EMO to optimize the hidden layer of ELM, then executing feature extraction according to the optimal hidden layer. To let EMO-ELM have the ability to learn sparse features, the hidden layer activations are constrained by an objective function. The other objective involved in EMO-ELM is the cross-validation RMSE, which ensures the learned features are accurate. Unlike ELM-AE which linearly transforms original features, EMO-ELM uses a nonlinear way to do this. This procedure is similar to neural network autoencoder, such AE, but EMO-ELM optimizes the hidden layer only. So that EMO-ELM can be updated using EMO. Experiments are carried out on two real hyperspectral image data sets, and the performance of EMO-ELM is extensively explored including the convergence, the sparsity, and the classification ability. According to the experimental results, we can draw the following conclusions:
  • The experimental results demonstrate that EMO-ELMs is more suitable to extract sparse features from the hyperspectral image, and EMO plays a more significant role in dealing with the nonlinear data of hyperspectral image.
  • The proposed EMO-ELM significantly improves the performance of the original ELM. These experimental results demonstrate that the optimized hidden layer of ELM is effective for HSI feature learning.
  • EMO-ELM generally outperforms ELM-AE, SELM-AE, AE, and SAE in terms of sparsity and classification ability because of two optimized objectives.
  • The knee-based solution selection strategy can accurately focus on the knee area of the PF curve. But, the RMSE-based solution selection strategy is more applicable in our experiments.

Author Contributions

X.F. was mainly responsible for mathematical modeling, experimental design and editing the paper. Y.C. carried out partial experiments. Z.C. (Zhihua Cai) and X.J. provided some helpful comments. Z.C. (Zhikun Chen) reviewed the paper. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported in part by the National Natural Science Foundation of China under Grant 61603355, Grant 61773355, Grant 61973285, and Grant 61873249, in part by the Fundamental Research Funds for National University, China University of Geosciences (Wuhan), under Grant 1910491T06, in part by the Qinzhou Scientific Research and Technology Development Plan Project under Grant 201714322, and in part by the High-Level Talents Research Projects of Beibu Gulf University under Grant 2019KYQD27.

Acknowledgments

The authors would like to thank Y.Zhang for sharing the related code for comparison purpose, and J.Jiang for helpful comments.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Li, W.; Tramel, E.W.; Prasad, S.; Fowler, J.E. Nearest regularized subspace for hyperspectral classification. IEEE Trans. Geosci. Remote Sens. 2014, 52, 477–489. [Google Scholar] [CrossRef] [Green Version]
  2. Chen, Y.; Li, C.; Ghamisi, P.; Jia, X.; Gu, Y. Deep fusion of remote sensing data for accurate classification. IEEE Geosci. Remote Sens. Lett. 2017. [Google Scholar] [CrossRef]
  3. Thenkabail, P.S. Hyperspectral data processing: Algorithm design and analysis. Photogramm. Eng. Remote Sens. 2015, 81, 441–442. [Google Scholar] [CrossRef]
  4. Sun, W.; Tian, L.; Xu, Y.; Zhang, D.; Du, Q. Fast and robust self-representation method for hyperspectral band selection. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2017, 10, 5087–5098. [Google Scholar] [CrossRef]
  5. Shippert, P. Introduction to hyperspectral image analysis. Online J. Space Commun. 2003, 3, 13. [Google Scholar]
  6. Cai, Y.; Liu, X.; Cai, Z. BS-Nets: An End-to-End framework for band selection of hyperspectral image. IEEE Trans. Geosci. Remote Sens. 2019. [Google Scholar] [CrossRef] [Green Version]
  7. Mei, S.; Ji, J.; Geng, Y.; Zhang, Z.; Li, X.; Du, Q. Unsupervised spatial-spectral feature learning by 3D convolutional autoencoder for hyperspectral classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6808–6820. [Google Scholar] [CrossRef]
  8. Luo, F.; Bo, D.; Zhang, L.; Zhang, L.; Tao, D. Feature learning using spatial-spectral hypergraph discriminant analysis for hyperspectral image. IEEE Trans. Cybern. 2019, 49, 2406–2419. [Google Scholar] [CrossRef]
  9. Rodarmel, C.; Shan, J. Principal component analysis for hyperspectral image classification. Geo. Spat. Inf. Sci. 2002, 62, 115. [Google Scholar]
  10. Kemker, R.; Kanan, C. Self-taught feature learning for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2693–2705. [Google Scholar] [CrossRef]
  11. Zou, H.; Hastie, T.; Tibshirani, R. Sparse principal component analysis. J. Comput. Graph. Stat. 2006, 15, 265–286. [Google Scholar] [CrossRef] [Green Version]
  12. Zhang, Y.; Wu, J.; Cai, Z.; Yu, P. Multi-view Multi-label Learning with Sparse Feature Selection for Image Annotation. IEEE Trans. Multimed. 2020, 1–14. [Google Scholar] [CrossRef]
  13. Zhang, Y.; Wu, J.; Cai, Z.; Du, B.; Yu, P. An unsupervised parameter learning model for RVFL neural network. Neural Netw. 2019, 112, 85–97. [Google Scholar] [CrossRef] [PubMed]
  14. Agarwal, A.; El-Ghazawi, T.; El-Askary, H.; Le-Moigne, J. Efficient hierarchical-PCA dimension reduction for hyperspectral imagery. In Proceedings of the 2007 IEEE International Symposium on Signal Processing and Information Technology, Giza, Egypt, 15–18 December 2007; pp. 353–356. [Google Scholar]
  15. Cheng, X.; Chen, Y.; Tao, Y.; Wang, C.; Kim, M.; Lefcourt, A. A novel integrated PCA and FLD method on hyperspectral image feature extraction for cucumber chilling damage inspection. Trans. ASABE 2004, 47, 1313. [Google Scholar] [CrossRef]
  16. Lazcano, R.; Madroñal, D.; Fabelo, H.; Ortega, S.; Salvador, R.; Callico, G.; Juarez, E.; Sanz, C. Adaptation of an iterative PCA to a manycore architecture for hyperspectral image processing. IET Signal Process. 2019, 91, 759–771. [Google Scholar] [CrossRef]
  17. Hinton, G.E.; Salakhutdinov, R.R. Reducing the dimensionality of data with neural networks. Science 2006, 313, 504–507. [Google Scholar] [CrossRef] [Green Version]
  18. Lin, Z.H.; Chen, Y.S.; Zhao, X.; Wang, G. Spectral-spatial classification of hyperspectral image using autoencoders. In Proceedings of the 2013 9th International Conference on Information, Communications and Signal Processing (ICICS), Tainan, Taiwan, 10–13 December 2013; pp. 1–5. [Google Scholar]
  19. Windrim, L.; Ramakrishnan, R.; Melkumyan, A.; Murphy, R.J.; Chlingaryan, A. Unsupervised feature-learning for hyperspectral data with autoencoders. Remote Sens. 2019, 11, 864. [Google Scholar] [CrossRef] [Green Version]
  20. Koda, S.; Melgani, F.; Nishii, R. Unsupervised spectral-spatial feature extraction with generalized autoencoder for hyperspectral imagery. IEEE Geosci. Remote Sens. Lett. 2019, 1–5. [Google Scholar] [CrossRef]
  21. Liao, Y.; Wang, Y.; Liu, Y. Graph regularized auto-encoders for image representation. IEEE Trans. Image Process. 2016, 26, 2839–2852. [Google Scholar] [CrossRef]
  22. Liang, M.; Jiao, L.; Meng, Z. A superpixel-based relational auto-encoder for feature extraction of hyperspectral images. Remote Sens. 2019, 11, 2454. [Google Scholar] [CrossRef] [Green Version]
  23. He, K.; Zhang, X.; Ren, S.; Sun, J. Deep residual learning for image recognition. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Lasvegas, NV, USA, 26 June–1 July 2016; pp. 770–778. [Google Scholar]
  24. Tao, C.; Pan, H.B.; Li, Y.S.; Zou, Z.R. Unsupervised spectral-spatial feature learning with stacked sparse autoencoder for hyperspectral imagery classification. IEEE Geosci. Remote Sens. Lett. 2015, 12, 2438–2442. [Google Scholar]
  25. Huang, G.B.; Zhu, Q.Y.; Siew, C.K. Extreme learning machine: a new learning scheme of feedforward neural networks. In Proceedings of the 2004 IEEE International Joint Conference on Neural Networks, Budapest, Hungary, 25–29 July 2004; pp. 985–990. [Google Scholar]
  26. Cai, Y.; Liu, X.; Zhang, Y.; Cai, Z. Hierarchical ensemble of extreme learning machine. Pattern Recognit. Lett. 2018, 116, 101–106. [Google Scholar] [CrossRef]
  27. Zhang, Y.; Wu, J.; Zhou, C.; Cai, Z.; Yang, J.; Yu, P. Multi-View Fusion with Extreme Learning Machine for Clustering. ACM Trans. Intell. Syst. Technol. 2019, 10, 1–23. [Google Scholar] [CrossRef] [Green Version]
  28. Han, M.; Liu, B. Ensemble of extreme learning machine for remote sensing image classification. Neurocomputing 2015, 149, 65–70. [Google Scholar] [CrossRef]
  29. Lv, Q.; Niu, X.; Dou, Y.; Xu, J.; Lei, Y. Classification of hyperspectral remote sensing image using hierarchical local-receptive-field-based extreme learning machine. IEEE Geosci. Remote Sens. Lett. 2016, 13, 434–438. [Google Scholar] [CrossRef]
  30. Zhou, Y.; Lian, J.; Han, M. Remote sensing image transfer classification based on weighted extreme learning machine. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1405–1409. [Google Scholar] [CrossRef]
  31. Zhou, Y.; Peng, J.; Chen, C.L.P. Extreme learning machine with composite kernels for hyperspectral image classification. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2015, 8, 2351–2360. [Google Scholar] [CrossRef]
  32. Zhang, Y.; Jiang, X.; Wang, X.; Cai, Z. Spectral-Spatial Hyperspectral Image Classification with Superpixel Pattern and Extreme Learning Machine. Remote Sens. 2019, 11, 1983. [Google Scholar] [CrossRef] [Green Version]
  33. Kasun, L.L.C.; Zhou, H.; Huang, G.B.; Vong, C.M. Representational learning with ELMs for big data. IEEE Intell. Syst. 2013, 28, 31–34. [Google Scholar]
  34. Tang, J.; Deng, C.; Huang, G.B. Extreme learning machine for multilayer perceptron. IEEE Trans. Neural Netw. Learn Syst. 2016, 27, 809–821. [Google Scholar] [CrossRef]
  35. Lv, F.; Han, M.; Qiu, T. Remote sensing image classification based on ensemble extreme learning machine with stacked autoencoder. IEEE Access 2017, 5, 9021–9031. [Google Scholar] [CrossRef]
  36. Ahmad, M.; Khan, A.M.; Mazzara, M.; Distefano, S. Multi-layer extreme learning machine-based autoencoder for hyperspectral image classification. In Proceedings of the 14th International Conference on Computer Vision Theory and Applications (VISAPP’19), Prague, Czech Republic, 25–27 February 2019; pp. 25–27. [Google Scholar]
  37. Kasun, L.L.C.; Yang, Y.; Huang, G.B.; Zhang, Z. Dimension reduction with extreme learning machine. IEEE Trans. Image Process. 2016, 25, 3906–3918. [Google Scholar] [CrossRef] [PubMed]
  38. Li, P.; Hastie, T.J.; Church, K.W. Very sparse random projections. In Proceedings of the Twelfth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Philadelphia, PA, USA, 20–23 August 2006; pp. 287–296. [Google Scholar]
  39. Luo, X.; Xu, Y.; Wang, W.; Yuan, M.; Ban, X.; Zhu, Y.; Zhao, W. Towards enhancing stacked extreme learning machine with sparse autoencoder by correntropy. J. Franklin Inst. 2018, 355, 1945–1966. [Google Scholar] [CrossRef]
  40. Huang, G.; Huang, G.B.; Song, S.; You, K. Trends in extreme learning machines: A review. Neural Netw. 2015, 61, 32–48. [Google Scholar] [CrossRef] [PubMed]
  41. Huang, G.B. An insight into extreme learning machines: Random neurons, random features and kernels. Cognit. Comput. 2014, 6, 376–390. [Google Scholar] [CrossRef]
  42. Deb, K.; Pratap, A.; Agarwal, S.; Meyarivan, T. A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans. Evol. Comput. 2002, 6, 182–197. [Google Scholar] [CrossRef] [Green Version]
  43. Zhang, Q.; Li, H. MOEA/D: A multiobjective evolutionary algorithm based on decomposition. IEEE Trans. Evol. Comput. 2007, 11, 712–731. [Google Scholar] [CrossRef]
  44. Coello, C.A.; Pulido, G.T.; Lechuga, M.S. Handling multiple objectives with particle swarm optimization. IEEE Trans. Evol. Comput. 2004, 8, 256–279. [Google Scholar] [CrossRef]
  45. Chollet, F. Keras. Available online: https://github.com/fchollet/keras (accessed on 6 November 2019).
  46. Hurley, N.; Rickard, S. Comparing measures of sparsity. IEEE Trans. Inf. Theory 2009, 55, 4723–4741. [Google Scholar] [CrossRef] [Green Version]
Figure 1. (a) ELM-AE includes NPR and LR. Using β T as the transformation matrix to transform features. (b) AE consists of an encoder (red rhomboid box) and a decoder (green rhomboid box). The outputs of encoder represent the learned features.
Figure 1. (a) ELM-AE includes NPR and LR. Using β T as the transformation matrix to transform features. (b) AE consists of an encoder (red rhomboid box) and a decoder (green rhomboid box). The outputs of encoder represent the learned features.
Sensors 20 01262 g001
Figure 2. Structure of EMO-ELM which consists of an encoder (red rhomboid box) and decoder (green rhomboid box). Where the red neurons in hidden layer denote they are activated while the orange neurons represent they are limited.
Figure 2. Structure of EMO-ELM which consists of an encoder (red rhomboid box) and decoder (green rhomboid box). Where the red neurons in hidden layer denote they are activated while the orange neurons represent they are limited.
Sensors 20 01262 g002
Figure 3. Pseudo-color image (a) and ground truth (b) of Salinas-A data set.
Figure 3. Pseudo-color image (a) and ground truth (b) of Salinas-A data set.
Sensors 20 01262 g003
Figure 4. Pseudo-color image (a) and ground truth (b) of KSC data set.
Figure 4. Pseudo-color image (a) and ground truth (b) of KSC data set.
Sensors 20 01262 g004
Figure 5. Normalized Pareto front and solution selection of (a) SalinasA and (b) KSC data sets. The curvatures are normalized for plotting it in a same coordinate. The best compromise is denoted as the top three points of closing to the maximum curvature.
Figure 5. Normalized Pareto front and solution selection of (a) SalinasA and (b) KSC data sets. The curvatures are normalized for plotting it in a same coordinate. The best compromise is denoted as the top three points of closing to the maximum curvature.
Sensors 20 01262 g005
Figure 6. The 2-dimensional visualization of Iris dataset of (a) NRP, (b) SPCA, (c) ELM-AE, (d) SELM-AE, (e) AE, (f) SAE, (g) EMO-ELM( f 1 ), (h) EMO-ELM( f 2 ) and (i) EMO-ELM(best).
Figure 6. The 2-dimensional visualization of Iris dataset of (a) NRP, (b) SPCA, (c) ELM-AE, (d) SELM-AE, (e) AE, (f) SAE, (g) EMO-ELM( f 1 ), (h) EMO-ELM( f 2 ) and (i) EMO-ELM(best).
Sensors 20 01262 g006
Figure 7. The sparsity of different algorithms of (a) SalinasA and (b) KSC data set.
Figure 7. The sparsity of different algorithms of (a) SalinasA and (b) KSC data set.
Sensors 20 01262 g007
Figure 8. Box plot of SalinasA and KSC data sets. (ac) denotes the box plot of the SalinasA data set in terms of OA, AA, and Kappa, respectively; (df) represents the box plot of the KSC data set with respect to OA, AA, and Kappa, respectively. The edges of boxes are the 25th and 75th percentiles and the middle lines indicate the median line. Whiskers extend to the maximum and minimum points. Abnormal outliers are shown as “∘”s.
Figure 8. Box plot of SalinasA and KSC data sets. (ac) denotes the box plot of the SalinasA data set in terms of OA, AA, and Kappa, respectively; (df) represents the box plot of the KSC data set with respect to OA, AA, and Kappa, respectively. The edges of boxes are the 25th and 75th percentiles and the middle lines indicate the median line. Whiskers extend to the maximum and minimum points. Abnormal outliers are shown as “∘”s.
Sensors 20 01262 g008
Table 1. Groundtruth classes of the Salinas-A scene and their respective samples number.
Table 1. Groundtruth classes of the Salinas-A scene and their respective samples number.
#ClassNumber of Samples
1Brocoli_green_weeds_1391
2Corn_senesced_green_weeds1343
3Lettuce_romaine_4wk616
4Lettuce_romaine_5wk1525
5Lettuce_romaine_6wk674
6Lettuce_romaine_7wk799
Table 2. Groundtruth classes of the KSC scene and their respective samples number.
Table 2. Groundtruth classes of the KSC scene and their respective samples number.
#ClassNumber of Samples
1Scrub761
2Willow swamp243
3CP/Oak256
4Slash pine252
5Oak/Broadleaf161
6Hardwood229
7swamp105
8Graminoid marsh431
9Spartina marsh520
10Cattail marsh404
11Salt marsh419
12Mud flats503
13Water927
Table 3. The performance comparison of SalinasA for NRP, SPCA, ELM-AE, SELM-AE, AE, SAE and EMO-ELM. (MEAN ± STD) L = 10 , i t e r = 5000 , ρ = 0.05 , N P = 50 .
Table 3. The performance comparison of SalinasA for NRP, SPCA, ELM-AE, SELM-AE, AE, SAE and EMO-ELM. (MEAN ± STD) L = 10 , i t e r = 5000 , ρ = 0.05 , N P = 50 .
ClassAlgorithm
NRPSPCAELM-AESELM-AEAESAEEMO-ELM( f 1 )EMO-ELM( f 2 )EMO-ELM(best)
199.49 ± 0.7299.49 ± 0.7299.16 ± 0.9699.16 ± 0.9499.49 ± 0.7299.49 ± 0.7299.49 ± 0.7299.49 ± 0.7299.49 ± 0.72
297.80 ± 0.4498.73 ± 0.4195.95 ± 2.2898.81 ± 0.4197.83 ± 0.5897.57 ± 0.1698.18 ± 0.6499.09 ± 0.4398.73 ± 0.40
396.12 ± 1.3792.03 ± 1.9686.54 ± 6.2296.36 ± 1.5997.50 ± 0.8096.64 ± 0.7496.14 ± 0.4396.43 ± 0.8392.94 ± 0.64
499.93 ± 0.0999.8 ± 0.1699.13 ± 1.9199.72 ± 0.26100.00 ± 0.0099.87 ± 0.0999.99 ± 0.04100.00 ± 0.00100.00 ± 0.00
5100.00 ± 0.0099.87 ± 0.299.44 ± 0.5799.54 ± 0.5799.70 ± 0.4299.70 ± 0.4299.99 ± 0.0899.70 ± 0.4299.81 ± 0.22
699.37 ± 0.4797.12 ± 1.298.84 ± 0.4499.20 ± 0.4299.50 ± 0.3599.25 ± 0.3198.62 ± 0.3498.89 ± 0.3098.75 ± 0.71
AA98.79 ± 0.2397.84 ± 0.2796.51 ± 1.2998.80 ± 0.3099.00 ± 0.0598.75 ± 0.0798.73 ± 0.2298.93 ± 0.0898.29 ± 0.07
OA98.85 ± 0.1698.22 ± 0.2896.88 ± 1.2898.96 ± 0.2399.02 ± 0.1498.78 ± 0.0498.85 ± 0.2399.12 ± 0.1298.62 ± 0.09
Kappa0.986 ± 0.0020.978 ± 0.0040.961 ± 0.0160.987 ± 0.0030.988 ± 0.0020.985 ± 0.0000.986 ± 0.0030.989 ± 0.0020.983 ± 0.001
Table 4. The performance comparison of KSC for NRP, SPCA, ELM-AE, SELM-AE, AE, SAE and EMO-ELM. (MEAN ± STD) L = 10 , i t e r = 5000 , ρ = 0.05 , N P = 50 .
Table 4. The performance comparison of KSC for NRP, SPCA, ELM-AE, SELM-AE, AE, SAE and EMO-ELM. (MEAN ± STD) L = 10 , i t e r = 5000 , ρ = 0.05 , N P = 50 .
ClassAlgorithm
NRPSPCAELM-AESELM-AEAESAEEMO-ELM( f 1 )EMO-ELM( f 2 )EMO-ELM(best)
197.76 ± 1.3297.66 ± 0.6696.57 ± 1.0396.36 ± 1.1097.83 ± 0.2797.84 ± 0.4995.11 ± 0.2798.09 ± 0.5096.19 ± 0.48
289.59 ± 3.8192.10 ± 2.4683.50 ± 3.6385.72 ± 3.6986.87 ± 2.6086.71 ± 1.9394.40 ± 2.0889.09 ± 2.8790.78 ± 0.83
388.32 ± 1.4193.73 ± 3.3890.12 ± 3.1489.80 ± 3.2092.13 ± 3.8290.02 ± 4.4492.97 ± 0.0488.07 ± 3.5991.68 ± 1.66
463.53 ± 1.4231.15 ± 1.7957.46 ± 5.5756.19 ± 4.2747.46 ± 2.0651.43 ± 2.122.02 ± 0.5567.74 ± 3.2626.71 ± 5.96
555.58 ± 6.5951.84 ± 4.3443.50 ± 4.7243.99 ± 3.7358.15 ± 2.1357.89 ± 2.8116.92 ± 2.8358.68 ± 5.9154.23 ± 3.55
654.09 ± 5.9129.48 ± 2.3646.64 ± 7.2448.78 ± 6.8852.04 ± 5.9847.96 ± 2.630.09 ± 0.3350.41 ± 3.8635.42 ± 3.37
790.48 ± 5.3961.62 ± 10.3282.29 ± 7.7580.19 ± 9.6786.10 ± 5.4686.67 ± 8.3241.05 ± 11.5887.81 ± 7.6377.81 ± 12.64
884.32 ± 4.9681.24 ± 5.9682.53 ± 6.4383.67 ± 6.6688.82 ± 6.1389.89 ± 4.2970.79 ± 3.4088.59 ± 4.0884.92 ± 6.83
997.06 ± 0.2792.48 ± 0.9996.62 ± 2.2497.50 ± 1.6997.56 ± 1.1097.50 ± 1.3279.76 ± 4.5798.81 ± 1.2797.94 ± 0.91
1092.73 ± 1.2897.85 ± 1.6091.24 ± 2.8890.55 ± 2.5893.77 ± 2.3896.49 ± 2.3077.37 ± 1.4194.66 ± 1.8590.65 ± 1.87
1198.59 ± 0.9897.66 ± 0.3694.41 ± 1.6995.49 ± 1.8198.83 ± 0.3499.16 ± 0.2783.60 ± 3.6298.73 ± 0.3693.89 ± 0.45
1284.97 ± 1.5395.63 ± 0.7978.85 ± 2.9881.63 ± 2.6894.69 ± 0.9696.28 ± 0.9085.39 ± 1.0594.77 ± 1.7390.96 ± 1.74
1399.81 ± 0.16100.00 ± 0.0098.34 ± 0.7397.93 ± 0.59100.00 ± 0.00100.00 ± 0.0099.40 ± 0.2799.61 ± 0.1999.32 ± 0.19
AA84.37 ± 1.1778.65 ± 0.3880.16 ± 0.9280.60 ± 1.1184.17 ± 1.0184.45 ± 0.8564.53 ± 0.2385.77 ± 0.8979.27 ± 0.73
OA89.51 ± 0.5987.21 ± 0.6086.49 ± 0.6586.96 ± 0.7290.20 ± 0.7090.59 ± 0.5676.77 ± 0.6791.21 ± 0.4786.70 ± 0.76
Kappa0.883 ± 0.0070.857 ± 0.0070.849 ± 0.0070.855 ± 0.0080.891 ± 0.0080.895 ± 0.0060.739 ± 0.0080.902 ± 0.0050.851 ± 0.008

Share and Cite

MDPI and ACS Style

Fang, X.; Cai, Y.; Cai, Z.; Jiang, X.; Chen, Z. Sparse Feature Learning of Hyperspectral Imagery via Multiobjective-Based Extreme Learning Machine. Sensors 2020, 20, 1262. https://doi.org/10.3390/s20051262

AMA Style

Fang X, Cai Y, Cai Z, Jiang X, Chen Z. Sparse Feature Learning of Hyperspectral Imagery via Multiobjective-Based Extreme Learning Machine. Sensors. 2020; 20(5):1262. https://doi.org/10.3390/s20051262

Chicago/Turabian Style

Fang, Xiaoping, Yaoming Cai, Zhihua Cai, Xinwei Jiang, and Zhikun Chen. 2020. "Sparse Feature Learning of Hyperspectral Imagery via Multiobjective-Based Extreme Learning Machine" Sensors 20, no. 5: 1262. https://doi.org/10.3390/s20051262

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop