Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Robust Multiplexing Waveform Design for Integrated OFDM Radar and Communication via Complex Weight Optimization
Next Article in Special Issue
CD-MQANet: Enhancing Multi-Objective Semantic Segmentation of Remote Sensing Images through Channel Creation and Dual-Path Encoding
Previous Article in Journal
The Influence of Visual Landscapes on Road Traffic Safety: An Assessment Using Remote Sensing and Deep Learning
Previous Article in Special Issue
MCSNet: A Radio Frequency Interference Suppression Network for Spaceborne SAR Images via Multi-Dimensional Feature Transform
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

A Hybrid 3D–2D Feature Hierarchy CNN with Focal Loss for Hyperspectral Image Classification

1
School of Computer Science and Information Engineering, Harbin Normal University, Harbin 150025, China
2
Key Laboratory of Intelligent Education and Information Engineering, Heilongjiang Universities, Harbin 150025, China
3
Department of Municipal and Environmental Engineering, Heilongjiang Institute of Construction Technology, Harbin 150025, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2023, 15(18), 4439; https://doi.org/10.3390/rs15184439
Submission received: 1 August 2023 / Revised: 3 September 2023 / Accepted: 6 September 2023 / Published: 9 September 2023
Graphical abstract
">
Figure 1
<p>Overview of the proposed feature extractor FE-HybridSN and entire framework FEHN-FL. Note that batch normalization (BN) and rectified linear unit (ReLU), following every convolutional operation, are omitted in the figure.</p> ">
Figure 2
<p>Classification accuracies (%) by different methods with 10% training samples and different spatial sizes over SV dataset (the left one is <math display="inline"><semantics> <mrow> <mn>11</mn> <mo>×</mo> <mn>11</mn> </mrow> </semantics></math>, another one is <math display="inline"><semantics> <mrow> <mn>23</mn> <mo>×</mo> <mn>23</mn> </mrow> </semantics></math>).</p> ">
Figure 3
<p>Classification maps for the IP dataset with 15% labeled training samples and <math display="inline"><semantics> <mrow> <mn>19</mn> <mo>×</mo> <mn>19</mn> </mrow> </semantics></math> spatial window. (<b>a</b>) Ground truth. (<b>b</b>) 2D-CNN [<a href="#B53-remotesensing-15-04439" class="html-bibr">53</a>]. (<b>c</b>) 3D-CNN [<a href="#B54-remotesensing-15-04439" class="html-bibr">54</a>]. (<b>d</b>) 3D2D-HybridSN [<a href="#B31-remotesensing-15-04439" class="html-bibr">31</a>]. (<b>e</b>) M3D-CNN [<a href="#B55-remotesensing-15-04439" class="html-bibr">55</a>]. (<b>f</b>) Proposed. In the subfigure (<b>b</b>–<b>f</b>), the right brackets encompass the overall classification accuracies, and the best classfication result is highlighted in bold.</p> ">
Figure 4
<p>Classification maps for the UP dataset with 10% labeled training samples and <math display="inline"><semantics> <mrow> <mn>19</mn> <mo>×</mo> <mn>19</mn> </mrow> </semantics></math> spatial window. (<b>a</b>) Ground truth. (<b>b</b>) 2D-CNN [<a href="#B53-remotesensing-15-04439" class="html-bibr">53</a>]. (<b>c</b>) 3D-CNN [<a href="#B54-remotesensing-15-04439" class="html-bibr">54</a>]. (<b>d</b>) 3D2D-HybridSN [<a href="#B31-remotesensing-15-04439" class="html-bibr">31</a>]. (<b>e</b>) M3D-CNN [<a href="#B55-remotesensing-15-04439" class="html-bibr">55</a>]. (<b>f</b>) Proposed. In the subfigure (<b>b</b>–<b>f</b>), the right brackets encompass the overall classification accuracies, and the best classification result is highlighted in bold.</p> ">
Figure 5
<p>Classification maps for the SV dataset with 5% labeled training samples and <math display="inline"><semantics> <mrow> <mn>19</mn> <mo>×</mo> <mn>19</mn> </mrow> </semantics></math> spatial window. (<b>a</b>) Ground truth. (<b>b</b>) 2D-CNN [<a href="#B53-remotesensing-15-04439" class="html-bibr">53</a>]. (<b>c</b>) 3D-CNN [<a href="#B54-remotesensing-15-04439" class="html-bibr">54</a>]. (<b>d</b>) 3D2D-HybridSN [<a href="#B31-remotesensing-15-04439" class="html-bibr">31</a>]. (<b>e</b>) M3D-CNN [<a href="#B55-remotesensing-15-04439" class="html-bibr">55</a>]. (<b>f</b>) Proposed. In the subfigure (<b>b</b>–<b>f</b>), the right brackets encompass the overall classification accuracies, and the best classification result is highlighted in bold.</p> ">
Figure 5 Cont.
<p>Classification maps for the SV dataset with 5% labeled training samples and <math display="inline"><semantics> <mrow> <mn>19</mn> <mo>×</mo> <mn>19</mn> </mrow> </semantics></math> spatial window. (<b>a</b>) Ground truth. (<b>b</b>) 2D-CNN [<a href="#B53-remotesensing-15-04439" class="html-bibr">53</a>]. (<b>c</b>) 3D-CNN [<a href="#B54-remotesensing-15-04439" class="html-bibr">54</a>]. (<b>d</b>) 3D2D-HybridSN [<a href="#B31-remotesensing-15-04439" class="html-bibr">31</a>]. (<b>e</b>) M3D-CNN [<a href="#B55-remotesensing-15-04439" class="html-bibr">55</a>]. (<b>f</b>) Proposed. In the subfigure (<b>b</b>–<b>f</b>), the right brackets encompass the overall classification accuracies, and the best classification result is highlighted in bold.</p> ">
Review Reports Versions Notes

Abstract

:
Hyperspectral image (HSI) classification has been extensively applied for analyzing remotely sensed images. HSI data consist of multiple bands that provide abundant spatial information. Convolutional neural networks (CNNs) have emerged as powerful deep learning methods for processing visual data. In recent work, CNNs have shown impressive results in HSI classification. In this paper, we propose a hierarchical neural network architecture called feature extraction with hybrid spectral CNN (FE-HybridSN) to extract superior spectral–spatial features. FE-HybridSN effectively captures more spectral–spatial information while reducing computational complexity. Considering the prevalent issue of class imbalance in experimental datasets (IP, UP, SV) and real-world hyperspectral datasets, we apply the focal loss to mitigate these problems. The focal loss reconstructs the loss function and facilitates effective achievement of the aforementioned goals. We propose a framework (FEHN-FL) that combines FE-HybridSN and the focal loss for HSI classification and then conduct extensive HSI classification experiments using three remote sensing datasets: Indian Pines (IP), University of Pavia (UP), and Salinas Scene (SV). Using cross-entropy loss as a baseline, we assess the hyperspectral classification performance of various backbone networks and examine the influence of different spatial sizes on classification accuracy. After incorporating focal loss as our loss function, we not only compare the classification performance of the FE-HybridSN backbone network under different loss functions but also evaluate their convergence rates during training. The proposed classification framework demonstrates satisfactory performance compared to state-of-the-art end-to-end deep-learning-based methods, such as 2D-CNN, 3D-CNN, etc.

Graphical Abstract">

Graphical Abstract

1. Introduction

Hyperspectral images (HSIs) are generated by imaging spectrometers mounted on various space platforms, capturing spatial–spectral information [1]. The advancement of hyperspectral imaging technology has enabled sensors to capture hundreds of continuous spectral bands with nanometer-level resolution [2,3]. As a result, HSIs have found numerous applications in diverse fields, including environmental monitoring [4,5,6], hyperspectral anomaly detection [7], and hyperspectral image classification [8]. HSI classification, in particular, serves as a fundamental technique in many hyperspectral remote sensing applications, and it has proven to be invaluable in precision agriculture [9], geological exploration [10], and other domains [11,12].
The task of HSI classification is mainly tackled by two schemes: one with a hand-crafted feature extraction method and another with learning-based feature extraction method. In the early phase of HSI classification, the strategy of extracting more spectral or spatial features is conventional machine learning. For instance, Yang and Qian [13] introduced a novel approach for hyperspectral image classification called multiscale joint collaborative representation with a locally adaptive dictionary. It constrains the adverse impact to HSI classification from useless pixels. Camps-Valls et al. [14] proposed and validated the effectiveness of composite kernels, a novel technique that combines multiple kernel functions to improve hyperspectral image classification performance. Other hand-crafted approaches, one after another, were proposed. One prominent method is the joint sparse model and discontinuity preserving relaxation [15]. The approach preprocesses each pixel and calculates relevant statistical measures, aiming at elegantly integrating spatial context and spectral signatures. Similarly, sparse self-representation [16] addresses band selection using optimization-based sparse self-representation, optimizing for efficient feature selection. To improve classification accuracy, researchers have proposed fusing correlation coefficient and sparse representation [17], aiming to harness the strengths of both methods. Additionally, multiscale superpixels and guided filter [18] have been explored for sparse-representation-based hyperspectral image classification, promising effective feature extraction and classification. The Boltzmann-entropy-based unsupervised band selection [19] has been investigated, targeting informative band selection to enhance classification performance.
However, the classification processes, based on the abovementioned approaches, are relatively cumbersome due to that their accomplishments rely on manually extracted features. Furthermore, faced with the inherent high-dimensional complexity of HSIs, the researchers find it hard to obtain ideal classification just using the above approaches, especially in challenging scenes [20]. Regrettably, the lack of labeled samples in the HSI field is in sharp contrast to the richness of spectral data. This fact poses the challenge of learning better feature representation and is prone to overfitting of methods. In view of the above problems, some schemes have been proposed to alleviate them, mainly including feature extraction [21,22,23], dimension reduction [24,25], and data augmentation [26].
Deep learning has emerged as a powerful method for feature extraction, enabling the identification of features from hyperspectral images (HSIs). Among various deep learning models, the convolutional neural network (CNN) stands out as one of the most widely applied models to address HSI classification challenges. CNNs have shown remarkable performance gains over conventional hand-designed features. CNN models are capable of processing spatial HSI patches as data inputs, leading to the development of progressive CNN-based methods that leverage both spectral and spatial features. For instance, Mei et al. [27] proposed a CNN model that adopts a pixel-wise approach and involves preprocessing each pixel by calculating the mean and standard deviation of the pixel neighborhood for each spectral band. On the other hand, Paoletti et al. [28] and Li et al. [29] introduced two distinct CNN models—one for extracting spatial features and the other for extracting spectral features. These models utilize a softmax classifier to achieve desirable classification results. While CNN-based technologies can effectively extract spatial and spectral features for HSI classification and other applications, they still encounter challenges in effectively utilizing information related to spatial and spectral associations.
In contrast, some technologies, which learn spatial and spectral features simultaneously for HSI classification, have been proposed. Yang et al. [30] proposed a multiscale wavelet 3D-CNN (MW-3D-CNN). Great importance was attached to the relationship information amid adjacent channels, while the corresponding model’s calculating complexity was augmented intensively. Accordingly, Roy et al. [31] proposed the 3D–2D CNN feature hierarchy model for HSI classification. On one hand, a few 3D-CNN layers were utilized to extract spectral information amid spectral bands. On the other hand, 2D-CNN layers concentrated on much spatial texture and context. Liu et al. [32] extended the CNN model by incorporating the attention mechanism to enhance feature extraction from HSI. More recently, Zhong et al. [33] introduced the spectral–spatial residual network (SSRN). In the SSRN’s designed residual blocks, identity mapping is utilized to connect 3D convolutional layers. These innovative models and technologies address the limitations of previous approaches and effectively leverage spectral and spatial features to significantly improve the performance of HSI classification.
In recent times, large language models such as Transformer [34] have become a new paradigm in the field of natural language processing (NLP). Transformer introduces the attention mechanism, which allows for interactions between different tokens, capturing long-range semantic dependencies. Inspired by this success, researchers have extended these methods to the field of computer vision (CV). Similar to Transformer in NLP, Vision Transformer (ViT) [35] is typically pretrained on unlabeled image streams or video streams and then fine-tuned on downstream tasks to train model parameters. In the domain of remote sensing imagery, more researchers are adopting ViT or its variants [36] as foundational models for their studies. Liu et al. [37] utilized a customized Swin Transformer to reduce computational complexity and obtained strong generalization. Ayas and Tunc-Gormus [38] introduced a novel spectral Swin Transformer (SpectralSWIN) network, but without using attention mechanism, to hierarchically fuse spatial–spectral features and achieve a significant superiority. Zhao et al. [39] proposed a spectral–spatial axial aggregation Transformer framework to perform multiscale feature extraction and fusion on the input data while utilizing spectral shift operations to ensure information aggregation and feature extraction across different spectral components.
With regard to the issue of HSI classification, currently, many loss functions have been designed and obtained stunning performance. Multiclass hinge (MCH) loss in support vector machine (SVM), as one of the traditional classification strategies, is the main solution to HSI classification in the early phase. Wang et al. [40] introduced a novel classification framework, incorporating spatial, spectral, and hierarchical structure features, for hyperspectral images. The approach involves integrating three different and important types of information into the SVM classifier. By leveraging this joint integration, the proposed framework intends to enhance the classification performance of hyperspectral images. Furthermore, cross-entropy loss, as a standard metric, is widely applied to complete different classification tasks in hyperspectral image recognition, such as [41,42]. Regrettably, common and publicly available datasets Indian Pines (IP), Salinas (SA) and University of Pavia (UP), as well as the overwhelming majority of other hyperspectral imagery datasets, are class-imbalanced [43,44] (See Section 2.2). This fact poses great challenges for deep-learning-based HSI classification models regarding how to handle the class imbalance problem [45].
In this paper, spatial–spectral feature obtained by the deep learning approach is proposed for the HSI classification task. First, to extract more vital HSI features, including spatial and spectral information, we propose a novel convolutional neural network (FE-HybridSN), which pays attention to the correlation of adjacent bands from HSI. Compared with the plain 3D-CNN model for HSI classification, the FE-HybridSN alleviates the burden from complex computation. The proposed framework results in better HSI classification performance when the 3D-CNN model and the FE-HybridSN have similar scale in model structure. Second, to cope with the class-imbalanced problem in HSI classification, we apply the focal loss as loss function. In fact, we obtain expected classification results based on the focal loss after much rigorous research. In summary, the main contributions of this paper are as follows:
  • We fashion a novel five-layer FE-HybridSN of hyperspectral images for mining spatial context and spectral features;
  • We apply the focal loss as the loss function to alleviate the class-imbalanced problem in the HSI classification task;
  • We explore feature learning and classification of hyperspectral images using systematic experiments, and inspire new deep learning ideas for hyperspectral applications.
The paper is structured as follows. Section 2 outlines the challenges and presents our proposed approach to address them. In Section 3, we provide a detailed description of the experimental background and present the fundamental experimental configurations. Additionally, we compare our method with other state-of-the-art approaches. Building upon the results presented in Section 3, Section 4 offers a comprehensive discussion to provide insights and interpretations. Finally, in Section 5, we summarize the findings of the paper and propose promising research directions for future exploration.

2. Methodology

We present a novel joint network called FE-HybridSN, which combines a hybrid 3D–2D CNN architecture with the focal loss to achieve hyperspectral image classification. The proposed approach is designed to capitalize on spectral–spatial feature maps and extract more abstract representations from hierarchical space, as depicted in Figure 1. By leveraging the hybrid 3D–2D CNN model, we aim to effectively mine valuable information from the hyperspectral data. Additionally, we utilize the focal loss to mitigate any adverse effects on hyperspectral image classification, as illustrated in Figure 1. The focal loss helps in handling challenging samples, thereby enhancing the overall performance of the classification process.

2.1. Proposed Model

Assume the spatial–spectral hyperspectral data cube is denoted by I o r i g i n R H × W × B , where H, W, and B represent the height, width, and the number of spectral bands in the hyperspectral image (HSI), respectively. Each pixel in I o r i g i n is associated with a one-hot label vector Y = { y 1 , y 2 , , y C } R 1 × 1 × C , where C signifies the number of land cover classes. However, a real-world challenge arises from the fact that high-dimensional hyperspectral pixels may exhibit mixtures of multiple land cover classes, leading to substantial spectral intravariability and significant interclass similarity.
To address this issue, we employ principal component analysis (PCA) to preprocess the original HSI data I o r i g i n and eliminate redundant spectral information. PCA reduces the original B spectral bands to K bands while preserving the spatial dimensions, thus retaining the spatial information even after the PCA process. Consequently, we represent the processed spatial–spectral hyperspectral data cube as I P C A R H × W × K . Here, H and W still represent the height and width of the spectral bands, and K indicates the number of retained spectral bands after PCA processing.
During the data preprocessing stage, besides reducing the dimensionality of the hyperspectral data using PCA, it is also necessary to segment the image into small, overlapping 3D blocks. The primary purpose of this is to apply our deep learning method on each smaller 3D block. Each adjacent 3D block originates from the PCA-reduced hyperspectral data cube, and they are uniquely identified by the central spatial coordinates. We denote each 3D block as I b l o c k R s × s × K , where s represents the size of each spatial window and K denotes the depth of spectral bands in each block. Notably, K also corresponds to the number of bands retained after PCA dimensionality reduction. Ultimately, the PCA-reduced hyperspectral data cube generates ( H s + 1 ) × ( W s + 1 ) 3D blocks. Specifically, for any given hyperspectral 3D block with central spatial coordinates ( α , β ) , the corresponding spatial window will cover the width range [ α ( s 1 ) / 2 , α + ( s 1 ) / 2 ] and the height range [ β ( s 1 ) / 2 , β + ( s 1 ) / 2 ] .
We propose a framework named FEHN-FL with hierarchical convolutional structure for HSI classification. CNN parameters are trained using supervised methods [46] with gradient descent optimization. Conventional 2D CNNs compute 2D discriminative feature maps by applying convolutions solely over spatial dimensions, but they lack the ability to identify and handle spectral information. In contrast, a single 3D-CNN can slide the convolutional kernel along all three dimensions (height, width, and spectral) and interact with both spatial and spectral dimensions. This capability allows the 3D-CNN to comprehensively capture spatial and spectral information from the high-dimensional hyperspectral data. FE-HybridSN, the backbone of the FEHN-FL, hierarchically combines three 3D convolutions (refer to Equation (2)), two 2D convolutions (refer to Equation (1)), and three fully connected layers, achieving a balanced integration of spectral and spatial information for more effective HSI classification.
In 2D-CNN, the output results are generated through the process of convolution, where the input image is convolved with 2D filters such that the size is predesigned, also known as convolutional kernels. This operation involves element-wise multiplication between the filter’s weights and the pixels in the input image, followed by summation to obtain the new output pixel value. The convolutional process slides the filter across the entire input image, computing the output pixel value at each position. By utilizing distinct filters at different layers, 2D-CNN can effectively learn various features present in the image. These learned features are then combined to form higher-level representations, enabling the model to perform image classification and feature extraction tasks. The resulting features from the convolution are passed through an activation function, introducing nonlinearity to the model. In 2D convolution, the activation value of the jth feature map at spatial position ( x , y ) in the ith layer is represented as v ( i , j ) ( x , y ) and can be expressed by the following equation:
v ( i , j ) ( x , y ) = ψ b i , j + τ = 1 d l 1 ρ = γ γ σ = δ δ w i , j , τ σ , ρ × v i 1 , τ x + σ , y + ρ
where ψ is the nonlinear activation function, b i , j is the bias parameter for the jth feature map of the ith layer, and d τ indicates the number of feature maps in the τ th layer. The size of the predesigned convolutional kernel is ( 2 γ + 1 ) × ( 2 δ + 1 ) . w i , j corresponds to the weight parameter for the jth feature map of the ith layer.
According to the definition of three dimensional convolution [47], we perform convolutional operations by applying 3D convolutional kernels to the hyperspectral images. In the FEHN-FL, the feature maps in the convolutional layer are generated by applying 3D convolutional kernels on discrete or consecutive spectral bands of the input layer, thus capturing spectral information. In 3D convolution, the activation value v ( i , j ) ( x , y , z ) of the jth feature map at spatial position ( x , y , z ) in the ith layer is expressed as follows:
v ( i , j ) ( x , y , z ) = ψ b i , j + τ = 1 d l 1 λ = η η ρ = γ γ σ = δ δ w i , j , τ σ , ρ , λ × v i 1 , τ x + σ , y + ρ , z + λ
where 2 η + 1 is the depth of kernel along the spectral dimension and other parameters are the same as in Equation (1).
In the FE-HybridSN, we design different 3D convolution kernels. We denote the structure of the 3D kernel as L @ K 1 u × K 2 u × K 3 u , where u and L, respectively, represent the layer index of current kernels and the number of output channels for the current convolution layer. Moreover, we denote the structure of the 2D kernel as L @ K 1 d × K 2 d , where d represents the layer index of current kernels, and L is same as the 3D kernel. One of our designed kernel sequence is ( 8 @ 3 × 3 × 7 , 16 @ 3 × 3 × 5 , 32 @ 3 × 3 × 3 , 128 @ 3 × 3 , 256 @ 3 × 3 ). Here, 8 @ 3 × 3 × 7 means that the first 3D kernel size is 3 × 3 × 7 and the number of output channel is eight when a single channel from P is the first input. The explanation of subsequent kernel parameters is the same as 8 @ 3 × 3 × 7 . The proposed model is comprehensively summarized in Table 1, presenting the layer index, output map dimensions, and the corresponding number of parameters.
In the proposed method, the batch normalization (BN) layers are introduced as the important elements since they make the distribution of input data in every layer of the network relatively stable and accelerate the training speed of the entire framework. The formula of BN is as follows:
N ( x ) = x x ¯ D ( x ) + ε · γ + β
where x ¯ is the average of the summation, D ( x ) is the variance, γ and β are learnable parameter vectors, and ε is a parameter for numerical stability. In addition, the nonlinear layer aims at adding some nonlinear features to the network. Then, it is notable that the rectified linear unit (ReLU) [48] is selected into each 3D/2D convolutional layer.

2.2. Focal Loss

The class-imbalanced problem commonly occurs in tasks with a long-tailed data distribution, where a few classes dominate the data, while most classes have very few samples. In traditional classification and visual recognition tasks, the training distribution can be balanced through manual intervention using resampling strategies. This strategy ensures that the number of samples from different classes does not significantly differ. However, as the number of categories increases, maintaining a balance between all categories becomes increasingly challenging, resulting in an exponential growth in the collection of samples.
In the case of HSI classification, neither using resampling strategies nor not using them are feasible or rational solutions. Hence, we employ the focal loss as the loss function. Compared to traditional loss functions like multiclass hinge loss and cross-entropy loss, the focal loss provides better performance and tackles the class imbalance issue effectively.

2.2.1. Balanced Cross Entropy

Introducing weighted factors α and 1 α for positive and negative classes, respectively, is a common approach to address class imbalance issues. In practice, α is often initialized as the inverse class frequency, which is the reciprocal of the ratio of positive class samples to negative class samples. The purpose is to assign larger weights to the classes with fewer samples during the training phase, aiming to balance the class distribution. We denote the α -balanced cross entropy (CE) loss as
C E ( p , y ) = C E ( p t ) = α t log ( p t ) .
where p t is expressed as follows:
p t = p , y = 1 1 p , y = 0
Here, the definition of α t is analogous to how we defined p t , and y { 0 , 1 } indicates negative and positive class, respectively.

2.2.2. Focal Loss Definition

Easily classified negatives comprise the majority of the loss and dominate the gradient [49]. While α balances the importance of positive/negative examples, it does not differentiate between easy and hard examples. The focal loss reconstructs the balanced cross-entropy loss to down-weight easy examples and thus focuses training on hard negatives. The focal loss is defined as follows:
F L ( p t ) = ( 1 p t ) γ log ( p t ) .
Formally, we use an α -balanced variant of the focal loss as the loss function:
F L ( p t ) = α t ( 1 p t ) γ log ( p t ) .

3. Experiments

3.1. Description of Experiment Datasets

Extensive experiments on three benchmark datasets of HSIs, including Indian Pines (IP) [50], University of Pavia (UP) [51], and Salinas Valley (SV) [52], were conducted to assess the classification performance of the proposed framework. Table 2 provides a concise overview of the hyperspectral image (HSI) datasets that were considered in the study. It includes information such as the number of samples available per class in each dataset and the corresponding ground truth (GT) provided.
Table 2. The total number of samples and the sample quantity for each land cover.
Table 2. The total number of samples and the sample quantity for each land cover.
Indian PinesUniversity Of PaviaSalinas
Remotesensing 15 04439 i045Remotesensing 15 04439 i046Remotesensing 15 04439 i047
ColorLand-Cover typeSamplesColorLand-Cover typeSamplesColorLand-Cover typeSamples
Remotesensing 15 04439 i001Background10,776Remotesensing 15 04439 i002Background164,624Remotesensing 15 04439 i003Background56,975
Remotesensing 15 04439 i004Alfalfa46Remotesensing 15 04439 i005Asphalt6631Remotesensing 15 04439 i006Brocoli-green-weeds-12009
Remotesensing 15 04439 i007Corn-notill1428Remotesensing 15 04439 i008Meadows18,649Remotesensing 15 04439 i009Brocoli-green-weeds-23726
Remotesensing 15 04439 i010Corn-min830Remotesensing 15 04439 i011Gravel2099Remotesensing 15 04439 i012Fallow1976
Remotesensing 15 04439 i013Corn237Remotesensing 15 04439 i014Trees3064Remotesensing 15 04439 i015Fallow-rough-plow1394
Remotesensing 15 04439 i016Pasture483Remotesensing 15 04439 i017Painted metal sheets1345Remotesensing 15 04439 i018Fallow-smooth2678
Remotesensing 15 04439 i019Trees730Remotesensing 15 04439 i020Bare Soil5029Remotesensing 15 04439 i021Stubble3959
Remotesensing 15 04439 i022Pasture-mowed28Remotesensing 15 04439 i023Bitumen1330Remotesensing 15 04439 i024Celery3579
Remotesensing 15 04439 i025Hay-windrowed478Remotesensing 15 04439 i026Self-Blocking Bricks3682Remotesensing 15 04439 i027Grapes-untrained11,271
Remotesensing 15 04439 i028Oats20Remotesensing 15 04439 i029Shadows947Remotesensing 15 04439 i030Soil-vineyard-develop6203
Remotesensing 15 04439 i031Soybean-notill972 Remotesensing 15 04439 i032Corn-senesced-green-weeds3278
Remotesensing 15 04439 i033Soybean-min2455Remotesensing 15 04439 i034Lettuce-romaine-4wk1068
Remotesensing 15 04439 i035Soybean-clean593Remotesensing 15 04439 i036Lettuce-romaine-5wk1927
Remotesensing 15 04439 i037Wheat205Remotesensing 15 04439 i038Lettuce-romaine-6wk916
Remotesensing 15 04439 i039Woods1265Remotesensing 15 04439 i040Lettuce-romaine-7wk1070
Remotesensing 15 04439 i041Bldgs-Grass-Trees-Drives386Remotesensing 15 04439 i042Vinyard-untrained7268
Remotesensing 15 04439 i043Stone-Steel towers93Remotesensing 15 04439 i044Vinyard-vertical-trellis1807
Total samples21,025 Total samples207,400 Total samples111,104

3.2. Experimental Configuration

The proposed framework (FEHN-FL) was compared to other methods available in the literature. Concretely, we compared the backbone FE-HybridSN used for feature extraction with the following methods: (1) 2D-CNN [53]; (2) 3D-CNN [54]; (3) M3D-CNN [55]; (4) 3D2D-HybridSN [31]. In addition, we introduced other loss functions as experimental comparison objects, including cross-entropy (CE) loss and multiclass hinge (MCH) loss in support vector machine (SVM), to evaluate the performance of the focal loss. For each method, we carefully fine-tuned and fixed all hyperparameters to their optimal values before conducting the experiments. This ensured that we obtained the best possible performance from each approach to facilitate a fair comparison of the results. The details are shown in Table 3.
Before original data were used the input of framework FEHN-FL, PCA was applied over the hyperspectral data so as to reduce the dimension of the original hyperspectral image and extract principal components. In our experiments, all principal components were set as fifteen. According to the experimental design, the datasets IP, UP, and SV were randomly divided into two parts in proportion: 5%, 10%, and 15% of IP, UP, or SV labeled samples were the training samples, and the remainder were testing datasets. Considering different schemes about backbone and loss function, as well as the aforementioned datasets, we implemented sets of experiments to validate the performance of the FE-HybridSN based on CE loss (Experiment 1). Then, to test the effects rooted from different patch sizes, we correspondingly designed four different spatial sizes for input data (Experiment 2). Furthermore, we considered different training data percentages (Experiment 3) aimed at comparing the FEHN-FL to other state-of-the-art methods. Finally, we compared the focal loss with other loss functions and studied its convergence speed.
  • In the first experiment, the FE-HybridSN was compared with 2D-CNN [53], 3D-CNN [54], M3D-CNN [55], and 3D2D-HybridSN [31] classification methods using a training set accounting for 15% of the whole labeled dataset over IP, UP, and SV datasets. Additionally, we set the input spatial size to be 15 × N × N (N could be 11, 15, 19, or 23) for the 2D-CNN [53], 3D-CNN [54], M3D-CNN [55], 3D2D-HybridSN [31], and FE-HybridSN, with N being the spatial size (i.e., patch size).
  • In the second experiment, the FEHN-FL was compared with the four aforementioned methods. We designed four different patch sizes, i.e., 11 × 11 , 15 × 15 , 19 × 19 , and 23 × 23 . Here, we considered 10% of available labeled data for the IP, UP, and SV datasets.
  • In the third experiment, we compared the proposed framework (FEHN-FL) with all the aforementioned methods [31,53,54,55]. The configurations of patch size were the same as the second experiment. We still focused on the performance just using our backbone (FE-HybridSN) with CE loss.
  • In the final experiment, we compared the convergence speed of different loss functions, including cross-entropy (CE) loss, multiclass hinge (MCH) loss, and focal loss, using training samples of the same scale on the Indian Pines (IP), University of Pavia (UP), and Salinas Valley (SV) datasets.
In our experimental evaluation, we measured the performance using three metrics: OA, AA, and kappa. OA represents the ratio of correctly predicted samples to the total testing samples, providing an overall accuracy measure. AA computes the average classification accuracy across different classes, giving insight into the performance of individual classes. Kappa serves as a statistical measure, indicating the agreement between the ground truth map and the classification map. The experiments were conducted on a hardware setup comprising an Intel Core processor (2.30 GHz), 16 GB of memory, and an NVIDIA GeForce GTX 1050Ti graphics processing unit with 8 GB RAM. To implement the HSI classification methods, we utilized Python in the Pytorch platform; specifically, Pytorch 1.12.0, CUDA 11.7, cuDNN 8.0, and Python 3.8 were the specific research environments used.

3.3. Experimental Results

3.3.1. Experiment 1

Table 4 presents the classification results with 15% training samples for the IP dataset. Specifically, the first column of Table 4 indicates other methods compared with the FE-HybridSN. The next four columns show the results of the OA, AA, and kappa coefficient, which were derived from using different spatial sizes: 11 × 11 , 15 × 15 , 19 × 19 , and 23 × 23 . In addition, Table 5 presents the classification results with 5% training samples for the UP dataset. The explanation for every item in Table 5 is the same as Table 4. Significantly, the last row contains the accuracies of the proposed backbone (i.e., FE-HybridSN).
Furthermore, as shown in Figure 2, we conducted two sets of experiments on the SV dataset based on spatial size. Two abscissae and two ordinates, respectively, indicate the names of different methods and classification accuracies (in percentages). Under the condition using established methods, we compared the classification performance using two different spatial sizes ( 11 × 11 and 23 × 23 ) and 10% training samples. The left one represents the classification result using the spatial size 11 × 11 . Correspondingly, the right one represents the classification result using the spatial size 23 × 23 .

3.3.2. Experiment 2

In the second experiment, we considered 5%, 10%, and 15% of the labeled data over the SV and UP datasets as training samples, and compared the classification performance of the proposed method with those of 2D-CNN [53], 3D-CNN [54], and 3D2D-HybridSN [31]. Table 6 shows the classification accuracies with 5%, 10%, and 15% labeled training samples and 11 × 11 spatial size by the abovementioned scheme.

3.3.3. Experiment 3

In order to comprehensively consider the performance of feature extractors and loss functions for hyperspectral image classification, we compared the proposed framework (FEHN-FL) with other methods. Concretely, spatial size 23 × 23 and three different training ratios (5%, 10%, and 15%) were considered for the SV and UP datasets. Table 7 presents the classification results according to the aforementioned experimental scheme.

3.3.4. Experiment 4

In Table 8, we propose the trainable parameter scales of various backbones. Table 9 demonstrates that when utilizing 10% labeled samples for training and setting the spatial size to 23 × 23 , our proposed framework with the feature extractor FE-HybridSN achieves the minimum training time overhead.

4. Discussion

After analyzing the results of Experiment 1 (refer to Section 3.3.1), several key observations can be made. Firstly, the FE-HybridSN, combined with cross-entropy loss, demonstrates high classification accuracy. When training with 15% and 5% labeled samples using different spatial sizes on the corresponding IP and UP datasets, our method outperforms other state-of-the-art techniques in terms of classification accuracy. Specifically, utilizing the CE loss as the terminal loss function and following the basic experimental configurations, the FE-HybridSN consistently achieves superior pixel-level classification results compared to 2D-CNN [53], 3D-CNN [54], 3D2D-HybridSN [31], and M3D-CNN [55].
In addition, Table 4 and Table 5 highlight the significant impact of spatial size on HSI classification accuracy. Notably, the classification accuracies obtained using 19 × 19 and 23 × 23 patch sizes are consistently lower than those achieved with a 15 × 15 patch size across all methods in Table 4. Similarly, Table 5 illustrates that increasing the patch size leads to improved classification accuracy for the FE-HybridSN framework utilizing cross-entropy loss on the UP dataset. This finding emphasizes the crucial role of spatial size in HSI classification and its ability to adjust the decision boundary. Specifically, a smaller spatial size leads to a greater loss of information, resulting in lower classification accuracy, as evidenced by the experimental results. Furthermore, Figure 2 provides a visual representation of the relationship between spatial size and classification accuracy, further supporting these conclusions.
Based on the results in Table 6 of Experiment 2 (refer to Section 3.3.2), it is evident that our proposed feature extractor, FE-HybridSN, achieves significantly better classification accuracy on both the SV and UP datasets. On one hand, when using training samples with different proportions of labeled data, it is evident that as the number of training samples decreases, the classification accuracy of hyperspectral images also decreases. With the small scale of the training set, the risk of overfitting increases. In other words, when training is completed, the resulting classification decision boundaries may lack robust generalization ability, particularly when facing new samples such as test samples. On the other hand, with the same training ratio, the OA of the SV dataset is noticeably higher compared to the classification accuracy of the UP dataset. This is primarily due to the fact that the SV dataset has a significantly larger training scale than the UP dataset.
In Experiment 3 (refer to Section 3.3.3), we considered the use of the focal loss as the terminal loss function. As depicted in Table 7, it contains a wealth of classification accuracy results for both the SV and UP datasets, making it highly valuable for analyzing the performance of various hyperspectral classification methods. Firstly, when considering the overall classification results, the SV dataset consistently outperforms the UP dataset for the same classification methods. This disparity can primarily be attributed to the larger scale of labeled training samples available in the SV dataset, which facilitates the construction of a more effective feature space. Similarly, when comparing different training ratios within the same dataset, the same trend persists. Secondly, it is evident that the 2D-CNN method, which solely focuses on spatial features and disregards spectral information, exhibits lower classification accuracy compared to other methods. This observation underscores the essentiality of leveraging spectral–spatial features for robust hyperspectral classification. Lastly, when comparing our proposed FEHN-FL framework with other methods, we consistently achieve superior classification accuracy on both the UP and SV datasets, irrespective of the training ratio employed. This demonstrates the efficacy of the focal loss in addressing class imbalance issues within hyperspectral images and refining decision boundaries during classification. Furthermore, it is noteworthy that even without considering the focal loss as the final loss function, we are still able to attain commendable classification accuracy. For instance, when utilizing the FE-HybridSN+CE framework on the SV dataset with 15% training samples, we observe exceptional accuracy levels of OA (99.97%), AA (99.95%), and kappa (99.98%). To summarize Experiment 3, we provide a set of classification performance charts in Figure 3,Figure 4 and Figure 5 showcasing the results obtained on three distinct experimental datasets. Despite the findings from Figure 5, where our proposed method did not show an improvement in classification accuracy on the SV dataset, this result was expected. The SV dataset inherently has friendly sample feature discrimination and an ample number of labeled samples.
In the final experiment, we provided the learnable parameter sizes of the methods involved, which are closely related to time complexity. As shown in Table 8 and Table 9, although our method’s learned parameter size increased to some extent compared to other methods, it can be observed that its training phase convergence speed is similar to, or even faster than, other methods.

5. Conclusions

The paper proposes a novel hierarchical deep neural network based on CNN for efficient extraction of spectral–spatial features from high-dimensional hyperspectral images. In the proposed framework, the hierarchical convolutional structure effectively reduces computational complexity while effectively capturing both channel and spatial texture information in the hyperspectral data. Specifically, the feature extractor FE-HybridSN consists of a three-layer 3D-CNN component dedicated to extracting spectral information across different channels, complemented by a two-layer 2D-CNN component focusing on spatial information within each channel. Compared with the state-of-the-art methods, FE-HybridSN demonstrates competitive classification performance on widely used datasets such as IP, UP, and SV. Furthermore, we introduce the focal loss as the loss function, which effectively mitigates the problem of biased classification decision boundaries caused by long-tailed distributions.
In conclusion, our research contributes to the field of hyperspectral image analysis by proposing a comprehensive framework that combines feature extraction, dimensionality reduction, and classification techniques. The experimental results demonstrate the potential of these techniques in enhancing the accuracy and efficiency of hyperspectral image classification. Although our method provides efficient performance in HSI classification, there are still several unresolved challenges that may pose future limitations. Our ongoing research will focus on the following directions to address these challenges and further improve our approach:
  • Enhancing the model design to enable adaptive adjustment of decision boundaries based on different hyperspectral datasets. This will allow our method to better accommodate the unique characteristics and variations present in different datasets.
  • Exploring and integrating advanced data augmentation techniques to tackle the issue of limited sample sizes. By generating synthetic data and applying transformational operations, we can effectively expand the training dataset and improve the model’s generalization capability.
  • Investigating alternative strategies to mitigate or alleviate the impact of spatial size during the convolutional process. This includes exploring methods such as multiscale feature extraction and attention mechanisms to capture both local and global spatial information more effectively.
By addressing these areas, we aim to overcome the current limitations and further enhance the robustness, adaptability, and overall performance of our approach in hyperspectral image classification.

Author Contributions

Methodology, X.Y.; Validation, Y.W. and C.Y.; Formal analysis, X.W.; Writing—original draft, X.W.; Writing—review & editing, Y.S. All authors have read and agreed to the published version of the manuscript.

Funding

This work was funded by the National Natural Science Foundation of China under Grant [U20A2082, 41971151] and the Harbin Normal University Postgraduate Innovation Project [HSDSSCX2022-117].

Data Availability Statement

The related datasets in this article is open source. Our proposed framework, FEHN-FL, is implemented in Python and the referential code is available https://github.com/KeithMalarkey/HSI-FEHN-FL, accessed on 1 August 2023.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Bioucas-Dias, J.M.; Plaza, A.; Camps-Valls, G.; Scheunders, P.; Nasrabadi, N.; Chanussot, J. Hyperspectral remote sensing data analysis and future challenges. IEEE Geosci. Remote Sens. Mag. 2013, 1, 6–36. [Google Scholar] [CrossRef]
  2. Keshava, N.; Mustard, J.F. Spectral unmixing. IEEE Signal Process. Mag. 2002, 19, 44–57. [Google Scholar] [CrossRef]
  3. Bioucas-Dias, J.M.; Plaza, A. An overview on hyperspectral unmixing: Geometrical, statistical, and sparse regression based approaches. In Proceedings of the 2011 IEEE International Geoscience and Remote Sensing Symposium, Vancouver, BC, Canada, 24–29 July 2011; pp. 1135–1138. [Google Scholar]
  4. Ye, Q.; Li, Z.; Fu, L.; Zhang, Z.; Yang, W.; Yang, G. Nonpeaked discriminant analysis for data representation. IEEE Trans. Neural Netw. Learn. Syst. 2019, 30, 3818–3832. [Google Scholar] [CrossRef] [PubMed]
  5. Fu, L.; Li, Z.; Ye, Q.; Yin, H.; Liu, Q.; Chen, X.; Fan, X.; Yang, W.; Yang, G. Learning Robust Discriminant Subspace Based on Joint L2,p -and L2,s -Norm Distance Metrics. IEEE Trans. Neural Netw. Learn. Syst. 2020, 33, 130–144. [Google Scholar] [CrossRef]
  6. He, C.; Sun, L.; Huang, W.; Zhang, J.; Zheng, Y.; Jeon, B. TSLRLN: Tensor subspace low-rank learning with non-local prior for hyperspectral image mixed denoising. Signal Process. 2021, 184, 108060. [Google Scholar] [CrossRef]
  7. Huang, J.; Liu, K.; Li, X. Locality constrained low rank representation and automatic dictionary learning for hyperspectral anomaly detection. Remote Sens. 2022, 14, 1327. [Google Scholar] [CrossRef]
  8. Bi, K.; Xiao, S.; Gao, S.; Zhang, C.; Huang, N.; Niu, Z. Estimating vertical chlorophyll concentrations in maize in different health states using hyperspectral LiDAR. IEEE Trans. Geosci. Remote Sens. 2020, 58, 8125–8133. [Google Scholar] [CrossRef]
  9. Liu, Y.; Li, M.; Wang, S.; Wu, T.; Jiang, W.; Liu, Z. Identification of heat damage in imported soybeans based on hyperspectral imaging technology. J. Sci. Food Agric. 2020, 100, 1775–1786. [Google Scholar] [CrossRef]
  10. Wang, W.; Qian, Y.; Tang, Y.Y. Hypergraph-regularized sparse NMF for hyperspectral unmixing. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 9, 681–694. [Google Scholar] [CrossRef]
  11. Shimoni, M.; Haelterman, R.; Perneel, C. Hypersectral imaging for military and security applications: Combining myriad processing and sensing techniques. IEEE Geosci. Remote Sens. Mag. 2019, 7, 101–117. [Google Scholar] [CrossRef]
  12. Barberio, M.; Benedicenti, S.; Pizzicannella, M.; Felli, E.; Collins, T.; Jansen-Winkeln, B.; Marescaux, J.; Viola, M.G.; Diana, M. Intraoperative guidance using hyperspectral imaging: A review for surgeons. Diagnostics 2021, 11, 2066. [Google Scholar] [CrossRef] [PubMed]
  13. Yang, J.; Qian, J. Hyperspectral image classification via multiscale joint collaborative representation with locally adaptive dictionary. IEEE Geosci. Remote Sens. Lett. 2017, 15, 112–116. [Google Scholar] [CrossRef]
  14. Camps-Valls, G.; Gomez-Chova, L.; Muñoz-Marí, J.; Vila-Francés, J.; Calpe-Maravilla, J. Composite kernels for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2006, 3, 93–97. [Google Scholar] [CrossRef]
  15. Gao, Q.; Lim, S.; Jia, X. Hyperspectral image classification using joint sparse model and discontinuity preserving relaxation. IEEE Geosci. Remote Sens. Lett. 2017, 15, 78–82. [Google Scholar] [CrossRef]
  16. Hu, P.; Liu, X.; Cai, Y.; Cai, Z. Band selection of hyperspectral images using multiobjective optimization-based sparse self-representation. IEEE Geosci. Remote Sens. Lett. 2018, 16, 452–456. [Google Scholar] [CrossRef]
  17. Tu, B.; Zhang, X.; Kang, X.; Zhang, G.; Wang, J.; Wu, J. Hyperspectral image classification via fusing correlation coefficient and joint sparse representation. IEEE Geosci. Remote Sens. Lett. 2018, 15, 340–344. [Google Scholar] [CrossRef]
  18. Dundar, T.; Ince, T. Sparse representation-based hyperspectral image classification using multiscale superpixels and guided filter. IEEE Geosci. Remote Sens. Lett. 2018, 16, 246–250. [Google Scholar] [CrossRef]
  19. Gao, P.; Wang, J.; Zhang, H.; Li, Z. Boltzmann entropy-based unsupervised band selection for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2018, 16, 462–466. [Google Scholar] [CrossRef]
  20. Landgrebe, D. Hyperspectral image data analysis. IEEE Signal Process. Mag. 2002, 19, 17–28. [Google Scholar] [CrossRef]
  21. Chen, Y.; Jiang, H.; Li, C.; Jia, X.; Ghamisi, P. Deep feature extraction and classification of hyperspectral images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2016, 54, 6232–6251. [Google Scholar] [CrossRef]
  22. Zhang, M.; Li, W.; Du, Q. Collaborative classification of hyperspectral and visible images with convolutional neural network. J. Appl. Remote Sens. 2017, 11, 042607. [Google Scholar] [CrossRef]
  23. Hamouda, M.; Ettabaa, K.S.; Bouhlel, M.S. Smart feature extraction and classification of hyperspectral images based on convolutional neural networks. IET Image Process. 2020, 14, 1999–2005. [Google Scholar] [CrossRef]
  24. An, J.; Zhang, X.; Jiao, L. Dimensionality reduction based on group-based tensor model for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1497–1501. [Google Scholar] [CrossRef]
  25. Yu, X.; Ding, R.; Shao, J.; Li, X. Hyperspectral Remote Sensing Image Feature Representation Method Based on CAE-H with Nuclear Norm Constraint. Electronics 2021, 10, 2667. [Google Scholar] [CrossRef]
  26. Li, W.; Chen, C.; Zhang, M.; Li, H.; Du, Q. Data augmentation for hyperspectral image classification with deep CNN. IEEE Geosci. Remote Sens. Lett. 2018, 16, 593–597. [Google Scholar] [CrossRef]
  27. Mei, S.; Ji, J.; Bi, Q.; Hou, J.; Du, Q.; Li, W. Integrating spectral and spatial information into deep convolutional neural networks for hyperspectral classification. In Proceedings of the 2016 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Beijing, China, 10–15 July 2016; pp. 5067–5070. [Google Scholar]
  28. Paoletti, M.E.; Haut, J.M.; Plaza, J.; Plaza, A. A new deep convolutional neural network for fast hyperspectral image classification. ISPRS J. Photogramm. Remote Sens. 2018, 145, 120–147. [Google Scholar] [CrossRef]
  29. Li, Y.; Xie, W.; Li, H. Hyperspectral image reconstruction by deep convolutional neural network for classification. Pattern Recognit. 2017, 63, 371–383. [Google Scholar] [CrossRef]
  30. Yang, J.; Zhao, Y.Q.; Chan, J.C.W.; Xiao, L. A multi-scale wavelet 3D-CNN for hyperspectral image super-resolution. Remote Sens. 2019, 11, 1557. [Google Scholar] [CrossRef]
  31. Roy, S.K.; Krishna, G.; Dubey, S.R.; Chaudhuri, B.B. HybridSN: Exploring 3-D–2-D CNN feature hierarchy for hyperspectral image classification. IEEE Geosci. Remote Sens. Lett. 2019, 17, 277–281. [Google Scholar] [CrossRef]
  32. Liu, K.; Zhong, P.; Zheng, Y.; Yang, K.; Liu, M. P_VggNet: A convolutional neural network (CNN) with pixel-based attention map. PLoS ONE 2018, 13, e0208497. [Google Scholar] [CrossRef]
  33. Zhong, Z.; Li, J.; Luo, Z.; Chapman, M. Spectral–spatial residual network for hyperspectral image classification: A 3-D deep learning framework. IEEE Trans. Geosci. Remote Sens. 2017, 56, 847–858. [Google Scholar] [CrossRef]
  34. Vaswani, A.; Shazeer, N.; Parmar, N.; Uszkoreit, J.; Jones, L.; Gomez, A.N.; Kaiser, Ł.; Polosukhin, I. Attention is all you need. Adv. Neural Inf. Process. Syst. 2017, 30, 5998–6008. [Google Scholar]
  35. Dosovitskiy, A.; Beyer, L.; Kolesnikov, A.; Weissenborn, D.; Zhai, X.; Unterthiner, T.; Dehghani, M.; Minderer, M.; Heigold, G.; Gelly, S.; et al. An image is worth 16x16 words: Transformers for image recognition at scale. arXiv 2020, arXiv:2010.11929. [Google Scholar]
  36. Liu, Z.; Lin, Y.; Cao, Y.; Hu, H.; Wei, Y.; Zhang, Z.; Lin, S.; Guo, B. Swin transformer: Hierarchical vision transformer using shifted windows. In Proceedings of the IEEE/CVF International Conference on Computer Vision, Montreal, BC, Canada, 11–17 October 2021; pp. 10012–10022. [Google Scholar]
  37. Liu, B.; Liu, Y.; Zhang, W.; Tian, Y.; Kong, W. Spectral Swin Transformer Network for Hyperspectral Image Classification. Remote Sens. 2023, 15, 3721. [Google Scholar] [CrossRef]
  38. Ayas, S.; Tunc-Gormus, E. SpectralSWIN: A spectral-swin transformer network for hyperspectral image classification. Int. J. Remote Sens. 2022, 43, 4025–4044. [Google Scholar] [CrossRef]
  39. Zhao, E.; Guo, Z.; Li, Y.; Zhang, D. SaaFormer: Spectral-spatial Axial Aggregation Transformer for Hyperspectral Image Classification. arXiv 2023, arXiv:2306.16759. [Google Scholar]
  40. Wang, Y.; Duan, H. Classification of Hyperspectral Images by SVM Using a Composite Kernel by Employing Spectral, Spatial and Hierarchical Structure Information. Remote Sens. 2018, 10, 441. [Google Scholar] [CrossRef]
  41. Haut, J.M.; Paoletti, M.E.; Plaza, J.; Plaza, A.; Li, J. Visual attention-driven hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 8065–8080. [Google Scholar] [CrossRef]
  42. Roy, S.K.; Manna, S.; Song, T.; Bruzzone, L. Attention-based adaptive spectral–spatial kernel ResNet for hyperspectral image classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 7831–7843. [Google Scholar] [CrossRef]
  43. Zhou, B.; Cui, Q.; Wei, X.S.; Chen, Z.M. Bbn: Bilateral-branch network with cumulative learning for long-tailed visual recognition. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Seattle, WA, USA, 13–19 June 2020; pp. 9719–9728. [Google Scholar]
  44. Ochal, M.; Patacchiola, M.; Storkey, A.; Vazquez, J.; Wang, S. Few-shot learning with class imbalance. arXiv 2021, arXiv:2101.02523. [Google Scholar] [CrossRef]
  45. Kubat, M.; Matwin, S. Addressing the curse of imbalanced training sets: One-sided selection. In Proceedings of the Fourteenth International Conference on Machine Learning (ICML 1997), Nashville, TN, USA, 8–12 July 1997; Volume 97, p. 179. [Google Scholar]
  46. Krizhevsky, A.; Sutskever, I.; Hinton, G.E. Imagenet classification with deep convolutional neural networks. Commun. ACM 2017, 60, 84–90. [Google Scholar] [CrossRef]
  47. Ji, S.; Xu, W.; Yang, M.; Yu, K. 3D convolutional neural networks for human action recognition. IEEE Trans. Pattern Anal. Mach. Intell. 2012, 35, 221–231. [Google Scholar] [CrossRef] [PubMed]
  48. Huang, C. ReLU networks are universal approximators via piecewise linear or constant functions. Neural Comput. 2020, 32, 2249–2278. [Google Scholar] [CrossRef]
  49. Lin, T.Y.; Goyal, P.; Girshick, R.; He, K.; Dollár, P. Focal loss for dense object detection. In Proceedings of the IEEE International Conference on Computer Vision, Venice, Italy, 22–29 October 2017; pp. 2980–2988. [Google Scholar]
  50. Audebert, N.; Le Saux, B.; Lefèvre, S. Deep learning for classification of hyperspectral data: A comparative review. IEEE Geosci. Remote Sens. Mag. 2019, 7, 159–173. [Google Scholar] [CrossRef]
  51. Tuia, D.; Persello, C.; Bruzzone, L. Domain adaptation for the classification of remote sensing data: An overview of recent advances. IEEE Geosci. Remote Sens. Mag. 2016, 4, 41–57. [Google Scholar] [CrossRef]
  52. Landgrebe, D.A. Signal Theory Methods in Multispectral Remote Sensing; John Wiley & Sons: Hoboken, NJ, USA, 2003; Volume 24. [Google Scholar]
  53. Makantasis, K.; Karantzalos, K.; Doulamis, A.; Doulamis, N. Deep supervised learning for hyperspectral data classification through convolutional neural networks. In Proceedings of the 2015 IEEE International Geoscience and Remote Sensing Symposium (IGARSS), Milan, Italy, 26–31 July 2015; pp. 4959–4962. [Google Scholar]
  54. Hamida, A.B.; Benoit, A.; Lambert, P.; Amar, C.B. 3-D deep learning approach for remote sensing image classification. IEEE Trans. Geosci. Remote Sens. 2018, 56, 4420–4434. [Google Scholar] [CrossRef]
  55. He, M.; Li, B.; Chen, H. Multi-scale 3D deep convolutional neural network for hyperspectral image classification. In Proceedings of the 2017 IEEE International Conference on Image Processing (ICIP), Beijing, China, 17–20 September 2017; pp. 3904–3908. [Google Scholar]
  56. Kingma, D.P.; Ba, J. Adam: A method for stochastic optimization. arXiv 2014, arXiv:1412.6980. [Google Scholar]
Figure 1. Overview of the proposed feature extractor FE-HybridSN and entire framework FEHN-FL. Note that batch normalization (BN) and rectified linear unit (ReLU), following every convolutional operation, are omitted in the figure.
Figure 1. Overview of the proposed feature extractor FE-HybridSN and entire framework FEHN-FL. Note that batch normalization (BN) and rectified linear unit (ReLU), following every convolutional operation, are omitted in the figure.
Remotesensing 15 04439 g001
Figure 2. Classification accuracies (%) by different methods with 10% training samples and different spatial sizes over SV dataset (the left one is 11 × 11 , another one is 23 × 23 ).
Figure 2. Classification accuracies (%) by different methods with 10% training samples and different spatial sizes over SV dataset (the left one is 11 × 11 , another one is 23 × 23 ).
Remotesensing 15 04439 g002
Figure 3. Classification maps for the IP dataset with 15% labeled training samples and 19 × 19 spatial window. (a) Ground truth. (b) 2D-CNN [53]. (c) 3D-CNN [54]. (d) 3D2D-HybridSN [31]. (e) M3D-CNN [55]. (f) Proposed. In the subfigure (bf), the right brackets encompass the overall classification accuracies, and the best classfication result is highlighted in bold.
Figure 3. Classification maps for the IP dataset with 15% labeled training samples and 19 × 19 spatial window. (a) Ground truth. (b) 2D-CNN [53]. (c) 3D-CNN [54]. (d) 3D2D-HybridSN [31]. (e) M3D-CNN [55]. (f) Proposed. In the subfigure (bf), the right brackets encompass the overall classification accuracies, and the best classfication result is highlighted in bold.
Remotesensing 15 04439 g003
Figure 4. Classification maps for the UP dataset with 10% labeled training samples and 19 × 19 spatial window. (a) Ground truth. (b) 2D-CNN [53]. (c) 3D-CNN [54]. (d) 3D2D-HybridSN [31]. (e) M3D-CNN [55]. (f) Proposed. In the subfigure (bf), the right brackets encompass the overall classification accuracies, and the best classification result is highlighted in bold.
Figure 4. Classification maps for the UP dataset with 10% labeled training samples and 19 × 19 spatial window. (a) Ground truth. (b) 2D-CNN [53]. (c) 3D-CNN [54]. (d) 3D2D-HybridSN [31]. (e) M3D-CNN [55]. (f) Proposed. In the subfigure (bf), the right brackets encompass the overall classification accuracies, and the best classification result is highlighted in bold.
Remotesensing 15 04439 g004
Figure 5. Classification maps for the SV dataset with 5% labeled training samples and 19 × 19 spatial window. (a) Ground truth. (b) 2D-CNN [53]. (c) 3D-CNN [54]. (d) 3D2D-HybridSN [31]. (e) M3D-CNN [55]. (f) Proposed. In the subfigure (bf), the right brackets encompass the overall classification accuracies, and the best classification result is highlighted in bold.
Figure 5. Classification maps for the SV dataset with 5% labeled training samples and 19 × 19 spatial window. (a) Ground truth. (b) 2D-CNN [53]. (c) 3D-CNN [54]. (d) 3D2D-HybridSN [31]. (e) M3D-CNN [55]. (f) Proposed. In the subfigure (bf), the right brackets encompass the overall classification accuracies, and the best classification result is highlighted in bold.
Remotesensing 15 04439 g005aRemotesensing 15 04439 g005b
Table 1. Lay-wise summary of the proposed FE-hybrid architecture with spatial size 15 × 15 .
Table 1. Lay-wise summary of the proposed FE-hybrid architecture with spatial size 15 × 15 .
Layer-IndexOutput Shape# Parameter
Input-0(1, 15, 15, 15)-
Conv3d-1(8, 9, 13, 13)512
BatchNorm3d-2(8, 9, 13, 13)16
ReLU-3(8, 9, 13, 13)0
Conv3d-4(16, 5, 11, 11)5776
BatchNorm3d-5(16, 5, 11, 11)32
ReLU-6(16, 5, 11, 11)0
Conv3d-7(32, 3, 9, 9)13,856
BatchNorm3d-8(32, 3, 9, 9)64
ReLU-9(16, 5, 11, 11)0
Conv2d-10(128, 7, 7)110,720
BatchNorm2d-11(128, 7, 7)256
ReLU-12(128, 7, 7)0
Conv2d-13(256, 5, 5)295,168
BatchNorm2d-14(256, 5, 5)512
ReLU-15(256, 5, 5)0
Linear-16(256)1,638,656
Dropout-17(256)0
Linear-18(128)32,896
Dropout-19(128)0
Linear-20(16)2064
Total parameters: 2,100,528
Trainable parameters: 2,100,528
Nontrainable parameters: 0
The parameters are based on the Indian Pines (IP) dataset.
Table 3. Training hyperparameter settings in the FEHN-FL.
Table 3. Training hyperparameter settings in the FEHN-FL.
IPUPSV
Epochs150
OptimizerAdam [56]
Batch size53128128
Learning rate (LR)1 × 10−3
α t and γ 1 and 1 × 10−6
Table 4. Classification accuracies (%) by different methods for IP dataset with 15% labeled training samples and various spatial sizes. Note that the bold represents the best performance with corresponding experimental configuration for different methods.
Table 4. Classification accuracies (%) by different methods for IP dataset with 15% labeled training samples and various spatial sizes. Note that the bold represents the best performance with corresponding experimental configuration for different methods.
Methods 11 × 11 15 × 15 19 × 19 23 × 23
OA AA Kappa OA AA Kappa OA AA Kappa OA AA Kappa
2D-CNN [53]88.3884.8286.798.2898.2598.0393.4290.3492.4895.9895.8595.41
3D-CNN [54]96.7596.1896.2998.8298.9698.6598.5598.2598.3598.2498.4298.00
3D2D-HybridSN [31]97.2397.0196.8498.5598.5698.3998.6098.5298.4098.2198.4198.30
M3D-CNN [55]96.9095.2996.4698.0796.6397.8098.3097.9798.0698.3595.5498.12
FE-HybridSN97.4197.8597.0498.8599.0298.7198.6598.5698.4698.7799.1198.60
Table 5. Classification accuracies (%) by different methods for UP dataset with 5% labeled training samples and various spatial sizes. Note that the bold represents the best performance with corresponding experimental configuration for different methods.
Table 5. Classification accuracies (%) by different methods for UP dataset with 5% labeled training samples and various spatial sizes. Note that the bold represents the best performance with corresponding experimental configuration for different methods.
Methods 11 × 11 15 × 15 19 × 19 23 × 23
OA AA Kappa OA AA Kappa OA AA Kappa OA AA Kappa
2D-CNN [53]95.7193.7594.3299.1598.7998.8798.3697.2397.8395.8094.0894.45
3D-CNN [54]98.9998.4398.6699.5399.2899.3899.5399.1999.3899.4998.9099.32
3D2D-HybridSN [31]99.4099.0299.2099.3899.2099.1799.5999.3799.4699.7099.3799.60
M3D-CNN [55]99.0898.1798.7899.5199.1599.4699.5799.1899.4499.6599.1399.54
FE-HybridSN99.4799.1399.3099.6199.4099.4899.7399.5099.6499.7399.3799.64
Table 6. Classification accuracies (%) by different methods over SV and UP datasets with 11 × 11 spatial size and different training ratios. Note that the bold represents the best performance with corresponding experimental configuration for different methods.
Table 6. Classification accuracies (%) by different methods over SV and UP datasets with 11 × 11 spatial size and different training ratios. Note that the bold represents the best performance with corresponding experimental configuration for different methods.
Training RatioMethodsSVUP
OA AA Kappa OA AA Kappa
15%2D-CNN [53]98.9199.4298.7898.4397.4897.91
3D-CNN [54]99.9099.8899.8999.6999.4499.59
3D2D-HybridSN [31]99.9099.8899.8999.3299.0999.09
FE-HybridSN99.9199.8999.9099.7399.6699.64
10%2D-CNN [53]98.7999.3998.6597.2196.4296.30
3D-CNN [54]99.7799.8799.7499.3499.0699.12
3D2D-HybridSN [31]99.8299.8599.8099.7199.6099.62
FE-HybridSN99.8599.9299.8399.8499.7799.79
5%2D-CNN [53]98.2199.0298.0196.0595.0594.77
3D-CNN [54]98.9299.1898.8099.0398.4898.71
3D2D-HybridSN [31]98.4898.4698.3199.2198.9798.95
FE-HybridSN98.9299.2299.4599.4599.1799.27
Table 7. Classification accuracies (%) by different methods combined with CE, MCH, and focal loss over SV and UP datasets with 23 × 23 spatial size and different training ratios. Note that the bold represents the best performance with corresponding experimental configuration for different methods.
Table 7. Classification accuracies (%) by different methods combined with CE, MCH, and focal loss over SV and UP datasets with 23 × 23 spatial size and different training ratios. Note that the bold represents the best performance with corresponding experimental configuration for different methods.
Training RatioMethodsSVUP
OA AA Kappa OA AA Kappa
15%2D-CNN [53] + CE99.9599.8999.9499.0998.7198.80
3D-CNN [54] + CE99.9999.9999.9999.9199.8499.88
3D2D-HybridSN [31] + CE99.9999.9999.9999.9599.9199.94
M3D-CNN [55] + CE99.9999.9999.9999.8799.7999.82
FE-HybridSN + CE99.9799.9599.9899.9399.9199.91
FE-HybridSN + MCH99.9699.9399.9699.2499.0198.99
Proposed99.9999.9999.9999.9599.9599.94
10%2D-CNN [53] + CE99.8999.8999.8898.1497.1297.54
3D-CNN [54] + CE99.9499.9499.9499.7499.2199.65
3D2D-HybridSN [31] + CE99.9499.9399.9399.8999.7799.84
M3D-CNN [55] + CE99.9399.9099.9299.8399.5499.77
FE-HybridSN + CE99.9699.9499.9399.8999.7999.83
FE-HybridSN + MCH99.8599.9499.8499.7299.5399.64
Proposed99.9699.9599.9699.8999.7999.85
5%2D-CNN [53] + CE98.8699.1798.8595.8094.0894.45
3D-CNN [54] + CE99.6899.8199.7599.4998.9099.32
3D2D-HybridSN [31] + CE99.6499.7999.6099.7099.3799.60
M3D-CNN [55] + CE99.6099.0399.4799.6599.1399.54
FE-HybridSN + CE99.3499.6799.2799.7399.3799.64
FE-HybridSN + MCH99.5599.7599.5099.3698.6199.15
Proposed99.7599.8899.7999.7599.4399.65
Table 8. Trainable parameter scales for different backbone networks.
Table 8. Trainable parameter scales for different backbone networks.
2D-CNN [53]3D-CNN [54]M3D-CNN [55]3D2D-HybridSN [31]Proposed
Params0.19 M0.51 M0.99 M1.79 M2.19 M
Table 9. The average training time using 10% labeled training samples with 23 × 23 spatial size over IP, UP, and SV datasets. Note that the minimum average training time is highlighted in bold.
Table 9. The average training time using 10% labeled training samples with 23 × 23 spatial size over IP, UP, and SV datasets. Note that the minimum average training time is highlighted in bold.
Focal Loss (min)CE Loss (min)MCH (min)
IP4.124.134.22
UP14.9315.3515.01
SV18.8221.7719.08
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Wen, X.; Yu, X.; Wang, Y.; Yang, C.; Sun, Y. A Hybrid 3D–2D Feature Hierarchy CNN with Focal Loss for Hyperspectral Image Classification. Remote Sens. 2023, 15, 4439. https://doi.org/10.3390/rs15184439

AMA Style

Wen X, Yu X, Wang Y, Yang C, Sun Y. A Hybrid 3D–2D Feature Hierarchy CNN with Focal Loss for Hyperspectral Image Classification. Remote Sensing. 2023; 15(18):4439. https://doi.org/10.3390/rs15184439

Chicago/Turabian Style

Wen, Xiaoyan, Xiaodong Yu, Yufan Wang, Cuiping Yang, and Yu Sun. 2023. "A Hybrid 3D–2D Feature Hierarchy CNN with Focal Loss for Hyperspectral Image Classification" Remote Sensing 15, no. 18: 4439. https://doi.org/10.3390/rs15184439

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop