Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
A Framework for Multivariate Analysis of Land Surface Dynamics and Driving Variables—A Case Study for Indo-Gangetic River Basins
Next Article in Special Issue
SII-Net: Spatial Information Integration Network for Small Target Detection in SAR Images
Previous Article in Journal
Design and Experimental Evaluation of an Aerial Solution for Visual Inspection of Tunnel-like Infrastructures
Previous Article in Special Issue
Low Contrast Infrared Target Detection Method Based on Residual Thermal Backbone Network and Weighting Loss Function
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

MCMS-STM: An Extension of Support Tensor Machine for Multiclass Multiscale Object Recognition in Remote Sensing Images

School of Electronics and Information Engineering, Harbin Institute of Technology, Harbin 150001, China
*
Author to whom correspondence should be addressed.
Remote Sens. 2022, 14(1), 196; https://doi.org/10.3390/rs14010196
Submission received: 24 November 2021 / Revised: 30 December 2021 / Accepted: 30 December 2021 / Published: 2 January 2022
Figure 1
<p>Examples of two types of image normalization methods.</p> ">
Figure 2
<p>Illustration of objects classification for MCMS-STM.</p> ">
Figure 3
<p>Illustration of classification hyperplane defined by rank-1 projection tensor and rank-<span class="html-italic">R</span> projection tensor. The bule solid line denotes the hyperplane defined by rank-1 or rank-<span class="html-italic">R</span> tensor, and the blue dotted line denotes the hyperplane defined by rank-<span class="html-italic">R</span> tensor.</p> ">
Figure 4
<p>Illustration of slices with proper sizes and multiscale slices. (<b>a</b>) Proper sizes of slices for different objects (<b>b</b>) illustration of multiscale slice.</p> ">
Figure 5
<p>Graph based constraints model. Each vertex in the graph indicates a set of dual variables, and the sum of each vertex with the same color is equal.</p> ">
Figure 6
<p>Airplanes slices with different types from two slices sets of dataset 1. (<b>a</b>–<b>l</b>) are from slices set 1 of dataset 1, and (<b>m</b>–<b>x</b>) are from slices set 2 of dataset 1.</p> ">
Figure 7
<p>Ship slices from two slices sets of dataset 2. (<b>a</b>–<b>e</b>) are from slices set 1 of dataset 2, and (<b>f</b>–<b>j</b>) are from slices set 2 of dataset 2.</p> ">
Figure 7 Cont.
<p>Ship slices from two slices sets of dataset 2. (<b>a</b>–<b>e</b>) are from slices set 1 of dataset 2, and (<b>f</b>–<b>j</b>) are from slices set 2 of dataset 2.</p> ">
Figure 8
<p>The classification accuracies of MCMS-STM under different <span class="html-italic">R</span> and <span class="html-italic">C</span> using dataset 1. (<b>a</b>) The classification accuracies of OVR version of MCMS-STM with different <span class="html-italic">R</span> and <span class="html-italic">C</span>; (<b>b</b>) The classification accuracies of OVO version of MCMS-STM with different <span class="html-italic">R</span> and <span class="html-italic">C</span>.</p> ">
Figure 8 Cont.
<p>The classification accuracies of MCMS-STM under different <span class="html-italic">R</span> and <span class="html-italic">C</span> using dataset 1. (<b>a</b>) The classification accuracies of OVR version of MCMS-STM with different <span class="html-italic">R</span> and <span class="html-italic">C</span>; (<b>b</b>) The classification accuracies of OVO version of MCMS-STM with different <span class="html-italic">R</span> and <span class="html-italic">C</span>.</p> ">
Figure 9
<p>The classification accuracies of MCMS-STM with or without different image resizing operations.</p> ">
Figure 10
<p>The time consumption of solving the dual problem of MCMS-STM and multiclass SVM using different optimization solving algorithms under different numbers of training samples.</p> ">
Versions Notes

Abstract

:
The support tensor machine (STM) extended from support vector machine (SVM) can maintain the inherent information of remote sensing image (RSI) represented as tensor and obtain effective recognition results using a few training samples. However, the conventional STM is binary and fails to handle multiclass classification directly. In addition, the existing STMs cannot process objects with different sizes represented as multiscale tensors and have to resize object slices to a fixed size, causing excessive background interferences or loss of object’s scale information. Therefore, the multiclass multiscale support tensor machine (MCMS-STM) is proposed to recognize effectively multiclass objects with different sizes in RSIs. To achieve multiclass classification, by embedding one-versus-rest and one-versus-one mechanisms, multiple hyperplanes described by rank-R tensors are built simultaneously instead of single hyperplane described by rank-1 tensor in STM to separate input with different classes. To handle multiscale objects, multiple slices of different sizes are extracted to cover the object with an unknown class and expressed as multiscale tensors. Then, M-dimensional hyperplanes are established to project the input of multiscale tensors into class space. To ensure an efficient training of MCMS-STM, a decomposition algorithm is presented to break the complex dual problem of MCMS-STM into a series of analytic sub-optimizations. Using publicly available RSIs, the experimental results demonstrate that the MCMS-STM achieves 89.5% and 91.4% accuracy for classifying airplanes and ships with different classes and sizes, which outperforms typical SVM and STM methods.

1. Introduction

The diverse types and sizes of objects bring a challenge to object recognition in remote sensing images (RSIs). Recently, mainly owing to its powerful feature abstraction ability, various deep learning technologies have achieved impressive success in different object recognition tasks, such as Fast R-CNN [1], Faster R-CNN [2], local attention based CNN [3], 2 CNN [4], SSD [5] and YOLO [6]. In addition to object recognition, deep learning-based methods are widely used for a wide variety of classification tasks based on remote sensing data acquired by different sensors, such as graph convolution neural networks [7] based hyperspectral image classification [8], and multimodal deep learning-based multisource image classification [9]. Despite the recent advances, deep learning-based methods rely heavily on massive available labeled samples. In comparison, the machine learning-based object recognition method can obtain effective results using a small number of samples.
For the object recognition in RSIs, the general procedure consists of two steps, i.e., extracting an object slice using object detection method and recognizing the type of object contained in the slice using a trained classifier. By virtue of the object detection methods, e.g., region proposal-based methods [10,11,12] and saliency-based methods [13,14,15,16], the positions of objects in RSIs can be obtained. Then, the slice of the object can be extracted for subsequent classification. It is worth noting that most existing classifiers, including deep learning methods [1,2,3,4,5] and machine learning methods [17,18,19,20], can only deal with the input of the fixed size. Thereby, two image resizing operations are always used to extract slices containing objects with the fixed size, as shown in Figure 1, one of which extracts the slices by the appropriate size to ensure that the object can be exactly covered, and then resizes the slice to the same size using the image interpolation method [10,11,12]. Another way is to extract slices according to the maximum size of objects to cover each size of object [13,14,15,16]. After obtaining the slices with the fixed size, a proper classifier should be designed to predict the corresponding class label. Among the various machine learning based classifiers, the support vector machine (SVM) as well as its tensor extension, i.e., support tensor machine (STM), are the most worth considering due to a sound theoretical foundation and excellent generality capability. Besides the standard SVM (i.e., C SVM ) [17], various variants of SVM were established to improve the performance of SVM from different perspectives, including ν SVM [18], least-square SVM [19], etc. Since the standard SVM can be considered as a bi-category classifier and thus cannot be used to address multiclass classifications directly, the One-Versus-One (OVO) [20], Error Correcting Output Codes (ECOC) [21], and One-Versus-Rest (OVR) [22] strategies were employed in order to use various binary SVMs to achieve multiclass classifications indirectly. Different from the classic SVM that relies on auxiliary strategies to achieve multiclass classification, the multiclass SVM was developed to classify multiclass objects by solving parameters of multiple classification hyperplanes simultaneously [22]. Besides, based on the multiclass classification mechanism of [22], some other multiclass-oriented SVM variants were developed, e.g., the graph embedded multiclass SVM [23] and the least-squares twin multiclass SVM [24].
Since the object slice can be represented naturally as a tensor in RSI, the tensor-based classifier, i.e., STM, was developed to exploit the structural information embedded in RSI better. According to the supervised tensor learning framework [25] established in 2005, the C SVM and ν SVM were extended to C STM and ν STM by applying multilinear operator of tensor space, respectively. Besides, a large number of variants of STMs were established to deal with different classification tasks, such as linear support higher-order tensor machine (SHTM) [26], multi-kernel STM [27], higher rank support tensor machine [28], and support tucker machine [29]. Moreover, the support multimode tensor machine (SMTM) [30] was built to exploit the multimode product of tensor to obtain the results of the multiclass classification for different perspectives, while its multiclass classification strategy is similar to conventional ECOC strategy.
For classifying objects with multiple classes and different sizes, most existing STM methods need to exploit OVO or OVR strategies to train a series of binary classifiers to perform multiclass classification indirectly. The OVO and OVR strategies inherently assume that multiclass classifications can be solved by multiple independent bi-category classifications, ignoring the correlation between multiple classes [23]. Although there exist a few SVM methods that can handle multiclass classification directly, e.g., multiclass SVM [22] and its variants [23,24], their multiclass classification mechanism that is similar to OVR may generate excessive constraints to limit the classification hyperplane, leading to the increase of complexity for training classifier [31]. On the other hand, since the existing STMs only accept tensors with a fixed size as input, it needs to resize the object slices of different classes in RSIs into a fixed size. Considering that objects with different classes usually present different sizes, the typical two resizing operations will bring adverse effect on object recognition, of which the first type of resizing operation will lead to the loss of objects’ scale information, and the second type of resizing operation is easy to cause the slice to contain more background interferences. The examples of two types of resizing operations are given in Figure 1. It is seen that using the second type of resizing operation will lose the scale information of the ships and thus cannot identify ships according to their size feature, and using the first type of resizing operation will generate slices with a large size so that the slice containing ship with class 1 covers excessive background interferences. To deal with the slices to be classified present different sizes, the deep learning method, i.e., the scale free convolution neural network, is built to utilize the global average pooling to map feature maps to unify size [32]. However, for representative machine learning methods, e.g., STM, there is no related work that can process the input slices with different sizes. In comparison to image resizing, the objects with different sizes should be contained by slices with proper sizes to reduce the impact of background interferences and maintain inherent scale information of the contained object, and these slices can be naturally represented as tensors with different dimensions, denoted as multiscale tensors in this paper, while the existing STMs can only process tensor with the same size and cannot process the slices with different sizes represented as multiscale tensors.
Motived by the abovementioned issues, the multiclass multiscale support tensor machine (MCMS-STM) is proposed in this paper. To deal with mutliclass classification, by integrating OVR and OVO strategies to optimization problems, a new multiclass classification mechanism is constructed to use multiple hyperplanes defined by rank-R tensors instead of a single hyperplane defined by rank-1 tensor of STM, where each hyperplane is needed to separate samples with specific classes. Furthermore, to classify objects with different sizes, according to positions of objects obtained from detection results, it is necessary to extract the objects slices with proper sizes rather than the fixed size to reduce the impact of background interferences and maintain inherent scale information of the contained object. These slices with different sizes can be naturally represented as tensors with different dimensions, denoted as multiscale tensors in this paper. Note that the existing STM methods can only process tensor with the same size and cannot process the slices with different sizes represented as multiscale tensors. To deal with input of multiscale tensors, instead of the fixed-dimensional hyperplane used in STM, the M-dimensional hyperplanes are built to separate input of multiscale tensors, and the resulting projecting value is used to predict the class label of input to achieve cross-scale object recognition. In addition, to train the OVO version of MCMS-STM efficiently, a decomposition algorithm is proposed to split the dual problem of the MCMS-STM into a series of sub-optimizations to accelerate the training.
The remainder of this paper is organized as follows. Section 2 consists of some preliminaries, such as the basic definitions and notions, the classical SVM and STM methods. In Section 3, the OVR version and OVO version of MCMS-STM and the corresponding solving methods are presented. Then, the decomposition algorithm is constructed to accelerate the training of the OVO version of MCMS-STM, and the relationship between the multiclass classification mechanism used in MCMS-STM and the existing methods is discussed. In Section 4, the experiments are conducted on publicly RSIs to analyze the parameter setting and the impact of image resizing operation and evaluate the performance of the MCMS-STM. Our conclusion is given in Section 5.

2. Preliminaries and Related Work

Before presenting the MCMS-STM, the notations, abbreviations, the basic tensor algebra used throughout this paper, and the traditional SVM and STM are introduced briefly as follows.

2.1. Notations, Abbreviations, and Tensor Operation

Tensor is the expansion of vector and matrix to the higher dimension, and tensor algebra, also known as multilinear algebra, is the extension of linear algebra to multiway data. To distinguish between tensors, matrices, and vectors, according to the convention in [33], the used symbols throughout this paper are summarized in Table 1.
Then, we summarize all the notations and abbreviations used throughout this paper in Table 2.
The used basic tensor operations throughout this paper are given as follows.
Definition 1.
Frobenius norm and inner product of tensors: The Frobenius norm F of a tensor A I 1 , , I M is calculated as
A F = i 1 = 1 , , i M = 1 i 1 = I 1 , , i M = I M A ( i 1 , , i M ) × A ( i 1 , , i M )
The inner product between tensors A I 1 , , I M , I 1 , , I M is defined as the sum of the products of their corresponding entries, i.e.,
A , B = i 1 = 1 , , i M = 1 i 1 = I 1 , , i M = I M A ( i 1 , , i M ) × B ( i 1 , , i M )  
Definition 2.
Mode-k product of tensor: Given tensor A I 1 , , I M and matrix B I k , I k , the mode-k product between A and B is denoted as A × k B , whose results are tensor C I 1 × × I k × × I M , as calculated by Equation (3).
C ( I 1 , , I k 1 , q , I k + 1 , , I M ) = A × k B = p = 1 I k A ( I 1 , , I k 1 , p , I k + 1 , , I M ) × B ( p , q )  
Definition 3.
Outer product: The outer product of M vectors (i.e., { a m I m } m = 1 M ) is denoted as A = a 1   a M , and the corresponding results are represented as an M-order tensor A I 1 , , I M , whose entry with coordinate ( i 1 , , i M ) is calculated by
A ( i 1 , , i M ) = a 1 ( i 1 ) × , , × a M ( i M )

2.2. Classical Binary and Multiclass Support Vector Machine

Consider bi-category classification task, given training set containing N samples, i.e., { x i , y i } | N i = 1 , where x i I 1 and y i { 1 , 1 } denote the input vector of ith sample and the corresponding label, respectively. SVM aims at learning the parameters of classification hyperplane with the largest classification margin from the training set, which can be drawn using the following optimization problem.
m i n w , b 1 2 w T w + C i = 1 N ξ i s . t . y i ( x i T w + b ) 1 ξ i   1 i N ξ i 01 i N
where w, b, ξ i and C denote the normal vector for classification hyperplane, the bias, the slacking variable, and the regularization parameter, respectively. To obtain the optimal w, the dual problem can be drawn as follows. The detailed derivation can refer to [17].
m i n w , b 1 2 a Q a a T e s . t . i y i a i = 0 0 a i C 1 i N ,
where a denotes the Lagrangian multiplier. Q is a positive semidefinite matrix, where Q ( i , j ) = y i y j x i T x j . After solving the optimal α , the w can be calculated by Equation (7).
w = i a i y i x i .
Then, the label of test sample x can be predicted by Equation (8).
y ( x ) = s i g n ( x T w + b ) ,
where s i g n ( ) denotes the sign function.
The standard SVM can only deal with the bi-category classification problem, while most practical applications can be regarded as multiclass classification problems. Facing this situation, the multiclass SVM, i.e., an extension of SVM in multiclass classification problems, is established to identify samples with different classes.
Given the training set { x i , y i } | i = 1 N , where x i and y i { 1 , , M } denote the input vector of the ith sample and the corresponding class label. The multiclass SVM aims at learning M classification hyperplanes simultaneously by solving the following optimization problem.
m i n w m | m = 1 M , b 1 2 m = 1 M w m T w m + C × i = 1 N m y i ξ i m s . t . x i T w y i + b y i x i T w m + b m + 2 ξ i m   1 i N   m y i ξ i m 0   1 i N   m y i ,
where w m | m = 1 M , bm and ξ i m denote the normal orientation of M classification hyperplanes, the mth bias, the slacking variable, respectively. The exact solving method can refer to [22].
According to the resulting w m | m = 1 M and bm, the class label of test sample x can be predicted using the following decision function.
y ( x ) = m a x m { x T w m + b m | 1 m M }

2.3. Support Tensor Machine

To better use the structural information of input data represented as a tensor, the STM is extended from SVM to directly separate input in tensor space.
The given training set containing N samples { X i , y i } | N i = 1 , where X i I 1 , , I M denotes the input tensor and y i { 1 , 1 } denotes the corresponding label. Compared to SVM that separates input in vector space, the STM utilizes the mode-k product to process input in tensor space. The corresponding optimization problem of STM is given in Equation (11).
m i n w i | M i = 1 1 2 w 1 , , w m F 2 + C i = 1 N x i s . t . y i ( X i j = 1 M × j w j + b ) 1 ξ i   1 i N ξ i 0   1 i N
where w j I j denotes the projection vector along with the jth mode of the input tensor. The W = w 1   w m I 1 , , I M denotes the projection tensor used to indicate the orientation of the normal orientation for the classification hyperplane. Since the parameters of STM w 1   w m R I 1 × , , × I M is the M-order tensor, it is difficult to be solved directly. In this way, the alternating optimization scheme is introduced to solve the above optimization, i.e., optimize each w i k in turn by fixing the rest w i | i i k in the kth iteration, where i k = m o d ( k , m ) + 1 denotes the index of order to be optimized. The optimization problem in kth iteration is equivalent to that in SVM (see Equation (5)) it can be solved by using SVM solver. After the alternating optimization procedure is terminated, the optimal parameters w j | j and b can be utilized to predict the label of the test sample X , as shown below.
y ( X ) = s i g n ( X j = 1 l × j w j + b )

3. Multiclass Multiscale Support Tensor Machine

As discussed above, the standard STM fails to deal with multiclass classification, and it requires resizing preprocessing to generate slices of the fixed size to comply with its input requirements, causing loss of scale information or increase of background interferences. To address multiclass multiscale object recognition in RSIs, the MCMS-STM is proposed by solving simultaneously hyperplanes of multiple dimensions defined by multiscale rank-R tensors to directly classify objects with different classes and sizes, as illustrated in Figure 2.

3.1. Construction of Multiclass Multiscale Support Tensor Machine

Starting from the standard STM, the classifier is extended gradually from two aspects, i.e., multiclass and multiscale, to form the proposed MCMS-STM.

3.1.1. Extend STM to Deal with Multiclass Classification

Consider the training set { X i , y i } i = 1 N set with N samples of M classes, where X i and y i { 1 , 2 , , M } denote the L-order tensor representation of i-th image slice and the corresponding class label, respectively. Note that the standard STM can process input represented as a tensor, while it cannot process multiclass classifications directly. In contrast, the multiclass SVM can deal with multiclass classifications, while it cannot process input represented as a tensor. To achieve multiclass classification in tensor space, a straightforward idea is to construct the following optimization problem to learn parameters of multiple hyperplanes in tensor space simultaneously by integrating the merits of STM (see Equation (11)) and multiclass SVM (see Equation (9)).
m i n w m l | m = 1 , l = 1 M , L , b 1 2 m = 1 M w m 1 w m L F 2 + C × i = 1 N m y i ξ i m s . t . X i l = 1 L × l w y i l + b y i X i l = 1 L × l w m l + b m + 2 ξ i m ξ i m 0     1 i N     m y i ,
where w m 1 w m L I 1 , I 2 , , I L  denotes the mth projection tensor used to determine the normal orientation of mth hyperplane. Compared with the optimization problem of multiclass SVM (see Equation (9)), the optimization problem in Equation (13) allows the tensor X i as input by constructing projection tensor W m with the same dimensions as X i , while more detailed modifies are essential to improve the performance of multiclass classification.
According to the definition of tensor rank, e.g., CANDECOMP/PARAFAC (CP) rank [34], the tensor W m belongs to the rank-1 tensor. Referring to the study in [28,29] that the single rank-1 tensor cannot be used to describe the classification hyperplane accurately, it is considered to utilize multiple rank-1 tensors to improve the effect of classification, as shown in Equation (14).
m i n w m ( r , l ) | m , g , l , b 1 2 m = 1 M r = 1 R w m ( r , 1 ) w m ( r , L ) F 2 + C i = 1 N m = 1 M ξ i m s . t . r X i l = 1 L × l w y i ( r , l ) + b y i r X i l = 1 L × l w m ( r , l ) + b m + 2 ξ i m ξ i m 0   m y i   1 i N ,
where R denotes the number of rank-1 projecting tensors. In Equation (14), the hyperplane is defined by the sum of R rank-1 projection tensors (i.e., the rank-R projecting tensor W m ). The illustration of hyperplanes defined by rank-1 projecting tensor and hyperplanes defined by rank-R tensor is shown in Figure 3. It is seen that the linear combination of rank-1 projecting tensors can be used to obtain the hyperplane with more orientations, which indicates that the rank-R projection tensor is likely to obtain a more effective hyperplane.
Note that the multiclass classification mechanism of Equation (14) is equivalent to that of multiclass SVM (see Equation (9)). In other words, the multiclass classification mechanism of Equation (14) can be described as an OVR strategy that solves M hyperplanes simultaneously, of which mth hyperplane defined by W m is used to separate class m from the other. Considering the related research that the OVO strategy based SVM always achieves better results than OVR strategy based SVM, it is taken into consideration to embed OVO strategy to tensor space, of which the hyperplane defined by projection tensor W m , m is used to separate samples from class m and samples from class m . In this way, the optimization problem in Equation (14) can be changed to the optimization problem in Equation (15).
m i n w m , m ( r , l ) , b 1 2 m m r = 1 R w m , m ( r , 1 ) w m , m ( r , L ) F 2 + C i = 1 N m = 1 M ξ i m s . t . r X i l = 1 L × l w y i , m ( r , l ) + b y i , m r X i l = 1 L × l w m , y i ( r , l ) + b m , y i + 2 ξ i m     m y i     1 i N ξ i m 0     m y i     1 i N ,
where W m , m = r = 1 R w m , m ( r , 1 ) w m , m ( r , L ) and b m , m denote the normal orientation of the hyperplane used to separate class m from class m and the corresponding bias. Since OVR and OVO strategies have their own advantages, these two classification mechanisms are used as different versions of MCMS-STM to cope with different classification tasks.

3.1.2. Classify the Multiscale Objects without Image Resizing Preprocessing

For the objects to be classified presenting different sizes, the convention manner is to adjust slices of different sizes into the fixed size, whereas this operation is easy to lead to the loss of scale information or increase of background interferences.
To maintain the scale information of objects and reduce the impact of background interferences, it is needed to extract slices with proper size to cover the object with a specific class, as shown in Figure 4. Assuming that each class of object presents a specific size, it is necessary to extract M sizes of slices of different sizes. Even for the same object, M slices of different sizes still need to be extracted because the category of the test sample is unknown (see Figure 4b). Therefore, different from the conventional manner that each object is described as a fixed-size slice, for each object, it is needed to extract M sizes of slices represented M-scale tensors X i ( m ) | m = 1 M . In this way, a multiscale sample set { X i ( m ) | m = 1 M , y i } i = 1 N is constructed, where X i ( m ) I 1 m , , I L m denotes the slice of mth size containing ith object. To separate input represented as M-scale tensors X i ( m ) | m = 1 M in tensor space, it is also required to construct projection tensors with M scales, i.e., { W m , m I 1 m , , I L m | 1 m M , m m } , where the projection tensor with specific dimension is used to separate input slice with specific scale. In this way, the classification model in Equation (15) can be further extended to identify the multiscale objects without image resizing operation, and the corresponding optimization problem is shown in Equation (16).
m i n w m , m ( r , l ) , b 1 2 m m r = 1 R w m , m ( r , 1 ) w m , m ( r , L ) F 2 + C i = 1 N m = 1 M ξ i m s . t . r X i ( y i ) l = 1 L × l w y i , m ( r , l ) + b y i , m r X i ( m ) l = 1 L × l w m , y i ( r , l ) + b m , y i + 2 ξ i m   m y i   1 i N ξ i m 0   m y i   1 i N ,
Equation (16) is the final optimization problem of OVO version of MCMS-STM. For OVR version of MCMS-STM, the final optimization problem is given directly as follows.
m i n w m , m ( r , l ) , b 1 2 m = 1 M r = 1 R w m ( r , 1 )   w m ( r , L ) F 2 + C i = 1 N m = 1 M ξ i m s . t . r X i ( y i ) l = 1 L × l w y i ( r , l ) + b y i r X i ( m ) l = 1 L × l w m ( r , l ) + b m + 2 x i m     m y i     1 i N ξ i m 0   m y i   1 i N ,
Note that the OVR version of MCMS-STM only uses M hyperplanes defined by W m = r = 1 R w m ( r , 1 ) w m ( r , L ) to separate samples.
On the one hand, the proposed MCMS-STM can use multiscale tensors to capture the category and scale information of objects and utilize grouped rank-R tensors to improve the effectiveness of multiclass classification problems. On the other hand, it can be found that the optimization problem of MCMS-STM is more complex than the binary STM, which needs to establish a specific solving method to obtain the optimal parameters.

3.2. Solving of the Optimization Problem for MCMS-STM

Since the procedures of training OVO and OVR versions of MCMS-STM are similar, we only give the solving methods for the OVO version of the MCMS-STM in this section and put the solving methods for the OVR version of the MCMS-STM in the Appendix A.
To solve the optimization problem in Equation (16) effectively, an alternating optimization scheme [25] is adopted for model training, i.e., optimize { w m , m ( r , l i t e r ) | 1 r R , 1 m M , m m } by fixing the other normal vectors in the kth iteration, where l i t e r = m o d ( k , L ) + 1 . First, all the projection vectors { w m , m ( r , l ) | 1 r R , 1 l L , 1 m M , m m } are initialized using uniformly distributed random numbers within 0 to 1. Then, we can get the Lagrangian function for alternating optimization in the kth iteration as follows.
L ( w m , m ( r , l i t e r ) | m , m , r , b m , m | m , m , a i m | i , m , b i m | i , m ) = 1 2 m m r = 1 R w m , m ( r , 1 ) w m , m ( r , L ) F 2 + C × i , m ξ i m i m a i m ( r X i ( y i ) l = 1 L × l w y i , m ( r , l ) + b y i , m r X i ( m ) l = 1 L × l w m , y i ( r , l ) b m , y i 2 + ξ i m ) i , m β i m ξ i m s . t . α i m 0   1 i N   m y i β i m 0   1 i N   m y i α i y i = 0   1 i N β i y i = 0   1 i N ,
where α i m and β i m denote the dual variables, and α i y i and β i y i denote the dummy dual variables. Let the partial derivatives of L ( w m , m ( r , l i t e r ) | m , m , r , b m , m | m , m , α i m | i , m , β i m | i , m ) with respect to the original variables be zeros. We have
L w m , m ( r , l i t e r ) = 0 w m , m ( r , l i t e r ) = 1 l l i t e r L w m , m ( r , l ) T w m , m ( r , l ) × ( { i | y i = m } α i m X i ( m ) l l i t e r L × l w m , m ( r , l ) { i | y i = m } a i m X i ( m ) l l i t e r L × l w m , m ( r , l ) )
b m , m = 0 { i | y i = m } α i m { i | y i = m } α i m = 0 ξ i m = 0 C β i m α i m = 0 .
Substituting Equation (19) and Equation (20) into Equation (18) yields the following Lagrangian dual problem.
m i n α i m | i , m 1 2 ( m m r = 1 R i = 1 N j = 1 N ( c j y i μ ( r , m ) y i a ¯ i , r , m , a ¯ j , r , m 2 c i m c j m μ ( r , m ) m a ¯ i , r , m , a ¯ j , r m + c i m c j m μ ( r , m ) m a ¯ i , r m , a ¯ j , r m ) ) 2 i , m α i m s . t .   { i | y i = m } α i m { i | y i = m } α i m = 0     m m 0 α i m C     1 i N     m y i α i y i = 0     1 i N ,
where
c j m = { 1   i f   y j = m 0   o t h e r w i s e μ ( r , m ) m = 1 l l i t e r L w m , m ( r , l ) T w m , m ( r , l ) a ¯ i , r , m = a i m X i ( y i ) l l i t e r L × l w y i , m ( r , l ) a ¯ i , r m = a i m X i ( m ) l l i t e r L × l w m , y i ( r , l ) .
For convenience, rewrite the dual problem in the matrix form as follows.
m i n a 1 2 a T Q a 2 a T e N × M s . t . { i | y i = m } α i m { i | y i = m } α i m = 0   m m 0 α i m C   1 i N   m y i α i y i = 0   1 i N ,
where Q = m = 1 M r = 1 R [ K 1 ( r ) K 2 ( r , m ) K 2 ( r , m ) T + K 3 ( r , m ) ] , and
e N × M = [ 1 , 1 , , 1 N × M ] K 1 ( r ) ( ( i 1 ) × M + m 1 , ( j 1 ) × M + m 2 ) = c j y i × μ ( r , m 1 ) y i × δ ˜ ( y i , m 2 ) δ ( m 1 , m 2 ) × X i ( y i ) l l i t e r L × l w y i , m 1 ( r , l ) , X j ( y j ) l l i t e r L × l w y j , m 2 ( r , l )
K 2 ( r , m ) ( ( i 1 ) × M + m 1 , ( j 1 ) × M + m 2 ) = c i m 2 c j m δ ˜ ( m , m 2 ) μ ( r , m ) y i × X i ( y i ) l l i t e r L × l w y i , m ( r , l ) , X j ( m 2 ) l l i t e r L × l w m 2 , m ( r , l ) K 3 ( r , m ) ( ( i 1 ) × M + m 1 , ( j 1 ) × M + m 2 ) = c i m c j m δ ( m 1 , m 2 ) × δ ˜ ( m , m 2 ) μ ( r , m ) m 1 × X i ( m 1 ) l l i t e r L × l w m 1 , m ( r , l ) , X j ( m 2 ) l l i t e r L × l w m 2 , m ( r , l ) a = [ a 1 1 , , a 1 M , a 2 1 , , a N M ] ,
where
δ ( m 1 , m 2 ) = { 1   i f   m 1 = m 2 0   o t h e r w i s e .
δ ˜ ( m 1 , m 2 ) = { 0   i f   m 1 = m 2 1   o t h e r w i s e .
It is found from Equation (23) that the dual problem of MCMS-STM in kth iteration is the quadratic optimization in terms of a , which can be solved by classical quadratic programing method (e.g., interior-point method [35]). According to the resulting a , the w m , m ( r , l iter ) can be updated using Equation (19).
Note that the values of the objective function (see Equation (16)) in each iteration form a bounded and nonincreasing sequence. Therefore, there is a finite limit for that sequence. To stop iterating at the proper time, the termination criteria are constructed in Equation (28).
m m r = 1 R l = 1 L ( w m , m ( r , l ) w m , m ( r , l ) ) F 2 δ | | k T i t e r ,
where w m , m ( r , l ) , δ and T i t e r denote the updated w m , m ( r , l ) in last iteration, a pre-established positive number and the threshold for the number of iterations. When the termination criteria are met, output current projection vectors w m , m ( r , l ) . To calculate bias b, according to the K.K.T. conditions [36], including complementary slackness, dual feasibility, and the equivalent constraints in Equation (19) and Equation (20), we obtain
a i m ( r X i ( y i ) l = 1 L × l w y i , m ( r , l ) + b y i , m r X i ( m ) l = 1 L × l w m , y i ( r , l ) b m , y i 2 + ξ i m ) = 0 β i m ξ i m = 0 β i m 0 α i m 0 C b i m a i m = 0 .
Combined with Equation (29), the equations and inequalities in terms of b can be drawn in following:
0 < α i m < C p i m = 0 α i m = C p i m 0 α i m = 0 , y i m p i m 0 ,
where p i m = r X i ( y i ) l = 1 L × l w y i , m ( r , l ) + b y i , m r X i ( m ) l = 1 L × l w m , y i ( r , l ) b m , y i 2 .
When the optimal solution is reached, the above equations and inequalities hold. Therefore, b can be calculated by following linear programming.
m i n b , u m , m , j | u j ( m , m ) | y i = m p i m + u 1 ( m , m ) = 0 y i = m p i m + u 2 ( m , m ) 0   y i = m p i m + u 3 ( m , m ) 0   1 m M   m m   α i m = 0 ,
where u 1 ( m , m ) , u 2 ( m , m ) and u 3 ( m , m ) denote the slacking variables used to maintain the feasibility of linear programming. Using the classical simplex method [37], the Equation (31) can be solved to obtain the optimal b. At present, the training procedure of MCMS-STM has been finished.
After training MCMS-STM, for any test sample X m | m = 1 M , the following decision function is used to predict the corresponding label.
a r g   m a x m # { r X i ( m ) l = 1 L × l w m , m ( r , l ) + b m , m > r X i ( m ) l = 1 L × l w m , m ( r , l ) + b m , m | m m } ,
where operator # { } denotes the number of elements in the set. Note that there may be multiple candidate classes that maximize the Equation (32). For this situation, just like the operation in [20], we simply select the candidate class with the smallest label as the predicted class.

3.3. Acceleration of Training of OVO Version of MCMS-STM

In this section, combining the existing decomposition algorithm [38], a MCMS-STM-oriented decomposition algorithm is presented to break the complex quadratic programming (QP) for dual problem of MCMS-STM (see Equation (23)) into a series of simple analytic QP problems to train the OVO version of MCMS-STM efficiently.
To solve the QP in Equation (23) efficiently, based on the idea of a general decomposition algorithm, the dual variables a are split into working set and non-working set, where variables in the working set are updated, and those in the non-working set are fixed in each iteration. To select the proper variables as the working set, a graph-based constraints model is illustrated in Figure 5 to demonstrate the constraints in Equation (23) clearly.
As shown in Figure 5, there are M × M vertices in the graph, where each vertex indicates a set of dual variables, i.e., a 1 1 = { α i m | 1 i N , y i = 1 , m = 1 } . The constraints in Equation (23) can be described visually via the graph-based constraints model, where the value of dual variables in each vertex should be between 0 and C, and the sum of the values for dual variables contained in the vertices (i,j) is equal to that in vertex (i,j), i.e., the vertex with the same color in Figure 5 should have the same sum. Since updating the variables contained in the vertices of the same color does not affect the feasibility of other variables, it is taken into consideration to select variables from the vertices of the same color as the working set.
First, the zeros values are assigned to all the dual variables as the initial feasible solution of the dual problem in Equation (23). Subsequently, select a working set from the vertices with the same color a i j and a i j . For variables in a i j and a i j , the constraints can be described as follows.
s . t . { k | y k = i } α k j { k | y k = j } α k i = 0 { y k = i | k } y k ( i , j ) α k j + { y k = j | k } y k ( j , i ) α k i = 0 ,
where y k ( j , i ) = 1 and y k ( j , i ) = 1 . It is noticed that the equivalent constraints in Equation (33) are equal to those in C-SVM (see Equation (6)). Therefore, similar to the variables selection strategy in C-SVM [38], the two variables that maximize the following objective function are selected as a working set.
m i n α i 1 j 1 , α i 2 j 2 d 1 f ( a ) α i 1 j 1 + d 2 f ( a ) α i 2 j 2 s . t . { d 1 , d 2 , α i 1 j 1 , α i 2 j 2 } S 1 | | S 2 | | S 3 | | S 4 S 1 = { d 1 = 1 , d 2 = 1 , α i 1 j 1 , α i 2 j 2 | α i 1 j 1 a i j , α i 2 j 2 a i j , α i 1 j 1 < C , α i 2 j 2 > 0 } S 2 = { d 1 = 1 , d 2 = 1 , α i 1 j 1 , α i 2 j 2 | α i 1 j 1 a j i , α i 2 j 2 a j i , α i 1 j 1 > 0 , α i 2 j 2 < C } S 3 = { d 1 = 1 , d 2 = 1 , α i 1 j 1 , α i 2 j 2 | α i 1 j 1 a i j , α i 2 j 2 a j i , α i 1 j 1 < C , α i 2 j 2 < C } S 4 = { d 1 = 1 , d 2 = 1 , α i 1 j 1 , α i 2 j 2 | α i 1 j 1 a i j , α i 2 j 2 a j i , α i 1 j 1 > 0 , α i 2 j 2 > 0 } ,
where f ( a ) = 1 2 a T Q a 2 a T e N × M , and d1, d1 denote the updating direction of variables in working set. The objective function in Equation (34) means selecting variables with the steepest descent direction as the working set. The constraints in Equation (34) ensure the feasibility of variables α i 1 j 1 , α i 2 j 2 in the working set after updating. If m i n α i 1 j 1 , α i 2 j 2 d 1 f ( a ) α i 1 j 1 + d 2 f ( a ) α i 2 j 2 0 , it indicates that there are no proper working variables in a j i and a j i to decrease the f ( a ) , otherwise, select the { α i 1 j 1 , α i 2 j 2 } and [ d 1 , d 2 ] as the working set and the updating direction, respectively. Then, solve the following optimization problem to determine the optimal step size for updating.
m i n d s t e p 1 2 ( a B + d s t e p × [ d 1 , d 2 ] ) T Q B B ( a B + d s t e p × [ d 1 , d 2 ] ) + ( a B + d s t e p × [ d 1 , d 2 ] ) Q B N a N + 1 2 a N Q N N a N 2 ( a B + d s t e p × [ d 1 , d 2 ] ) T e 2 2 a N T e N × M 2 s . t .   0 a B + d s t e p × [ d 1 , d 2 ] C ,
where d s t e p , a B and a N denote the step-size used for updating variables, the variables in working set, and the variables in non-working set, respectively. The [ Q B B , Q B N Q N B , Q N N ] is the permutation of the Q according to the selected working set. To obtain the optimal d s t e p , let the derivative of the objective function be zero. We have
d s t e p = ( [ d 1 , d 2 ] T Q B B [ d 1 , d 2 ] ) 1 × [ d 1 , d 2 ] T ( 2 e 2 [ Q B B , Q B N ] a ) ,
where d s t e p denotes the optimal step size. To avoid the updated a B lie outside the range from 0 to C, the final step size is determined by
d s t e p = d s t e p m a x { d t e m p 1 , d t e m p 2 } d t e m p 1 = { ( α i 1 j 1 + d s t e p × d 1 ) C   i f ( α i 1 j 1 + d s t e p × d 1 ) > C 0   o t h e r w i s e d t e m p 2 = { ( α i 2 j 2 + d s t e p × d 2 )   i f ( α i 2 j 2 + d s t e p × d 2 ) < 0 0   o t h e r w i s e ,
where d s t e p denotes the final step size. Use a B = a B + d s t e p × [ d 1 , d 2 ] to update the variables in working set. Then, use Equation (34) to re-select the working set for updating in the next iteration. This process is performed alternatively until the following terminal criteria are met, i.e., m i n α i 1 j 1 , α i 2 j 2 d 1 f ( a ) α i 1 j 1 + d 2 f ( a ) α i 2 j 2 0 for any vertex ( i , j ) , i j .
After the terminal criteria are satisfied, the current a is optimal. The decomposition algorithm for solving the dual problem of the OVO version of the MCMS-STM is summarized in Algorithm 1.
Algorithm 1. Decomposition algorithm of solving the dual problem of OVO version of the MCMS-STM.
Input: the Q of the dual problem of MCMS-STM and the labels y i   1 i N .
Output: optimal a
for vertex in ( i , j ) , i j
Step 1: Select working set using Equation (34).
If m i n α i 1 j 1 , α i 2 j 2 d 1 f ( a ) α i 1 j 1 + d 2 f ( a ) α i 2 j 2 0 , continue; otherwise, perform Step 2.
Step 2: Calculate the step size using Equations (36) and (37), and update the variables in working set.
Step 3: If the terminal criteria are met, output the a as optimal solution, and return.
end

4. Discussion of the Multiclass Classification Mechanism Used in MCMS-STM

To analyze the multiclass classification mechanism used in MCMS-STM compared with the existing OVO and OVR strategy, a detailed discussion is carried out as follows.

4.1. Discussion of OVR Version of MCMS-STM Compared with OVR Strategy Based STM

Considering the hard margin (i.e., the C = ) based C STM [25] with OVR strategy, it is required to construct M classifiers, where the mth classifier is used to separate the class m from the other M-1 classes, as shown in the following optimization problem.
m i n w m ( l ) | l , b 1 2 w m ( 1 ) w m ( L ) F 2 s . t . X i l = 1 L × l w m ( l ) + b m 1   y i = m X i l = 1 L × l w m ( l ) + b m 1   y i m ,
where w m ( 1 ) w m ( L ) denotes the normal orientation for the mth hyperplane.
Then, let R = 1 and C = for OVR version of MCMS-STM (see Equation (17)), the corresponding optimization problem can be converted to Equation (39).
m i n w m ( l ) | m , l , b 1 2 m = 1 M w m ( 1 ) w m ( L ) F 2 s . t . X i l = 1 L × l w y i ( l ) + b y i X i l = 1 L × l w m ( l ) + b m + 2 m y i   1 i N
By adding the two inequalities in Equation (38), it is easy to find that the feasible solution of Equation (38) is also the feasible solution of Equation (39). Thus, the solution of Equation (38) is the feasible solution of Equation (39), but not necessarily the optimal one. That means the MCMS-STM may obtain the optimal solution with a larger classification margin compared with C STM using OVR strategy, i.e., the better generalization ability.

4.2. Discussion of OVO Version of MCMS-STM Compared with OVO Strategy Based STM

When using OVO strategy [20] based C STM to process M class classification problems, M ( M 1 ) 2 classifiers are constructed, where each one is used to separate specific two classes. For separating classes m and m, the corresponding optimization problem of C STM is shown below.
m i n w m m ( l ) | l , b 1 2 w m , m ( 1 ) w m , m ( L ) F 2 s . t . X i l = 1 L × l w m , m ( l ) + b m , m 1   y i = m X j l = 1 L × l w m , m ( l ) + b m , m 1   y j = m ,
where w m ( 1 ) w m ( L ) b m m denote the normal orientation of the hyperplane and bias of the hyperplane, respectively. For OVO version of MCMS-STM (see Equation (16)), let the R = 1 and C = . The corresponding optimization problem is shown below.
m i n w m , m ( l ) | m , m , l , b 1 2 m = 1 M m ¹ m w m , m ( 1 ) w m , m ( L ) F 2 s . t . X i l = 1 L × l w m , m ( l ) + b m , m X i l = 1 L × l w m , m ( l ) + b m , m + 2         y i = m   m m   1 i N
For any feasible solution of w m m ( 1 ) and b m , m in Equation (40), let w m m ( 1 ) = w m , m ( 1 ) and b m , m = b m , m , we can obtain w m m ( 1 ) , b m , m , w m m ( 1 ) and b m , m as the feasible solution of Equation (41). On the contrary, for the constraints in Equation (41), we obtain:
X i l = 1 L × l w m , m ( l ) + b m , m X i l = 1 L × l w m , m ( l ) + b m , m + 2   y i = m X j l = 1 L × l w m , m ( l ) + b m , m X j l = 1 L × l w m , m ( l ) + b m , m + 2   y j = m X i l = 1 L × l ( w m , m ( l ) w m , m ( l ) ) + b m , m b m , m 1 1   y i = m X j l = 1 L × l ( w m , m ( l ) w m , m ( l ) ) + b m , m b m , m 1 1   y j = m
According to Equation (42), the constraints in Equation (41) can be converted to the same form as the constraints in Equation (40). Note that the feasible solutions of different methods can be transformed into each other using simple operations. Therefore, the MCMS-STM has the OVO interpretation.
Through the above analysis, the MCMS-STM present both the OVR and OVO interpretations under the specific parameter setting. More importantly, compared with the OVR and OVO strategies that need to learn multiple classification hyperplanes separately, a remarkable advantage is that the MCMS-STM can learn the multiple classification hyperplanes simultaneously to mine the correlation between multiple classes.

5. Experiments and Analysis

To demonstrate the superiority of the MCMS-STM for multiclass multiscale object recognition, two datasets containing image slices of multiclass multiscale objects are used to evaluate the performance of the proposed MCMS-STM, and the detailed information of datasets is introduced as follows.
(1)
Dataset 1: To verify the performance of MCMS-STM for multiclass multiscale airplane classification, the RSIs containing 218 airplanes with five types are collected from Google Earth service with a spatial resolution of 0.5 m and R, G, and B spectral bands. Then, using two image resizing operations, these 218 airplanes are cut separately from RSIs to build two slice sets. For slices set 1 generated from image resizing operation 1, the slices are cut according to the type of contained objects and then resized to a fixed size using a bilinear interpolation method. For slices set 2 generated from image resizing operation 2, these slices are cut at a size large enough (i.e., 120 × 120 ) so that all the types of objects are contained completely in the corresponding slice. These slices contain various backgrounds, and the contained multiclass airplanes present different orientations and sizes. Some representative slices from two slices sets of dataset 1 are displayed in Figure 6.
(2)
Dataset 2: The HRSC-2016 [39] dataset contains 1070 harbor RSIs with R, G, and B spectral bands collected from Google Earth service. To evaluate the performance of MCMS-STM for multiscale object recognition, 342 ships with five types are sliced from HRSC-2016 whose spatial resolution is equal to 1.07 m. Similar to dataset 1, these slices are cut by two image resizing operations to form two slices set. For slices set 1 generated from image resizing operation 1, the slices are cut according to the type of contained objects and then resized to a fixed size using a bilinear interpolation method. For slices set 2 generated from image resizing operation 2, these slices are cut at a size large enough (i.e., 770 × 170 ) to contain airplanes with different types. Some image slices with different types of ships in two slices sets of dataset 2 are shown in Figure 7.
The experiments are composed of four parts. In Section 5.1, the impact of parameter setting is analyzed. In Section 5.2, the affection of image resizing preprocessing for the recognition results is examined. In Section 5.3, the efficiency of the proposed decomposition algorithm is verified compared with the typical interior-point method and active-set method. In Section 5.4, the recognization accuracy of MCMS-STM is evaluated compared with typical SVM and STM methods using dataset 1 and dataset 2. In Section 5.5, the performance of MCMS-STM is further evaluated compared with typical deep learning methods. All the simulations are running on the computer with Windows 10 operating system and Intel i7-7700CPU at 3.6 G Hz.
The code of MCMS-STM can be found at https://github.com/supergt3/MCMS-STM (accessed on 23 November 2021).

5.1. Analysis of the Impact of Parameter Setting on Classification Performance

The main parameters in MCMS-STM include the R and C, which denote the CP rank of the projection tensor and the regularization parameter, respectively. To demonstrate the impact of parameter setting, the N-fold cross-validation method is adopted to partition the slices set 1 of dataset 1 into N subsets, and test the recognization accuracy using the one subset according to the trained classifier using the remaining N-1 subsets. This process is then repeated N times, with each of the N subsets used exactly once as the test samples. Through adjusting the values of R from { 1 , 2 , 3 , , 10 } and C from { 10 0 , 10 1 , 10 2 } , the obtained classification accuracies of different versions of MCMS-STM are shown in Figure 8.
From Figure 8, it is observed that the recognization accuracy of MCMS-STM is affected by different parameter settings. In detail, it is seen that the increase of C may have a positive or negative effect on recognization accuracy for different values of R. Therefore, it is difficult to set an effective C in advance. In addition, for the case of using the same C, MCMS-STM with R < 3 obtains smaller recognization accuracy than that with R 3 , because the projection tensor with a small R is difficult to define an accurate classification hyperplane. Through observing Figure 8, it is found that the best recognization accuracy is obtained by OVO version of the MCMS-STM. The reason is probably that OVO version of the MCMS-STM may obtain a larger classification margin compared with OVR version of the MCMS-STM.

5.2. Analysis of the Impact of Image Resizing on Classification Performance

One of the significant advantages of the MCMS-STM method is that it can use multiscale projection tensors to effectively classify objects with different sizes, avoiding the loss of objects’ scale information caused by image resizing. To highlight the superiority of MCMS-STM, two slices sets obtained by different image resizing methods are utilized to examine the recognization accuracy of the MCMS-STM with multiscale projection tensors and single-scale projection tensors. The experiments are implemented in three control groups. In detail, for control group 1, the MCMS-STM with single-scale projection tensors and slices set 1 are utilized to obtain the classification results. For control group 2, the MCMS-STM with single-scale projection tensors and slices set 2 are utilized to obtain the classification results. For control group 3, the MCMS-STM with multiscale projection tensors and the slices set 1 are utilized to obtain the classification results. To use multiscale projection tensors, it is needed to decide the sizes of multiscale tensors. In the experiments, the sizes of slices for five types of ships are set to 770 × 170 , 360 × 50 , 400 × 70 , 550 × 140 and 290 × 70 , respectively, according to the average sizes of objects for the different types. The obtained classification results for MCMS-STM under different conditions are plotted in Figure 9.
The accuracy in Figure 9 denotes the largest accuracy with parameters C and R selected from C = { 10 0 , 10 1 , 10 2 } and R = { 2 , 4 , 6 , 8 , 10 } under different versions of MCMS-STM. For MCMS-STM, the OVO version indicates the larger classification margin, but at the same time, it will also cause misclassification because the decision function (see Equation (32)) may generate multiple results. Therefore, for dataset 2, it is observed that the recognization accuracy of OVR version of MCMS-STM is higher than that of OVO version of MCMS-STM. In addition, note that the resizing operation used to generate slices set 1 will lead to the loss of object’s scale information, and the resizing operation used to generate slices set 2 will bring excessive background interferences, i.e., the two resizing operations have their own disadvantages. In comparison, the multiscale projection tensors used in MCMS-STM can maintain the scale information of objects while avoiding excessive background interferences. Therefore, it is observed that the recognization accuracy of MCMS-STM without resizing operations is better than using MCMS-STM with resizing operations. The above analysis concludes that the MCMS-STM with multiscale projection tensors is effective for multiscale object recognition tasks.

5.3. Evaluation of the Performance of the Decomposition Algorithm for Training MCMS-STM

To verify the efficiency of the proposed decomposition algorithm for training OVO version of MCMS-STM, the experiments are conducted on dataset 1 to evaluate the time consumption of training MCMS-STM and classical multiclass SVM using different optimization solving algorithms, including the proposed decomposition algorithm, interior-point method [35] and active-set method [40], under different numbers of training samples, as shown in Figure 10.
The time consumption in Figure 10 represents the performing time to solve the dual problem of the MCMS-STM or multiclass SVM once. From Figure 10, when using the same optimization algorithm, i.e., interior-point algorithm or active-set algorithm, it can be seen that the time consumption of solving dual problem of MCMS-STM is similar to that of solving dual problem of multiclass SVM. In addition, it is seen that the time consumption of the interior-point algorithm is less than that of the active-set algorithm under different sizes of training sample sets. Remarkably, it can be seen that the proposed decomposition algorithm significantly reduces the time consumption compared with the interior-point method and active-set method for training MCMS-STM, especially for the training set with a larger size. It indicates the efficiency of the proposed decomposition algorithm for training the MCMS-STM, especially for the training set with a large size.

5.4. Evaluation of the Performance of MCMS-STM Compared with Existing SVM and STM Methods

In this section, the experiments are conducted on dataset 1 and dataset 2 to evaluate the performance of MCMS-STM for multiclass multiscale airplane classification and ship classification compared with typical SVM and STM methods. To avoid the classification results from being affected by different feature extraction algorithms, the image slices represented as tensors are used as input directly for all the classification methods. In detail, according to the size of different objects, the dimensions of multiscale tensors for airplanes with five types are set to for 100 × 100 × 3 , 80 × 80 × 3 , 120 × 120 × 3 , 70 × 70 × 3 and 80 × 80 × 3 , respectively, and those for ships with five types are set to 770 × 170 × 3 , 360 × 50 × 3 , 400 × 70 × 3 , 550 × 140 × 3 and 290 × 70 × 3 , respectively. The first order, second order and third order for multiscale tensors denote the horizontal spatial order, vertical spatial order, and spectral order, respectively. For the comparison methods, two slice sets from dataset 1 and dataset 2 obtained by different slice cropping methods are utilized to obtain the classification results under two types of image resizing operations, respectively. In addition, note that the SVM and STM methods can only deal with bi-category classification problems. Therefore, the OVO and OVR strategies are introduced to perform multiclass classifications for binary SVM and STM methods indirectly. Considering that the SVM method cannot process tensors directly, the vectorization operation is utilized to convert image slice to vector to comply with its input requirement. Then, all the methods select parameters from C = { 10 0 , 10 1 , 10 2 } , ν = { 0.1 , 0.2 , 0.3 , 0.4 } and R = { 2 , 4 , 6 , 8 , 10 } that corresponds to the best classification results to obtain the final recognization accuracy. Using the N-fold cross-validation, the accuracies of MCMS-STM for different classification tasks under different conditions are displayed in Table 3, Table 4, Table 5, and Table 6, respectively.
In these tables, the bolding accuracy indicates the best results, and the notation C S V M (OVO) 1 denotes using OVO strategy based C S V M under slice set 1. Overall, note that the classification results for 10-fold cross-validation are better than those for 5-fold cross-validation, mainly because the 10-fold cross-validation indicates that more samples are used for training. From these tables, it is observed that applying OVO strategy based comparison methods present better classification results than applying OVR strategy based comparison methods because the OVO is a competitive multiclass classification strategy compared with OVR strategy. For N-fold cross-validation, by comparing the classification accuracies between N = 5 and N = 10, it can be found that there is a minor difference between different N for STM methods and a significant difference between different N for SVM methods. The reason is that the SVM method is embedded with much more parameters, and thus it is easy to improve performance as the number of samples increases. In addition, it can be seen that the recognization accuracy of multiclass SVM is better than most comparison methods under OVR or OVO strategy. Since the two resizing operations have their own advantages and disadvantages, i.e., the first type of resizing operation can reduce the impact of background interferences, while it loses the scale information of object, and the image the second type of resizing operation can maintain the scale information of objects, while it brings much more interferences for objects with small size, it is observed that using the slices set 1 obtains the largest accuracy for ship recognization under 10-fold cross-validation (see Table 6), and using the slices set 2 obtains the largest accuracy for airplane recognization under 10-fold cross-validation (see Table 4). That means two image resizing operations present comparable results. Moreover, it is seen that the optimal R that corresponds to the largest accuracy is located at the range from 4 to 8, as the projection with small R is difficult to describe the effective classification hyperplane, and the projection tensor with large R is easy to over-fit the training samples. It is worth observing that the MCMS-STM gets the best classification results among all the methods for different classification tasks. In detail, MCMS-STM gets 89.5% and 91.4% recognization accuracy for airplane recognization and ship recognization, respectively, while the largest accuracies of comparison methods for airplane recognization and ship recognization are equal to 88.6% and 89.9%, respectively. Therefore, it is concluded that the MCMS-STM is more effective for multiclass multiscale object recognition using remote sensing images compared with the comparison methods.
In addition, to verify whether the MCMS-STM makes improvements from a statistical view, the statistical test is applied to analyze the performance of the MCMS-STM against the SVM and STM methods.
For convenience, the most commonly applied test method, i.e., right-tailed significance t-test [41], is used to test the null hypothesis H0 that the difference of accuracy between MCMS-STM and the competitor is equal to zero. On the other hand, the alternate hypothesis H1 is that the difference of accuracy between MCMS-STM and the competitor is greater than zero.
To obtain effective results, half of samples in dataset 1 are used to train a classifier, and the rest of the samples in dataset 1 are used to evaluate the recognization accuracy of trained classifier for five classes of airplanes. The same processing is performed for dataset 2. In this way, ten paired observations are obtained for the sign test. According to obtained ten paired observations, use paired-sample right-tailed significance t-test [41], and the resulting p-value for MCMS-STM against the other methods is shown in Table 7.
Where positive, negative, ties, and p-value denote the number of times that the MCMS-STM outperforms the comparison method, the number of times that the comparison method outperforms the MCMS-STM, the number of times that both the MCMS-STM and comparison method obtain the same results, and the probability of the observation under H0. From Table 7, it is found that all the p-values are less than 0.05. Therefore, the null hypothesis can be rejected at the 0.05 significance level for all the comparison methods. It indicates that the proposed MCMS-STM outperforms the existing SVM and STM methods from the view of statistics.

5.5. Evaluation of the Performance of MCMS-STM Compared with Deep Learning Methods

Compared with deep learning methods relying on a large number of training samples, the proposed MCMS-STM can recognize multiclass multiscale objects effectively using a small number of training samples. To verify this advantage, the experiments are conducted on datasets 1 and 2 to compare the recognition accuracies of the MCMS-STM with two deep learning methods, i.e., GhostNet [42] and ResNeXt networks [43], of which the GhostNet utilizes the Ghost module to improve the feature representation power of the conventional convolutional layer and the ResNeXt improved from deep residual networks can exploit the aggregated residual transformations to mine the effective features to improve the recognization results. When training the GhostNet and ResNeXt, the learning rate and the batch size are set to 0.001 and 128, respectively. Since these two deep learning methods can only deal with slices with a fixed size, the slice sets 1 from two datasets are used to examine their performances. For MCMS-STM, the multiscale projection tensors are exploited, and the detailed sizes of projection tensors can be found in Section 5.4. To compare the recognization results of the proposed OVO version of MCMS-STM with the deep learning methods under different sizes of training sets, different proportions of samples in the dataset are selected as training samples and the rest as test samples, and the obtained recognization results are shown in Table 8 and Table 9.
Where the notation p in Table 8 and Table 9 denotes the proportion of the training samples occupying the data set. From the above tables, it is observed that the recognization accuracy is generally improved with the increase of training set size because more training samples can ensure sufficient training of the classifier. In addition, it is seen that the GhostNet obtains better results than ResNeXt under the same p. It indicates that the GhostNet can extract more effective features using its Ghost module. Since the deep learning methods rely on a large number of training samples, they obtain worse results under a small p. Remarkably, it is worth noting that the proposed MCMS-STM obtains significantly better results compared to deep learning methods, especially when the training samples are small. This funding implies the superiority of the proposed MCMS-STM for the small sample case compared with the deep learning methods.

6. Conclusions

To classify the multiclass multiscale objects in RSIs effectively, the MCMS-STM is proposed incorporating multiple hyperplanes defined by multiscale projection tensors to map the input of object slices with different sizes to multiclass class space, getting rid of the conventional image resizing operation. The main contributions of this paper can be summarized in the three folds below.
(1)
To achieve multiclass classifications for objects in RSIs, the MCMS-STM is proposed to learn multiple hyperplanes defined by rank-R projection tensors simultaneously to map input represented as tensor into class space. This new multiclass classification mechanism makes it easy to construct the corresponding decomposition algorithm to accelerate the training of the MCMS-STM and enables the classifier to present OVO and OVR interpretations, ensuring the MCMS-STM can deal with different classifications tasks effectively.
(2)
To identify multiscale objects in RSIs, instead of the conventional image resizing operation, according to the object position obtained from detection results, multiple slices of different sizes are extracted to describe the contained object with unknown class, and multidimensional classification hyperplanes are established to separate input of multiple slices with different sizes to achieve cross-scale object recognition. This multiscale classification mechanism can avoid the loss of scale information and reduce the impact of background interferences caused by conventional image resizing preprocessing.
(3)
To accelerate the training of OVO version of MCMS-STM, combining graph-based analysis, a decomposition algorithm is proposed to break the dual problem of OVO version of MCMS-STM as a series of small sub-optimizations, reducing the time consumption caused by the large Q. It ensures the MCMS-STM can be trained efficiently for more samples and classes.
The future work will focus on extending the MCMS-STM to adapt to the objects with different orientations and constructing a universal decomposition algorithm to accelerate the training of different versions of MCMS-STM.

Author Contributions

Methodology, T.G.; project administration, H.C.; data curation, W.C. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by the National Natural Science Foundation of China under Grant 61771170.

Institutional Review Board Statement

Not applicable.

Informed Consent Statement

Not applicable.

Data Availability Statement

Not applicable.

Conflicts of Interest

The authors declare no conflict of interest.

Appendix A

We give the procedure of training OVR version of MCMS-STM in this section. Similar to the solving method of OVO version of MCMS-STM, an alternating optimization scheme is adopted for training of OVR version of MCMS-STM, i.e., optimize { w m ( r , l i t e r ) | 1 r R , 1 m M } by fixing the other normal vectors in the kth iteration, where l i t e r = mod ( k , L ) + 1 . First, all the projection vectors { w m ( r , l ) | 1 r R , 1 l L , 1 m M } are initialized using uniformly distributed random numbers within 0 to 1. Then, get the Lagrangian function for alternating optimization in kth iteration as follows.
L ( w m ( r , l i t e r ) | m , r , b m | m , a i m | i , m , b i m | i , m ) = 1 2 m = 1 M r = 1 R w m ( r , 1 ) w m ( r , L ) F 2 + C × i , m ξ i m i m a i m ( r X i ( y i ) l = 1 L × l w y i ( r , l ) + b y i r X i ( m ) l = 1 L × l w m ( r , l ) b m 2 + x i m ) i , m β i m ξ i m s . t . α i m 0   1 i N   m y i β i m 0   1 i N   m y i α i y i = 0   1 i N β i y i = 0   1 i N .
Let the partial derivatives of ( w m ( r , l i t e r ) | m , r , b m | m , α i m | i , m , β i m | i , m ) with respect to the original variables be zeros. We have
L w m ( r , l i t e r ) = 0 w m ( r , l i t e r ) = 1 l ¹ l i t e r L w m ( r , l ) T w m ( r , l ) × ( { i | y i = m } m a i m X i ( m ) l ¹ l i t e r L × l w m ( r , l ) i a i m X i ( m ) l ¹ l i t e r L × l w m ( r , l ) )
b m = 0 { i | y i = m } m α i m i α i m = 0 ξ i m = 0 C β i m α i m = 0 .
Substituting Equations (A2) and (A3) into Equation (A1) yields the following Lagrangian dual problem.
m i n α i m | i , m 1 2 ( m = 1 M r = 1 R i = 1 N j = 1 N ( c j y i μ ( r , y i ) a ^ i , r , a ^ j , r 2 c i m μ ( r , y i ) a ^ i , r , a ^ j , r m + c i y j μ ( r , m ) a ^ i , r m , a ^ j , r m ) ) 2 i , m α i m s . t .   { i | y i = m } α i m { i | y i = m } α i m = 0   m m 0 α i m C   1 i N   m y i α i y i = 0   1 i N
where
μ ( r , m ) = 1 l l i t e r L w m ( r , l ) T w m ( r , l ) a ^ i , r = m a i m X i ( y i ) l l i t e r L × l w y i ( r , l ) a ^ i , r m = a i m X i ( m ) l l i t e r L × l w m ( r , l ) .
Using quadratic programing method (e.g., interior-point method [35]), the optimization problem in Equation (A4) can be solved. Then, according to the resulting α , the w m h ( r , l i t e r ) can be updated using Equation (A2). To stop iterating at the proper time, the termination criteria are constructed in Equation (A6).
m = 1 M r = 1 R l = 1 L ( w m ( r , l ) w m ( r , l ) ) F 2 δ || k T i t e r ,
where w m ( r , l ) denotes the updated w m ( r , l ) in the last iteration. To calculate bias b, we have
α i m ( r X i ( y i ) l = 1 L × l w y i ( r , l ) + b y i r X i ( m ) l = 1 L × l w m ( r , l ) b m 2 + x i m ) = 0 β i m ξ i m = 0 β i m ³ 0 a i m ³ 0 C β i m α i m = 0 .
Then, the equations and inequalities in terms of b can be drawn in following:
0 < α i m < C p ¯ i m = 0 α i m = C p ¯ i m 0 α i m = 0 , y i m p ¯ i m 0 ,
where p ¯ i m = r X i ( y i ) l = 1 L × l w y i ( r , l ) + b y i - r X i ( m ) l = 1 L × l w m ( r , l ) - b m - 2 .
When the optimal solution is reached, the above equations and inequalities hold. Therefore, b can be calculated by following linear programming using the classical simplex method [37].
m i n b , u m , m , j | u j ( m , m ) | y i = m p ¯ i m + u 1 ( m , m ) = 0   y i = m p ¯ i m + u 2 ( m , m ) 0   y i = m p ¯ i m + u 3 ( m , m ) 0   1 m M   m m   α i m = 0 ,
After training MCMS-STM, for any test sample X m | m = 1 M , the following decision function is used to predict the corresponding label.
a r g m a x m { r X i ( m ) l = 1 L × l w m ( r , l ) + b m } .

References

  1. Girshick, R. Fast r-cnn. In Proceedings of the IEEE International Conference on Computer Vision, Santiago, Chile, 13–16 December 2015; pp. 1440–1448. [Google Scholar]
  2. Ren, S.; He, K.; Girshick, R.; Sun, J. Faster r-cnn: Towards real-time object detection with region proposal networks. Adv. Neural Inf. Process. Syst. 2015, 28, 91–99. [Google Scholar] [CrossRef] [Green Version]
  3. Zhou, M.; Zou, Z.; Shi, Z.; Zeng, W.-J.; Gui, J. Local attention networks for occluded airplane detection in remote sensing images. IEEE Geosci. Remote Sens. Lett. 2019, 17, 381–385. [Google Scholar] [CrossRef]
  4. Pang, J.; Li, C.; Shi, J.; Xu, Z.; Feng, H. 2 -CNN: Fast Tiny Object Detection in Large-scale Remote Sensing Images. IEEE Trans. Geosci. Remote Sens. 2019, 57, 5512–5524. [Google Scholar] [CrossRef] [Green Version]
  5. Liu, W.; Anguelov, D.; Erhan, D.; Szegedy, C.; Reed, S.; Fu, C.-Y.; Berg, A.C. Ssd: Single shot multibox detector. In Proceedings of the European Conference on Computer Vision, Amsterdam, The Netherlands, 8–16 October 2016; pp. 21–37. [Google Scholar]
  6. Redmon, J.; Divvala, S.; Girshick, R.; Farhadi, A. You only look once: Unified, real-time object detection. In Proceedings of the IEEE conference on computer vision and pattern recognition, Las Vegas, NV, USA, 27–30 June 2016; 2016; pp. 779–788. [Google Scholar]
  7. Dadsetan, S.; Pichler, D.; Wilson, D.; Hovakimyan, N.; Hobbs, J. Superpixels and Graph Convolutional Neural Networks for Efficient Detection of Nutrient Deficiency Stress from Aerial Imagery. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 19–25 June 2021; pp. 2950–2959. [Google Scholar]
  8. Hong, D.; He, W.; Yokoya, N.; Yao, J.; Gao, L.; Zhang, L.; Chanussot, J.; Zhu, X. Interpretable hyperspectral artificial intelligence: When nonconvex modeling meets hyperspectral remote sensing. IEEE Geosci. Remote Sens. Mag. 2021, 9, 52–87. [Google Scholar] [CrossRef]
  9. Hong, D.; Gao, L.; Yokoya, N.; Yao, J.; Chanussot, J.; Du, Q.; Zhang, B. More diverse means better: Multimodal deep learning meets remote-sensing imagery classification. IEEE Trans. Geosci. Remote Sens. 2020, 59, 4340–4354. [Google Scholar] [CrossRef]
  10. Long, Y.; Gong, Y.; Xiao, Z.; Liu, Q. Accurate object localization in remote sensing images based on convolutional neural networks. IEEE Trans. Geosci. Remote Sens. 2017, 55, 2486–2498. [Google Scholar] [CrossRef]
  11. Cheng, G.; Zhou, P.; Han, J. Learning rotation-invariant convolutional neural networks for object detection in VHR optical remote sensing images. IEEE Trans. Geosci. Remote Sens. 2016, 54, 7405–7415. [Google Scholar] [CrossRef]
  12. Nguyen, V.D.; Tran, D.T.; Byun, J.Y.; Jeon, J.W. Real-time vehicle detection using an effective region proposal-based depth and 3-channel pattern. IEEE Trans. Intell. Transp. Syst. 2018, 20, 3634–3646. [Google Scholar] [CrossRef]
  13. Chen, H.; Zhao, J.; Gao, T.; Chen, W. Fast airplane detection with hierarchical structure in large scene remote sensing images at high spatial resolution. In Proceedings of the IGARSS 2018-2018 IEEE International Geoscience and Remote Sensing Symposium, Valencia, Spain, 22–27 July 2018; pp. 4846–4849. [Google Scholar]
  14. An, Z.; Shi, Z.; Teng, X.; Yu, X.; Tang, W. An automated airplane detection system for large panchromatic image with high spatial resolution. Optik 2014, 125, 2768–2775. [Google Scholar] [CrossRef]
  15. Zhang, L.; Zhang, Y. Airport detection and aircraft recognition based on two-layer saliency model in high spatial resolution remote-sensing images. IEEE J. Sel. Top. Appl. Earth Obs. Remote Sens. 2016, 10, 1511–1524. [Google Scholar] [CrossRef]
  16. Jing, M.; Zhao, D.; Zhou, M.; Gao, Y.; Jiang, Z.; Shi, Z. Unsupervised oil tank detection by shape-guide saliency model. IEEE Geosci. Remote Sens. Lett. 2018, 16, 477–481. [Google Scholar] [CrossRef]
  17. Cortes, C.; Vapnik, V. Support-vector networks. Mach. Learn. 1995, 20, 273–297. [Google Scholar] [CrossRef]
  18. Schölkopf, B.; Smola, A.J.; Williamson, R.C.; Bartlett, P.L. New support vector algorithms. Neural Comput. 2000, 12, 1207–1245. [Google Scholar] [CrossRef]
  19. Suykens, J.A.; Vandewalle, J. Least squares support vector machine classifiers. Neural Process. Lett. 1999, 9, 293–300. [Google Scholar] [CrossRef]
  20. Chang, C.-C.; Lin, C.-J. LIBSVM: A library for support vector machines. ACM Trans. Intell. Syst. Technol. (TIST) 2011, 2, 1–27. [Google Scholar] [CrossRef]
  21. Godbole, S.; Sarawagi, S.; Chakrabarti, S. Scaling multi-class support vector machines using inter-class confusion. In Proceedings of the Eighth ACM SIGKDD International Conference on Knowledge Discovery and Data Mining, Edmonton, AB, Canada, 23–26 July 2002; pp. 513–518. [Google Scholar]
  22. Weston, J.; Watkins, C. Support vector machines for multi-class pattern recognition. In Proceedings of the ESANN 1999, 7th European Symposium on Artificial Neural Networks, Bruges, Belgium, 21–23 April 1999; pp. 219–224. [Google Scholar]
  23. Iosifidis, A.; Gabbouj, M. Multi-class support vector machine classifiers using intrinsic and penalty graphs. Pattern Recognit. 2016, 55, 231–246. [Google Scholar] [CrossRef]
  24. De Lima, M.D.; Costa, N.L.; Barbosa, R. Improvements on least squares twin multi-class classification support vector machine. Neurocomputing 2018, 313, 196–205. [Google Scholar] [CrossRef]
  25. Tao, D.; Li, X.; Wu, X.; Hu, W.; Maybank, S.J. Supervised tensor learning. Knowl. Inf. Syst. 2007, 1, 1–42. [Google Scholar] [CrossRef]
  26. Hao, Z.; He, L.; Chen, B.; Yang, X. A linear support higher-order tensor machine for classification. IEEE Trans. Image Process. 2013, 22, 2911–2920. [Google Scholar]
  27. Chen, Z.-Y.; Fan, Z.-P.; Sun, M. A multi-kernel support tensor machine for classification with multitype multiway data and an application to cross-selling recommendations. Eur. J. Oper. Res. 2016, 255, 110–120. [Google Scholar] [CrossRef]
  28. Kotsia, I.; Guo, W.; Patras, I. Higher rank support tensor machines for visual recognition. Pattern Recognit. 2012, 45, 4192–4203. [Google Scholar] [CrossRef]
  29. Kotsia, I.; Patras, I. Support tucker machines. In Proceedings of the CVPR 2011, Colorado Springs, CO, USA, 20–25 June 2011; pp. 633–640. [Google Scholar]
  30. Ma, Z.; Yang, L.T.; Zhang, Q. Support Multimode Tensor Machine for Multiple Classification on Industrial Big Data. IEEE Trans. Ind. Inform. 2020, 17, 3382–3390. [Google Scholar] [CrossRef]
  31. Hsu, C.-W.; Lin, C.-J. A comparison of methods for multiclass support vector machines. IEEE Trans. Neural Netw. 2002, 13, 415–425. [Google Scholar] [PubMed] [Green Version]
  32. Xie, J.; He, N.; Fang, L.; Plaza, A. Scale-free convolutional neural network for remote sensing scene classification. IEEE Trans. Geosci. Remote Sens. 2019, 57, 6916–6928. [Google Scholar] [CrossRef]
  33. Kolda, T.G.; Bader, B.W. Tensor decompositions and applications. SIAM Rev. 2009, 51, 455–500. [Google Scholar] [CrossRef]
  34. Savicky, P.; Vomlel, J. Exploiting tensor rank-one decomposition in probabilistic inference. Kybernetika 2007, 43, 747–764. [Google Scholar]
  35. Momoh, J.A.; Guo, S.; Ogbuobiri, E.; Adapa, R. The quadratic interior point method solving power system optimization problems. IEEE Trans. Power Syst. 1994, 9, 1327–1336. [Google Scholar] [CrossRef]
  36. Boyd, S.; Boyd, S.P.; Vandenberghe, L. Convex Optimization; Cambridge University Press: Cambridge, UK, 2004. [Google Scholar]
  37. Brearley, A.L.; Mitra, G.; Williams, H.P. Analysis of mathematical programming problems prior to applying the simplex algorithm. Math. Program. 1975, 8, 54–83. [Google Scholar]
  38. Chang, C.-C.; Lin, C.-J. Trainings nu-support vector classifiers: Theory and algorithms. Neural Comput. 2001, 13, 2119–2147. [Google Scholar] [CrossRef]
  39. Liu, Z.; Wang, H.; Weng, L.; Yang, Y. Ship rotated bounding box space for ship extraction from high-resolution optical satellite images with complex backgrounds. IEEE Geosci. Remote Sens. Lett. 2016, 13, 1074–1078. [Google Scholar] [CrossRef]
  40. Bartlett, R.A.; Biegler, L.T. QPSchur: A dual, active-set, Schur-complement method for large-scale and structured convex quadratic programming. Optim. Eng. 2006, 7, 5–32. [Google Scholar] [CrossRef] [Green Version]
  41. Rohatgi, V.K.; Saleh, A.M.E. An Introduction to Probability and Statistics; John Wiley & Sons: Hoboken, NJ, USA, 2015. [Google Scholar]
  42. Han, K.; Wang, Y.; Tian, Q.; Guo, J.; Xu, C. GhostNet: More Features from Cheap Operations. In Proceedings of the 2020 IEEE/CVF Conference on Computer Vision and Pattern Recognition, Virtual, 14–29 June 2020. [Google Scholar]
  43. Xie, S.; Girshick, R.; Dollár, P.; Tu, Z.; He, K. Aggregated residual transformations for deep neural networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Honolulu, HI, USA, 21–26 July 2017; pp. 1492–1500. [Google Scholar]
Figure 1. Examples of two types of image normalization methods.
Figure 1. Examples of two types of image normalization methods.
Remotesensing 14 00196 g001
Figure 2. Illustration of objects classification for MCMS-STM.
Figure 2. Illustration of objects classification for MCMS-STM.
Remotesensing 14 00196 g002
Figure 3. Illustration of classification hyperplane defined by rank-1 projection tensor and rank-R projection tensor. The bule solid line denotes the hyperplane defined by rank-1 or rank-R tensor, and the blue dotted line denotes the hyperplane defined by rank-R tensor.
Figure 3. Illustration of classification hyperplane defined by rank-1 projection tensor and rank-R projection tensor. The bule solid line denotes the hyperplane defined by rank-1 or rank-R tensor, and the blue dotted line denotes the hyperplane defined by rank-R tensor.
Remotesensing 14 00196 g003
Figure 4. Illustration of slices with proper sizes and multiscale slices. (a) Proper sizes of slices for different objects (b) illustration of multiscale slice.
Figure 4. Illustration of slices with proper sizes and multiscale slices. (a) Proper sizes of slices for different objects (b) illustration of multiscale slice.
Remotesensing 14 00196 g004
Figure 5. Graph based constraints model. Each vertex in the graph indicates a set of dual variables, and the sum of each vertex with the same color is equal.
Figure 5. Graph based constraints model. Each vertex in the graph indicates a set of dual variables, and the sum of each vertex with the same color is equal.
Remotesensing 14 00196 g005
Figure 6. Airplanes slices with different types from two slices sets of dataset 1. (al) are from slices set 1 of dataset 1, and (mx) are from slices set 2 of dataset 1.
Figure 6. Airplanes slices with different types from two slices sets of dataset 1. (al) are from slices set 1 of dataset 1, and (mx) are from slices set 2 of dataset 1.
Remotesensing 14 00196 g006
Figure 7. Ship slices from two slices sets of dataset 2. (ae) are from slices set 1 of dataset 2, and (fj) are from slices set 2 of dataset 2.
Figure 7. Ship slices from two slices sets of dataset 2. (ae) are from slices set 1 of dataset 2, and (fj) are from slices set 2 of dataset 2.
Remotesensing 14 00196 g007aRemotesensing 14 00196 g007b
Figure 8. The classification accuracies of MCMS-STM under different R and C using dataset 1. (a) The classification accuracies of OVR version of MCMS-STM with different R and C; (b) The classification accuracies of OVO version of MCMS-STM with different R and C.
Figure 8. The classification accuracies of MCMS-STM under different R and C using dataset 1. (a) The classification accuracies of OVR version of MCMS-STM with different R and C; (b) The classification accuracies of OVO version of MCMS-STM with different R and C.
Remotesensing 14 00196 g008aRemotesensing 14 00196 g008b
Figure 9. The classification accuracies of MCMS-STM with or without different image resizing operations.
Figure 9. The classification accuracies of MCMS-STM with or without different image resizing operations.
Remotesensing 14 00196 g009
Figure 10. The time consumption of solving the dual problem of MCMS-STM and multiclass SVM using different optimization solving algorithms under different numbers of training samples.
Figure 10. The time consumption of solving the dual problem of MCMS-STM and multiclass SVM using different optimization solving algorithms under different numbers of training samples.
Remotesensing 14 00196 g010
Table 1. The symbols and their corresponding description used in this paper.
Table 1. The symbols and their corresponding description used in this paper.
SymbolDescription
lowercase letters (e.g., x,y) scalar
lowercase boldface letters (e.g., x,y)vector
uppercase boldface letter (e.g., M)matrix
calligraphy letter (e.g., X )tensor
Table 2. The description of notations and abbreviations for the proposed MCMS-STM.
Table 2. The description of notations and abbreviations for the proposed MCMS-STM.
Notations and AbbreviationsDescription
Mthe number of classes, as well as the number of scales.
Rthe CP rank of projection tensor.
Lthe number of orders for projection tensor.
w m , m ( r , l ) the   projection   vector   of   mode l   used   to   separate   class   m   from   class   m
w m ( r , l ) the projection vector of mode-l used to separate class m from others
ξ i m slacking variable
Cregularization parameter
b y i , m the bias used to separate class yi from class m
b y i the bias used to separate class yi from others
a i m dual variable
SVMsupport vector machine
STMsupport tensor machine
OVOone-versus-one
OVRone-versus-rest
CPCANDECOMP/PARAFAC
Table 3. The accuracies of different methods for multiscale airplanes recognization using 5-fold cross-validation.
Table 3. The accuracies of different methods for multiscale airplanes recognization using 5-fold cross-validation.
MethodParameter SettingAccuracy
OVR version of MCMS-STMC = 1 R = 884.9%
OVO version of MCMS-STM C = 1 R = 685.3%
C S V M (OVO) 1C = 1082.2%
C S V M (OVR) 1C = 10073.8%
ν S V M (OVO) 1 ν = 0.1 80.7%
ν S V M (OVR) 1 ν = 0.2 76.2%
Multi-class SVM1C = 10082.6%
C S T M (OVO) 1C = 1071.4%
C S T M (OVR) 1C = 166.5%
C S V M (OVO)2C = 1079.4%
C S V M (OVR) 2C = 10077.1%
ν S V M (OVO) 2 ν = 0.1 80.3%
ν S V M (OVR) 2 ν = 0.1 75.2%
Multi-class SVM2C = 1081.7%
C S T M (OVO) 2C = 175.2%
C S T M (OVR) 2C = 170.6%
Table 4. The accuracies of different methods for multiscale airplanes recognization using 10-fold cross-validation.
Table 4. The accuracies of different methods for multiscale airplanes recognization using 10-fold cross-validation.
Method C S T M Parameter SettingAccuracy
OVR version of MCMS-STM C = 10 R = 888.1%
OVO version of MCMS-STM C = 10 R = 889.5%
C S V M (OVO) 1C = 1084.0%
C S V M (OVR) 1C = 176.6%
ν S V M (OVO) 1 ν = 0.1 85.3%
ν S V M (OVR) 1 ν = 0.2 77.3%
Multi-class SVM1C = 185.3%
C S T M (OVO) 1C = 10079.4%
C S T M (OVR) 1C = 174.8%
C S V M (OVO)2C = 10087.6%
C S V M (OVR) 2C = 178.0%
ν S V M (OVO) 2 ν = 0.1 88.6%
ν S V M (OVR) 2 ν = 0.2 79.4%
Multi-class SVM2C = 10088.1%
C S T M (OVO) 2C = 177.5%
C S T M (OVR) 2C = 171.2%
Table 5. The accuracies of different methods for multiscale ships recognization using 5-fold cross-validation.
Table 5. The accuracies of different methods for multiscale ships recognization using 5-fold cross-validation.
MethodParameter SettingAccuracy
OVR version of MCMS-STM C = 100 R = 690.2%
OVO version of MCMS-STM C = 1 R = 687.4%
C S V M (OVO) 1C = 186.1%
C S V M (OVR) 1C = 180.7%
ν S V M (OVO) 1 ν = 0.1 86.1%
ν S V M (OVR) 1 ν = 0.1 79.0%
Multi-class SVM1C = 185.3%
C S T M (OVO) 1C = 1076.9%
C S T M (OVR) 1C = 173.1%
C S V M (OVO)2C = 1083.3%
C S V M (OVR) 2C = 10079.2%
ν S V M (OVO) 2 ν = 0.1 84.5%
ν S V M (OVR) 2 ν = 0.1 79.5%
Multi-class SVM2C = 10084.2%
C S T M (OVO) 2C = 177.0%
C S T M (OVR) 2C = 170.6%
Table 6. The accuracies of different methods for multiscale ships recognization using 10-fold cross-validation.
Table 6. The accuracies of different methods for multiscale ships recognization using 10-fold cross-validation.
MethodParameter SettingAccuracy
OVR version of MCMS-STM C = 10 R = 887.7%
OVO version of MCMS-STM C = 1 R = 891.4%
C S V M (OVO) 1C = 187.4%
C S V M (OVR) 1C = 10083.6%
ν S V M (OVO) 1 ν = 0.1 87.0%
ν S V M (OVR) 1 ν = 0.1 82.4%
Multi-class SVM1C = 189.9%
C S T M (OVO) 1C = 178.2%
C S T M (OVR) 1C = 174.4%
C S V M (OVO)2C = 10085.6%
C S V M (OVR) 2C = 181.0%
ν S V M (OVO) 2 ν = 0.1 86.2%
ν S V M (OVR) 2 ν = 0.2 80.5%
Multi-class SVM2C = 10086.2%
C S T M (OVO) 2C = 176.4%
C S T M (OVR) 2C = 1069.6%
Table 7. The accuracies of different methods using the second type of normalization operation for multiscale airplanes classification using 10-fold cross-validation.
Table 7. The accuracies of different methods using the second type of normalization operation for multiscale airplanes classification using 10-fold cross-validation.
MethodPositiveNegativeTiesp-Value
C S V M (OVO)901<0.0001
C S V M (OVR)8110.0045
ν S V M (OVO)901<0.0001
ν S V M (OVR)8200.0049
Multi-class SVM8020.0097
C S T M (OVO)901<0.0001
C S T M (OVR)1000<0.0001
Table 8. The accuracies of the MCMS-STM and deep learning methods for dataset 1 under different numbers of training samples.
Table 8. The accuracies of the MCMS-STM and deep learning methods for dataset 1 under different numbers of training samples.
p = 0.1p = 0.2p = 0.3p = 0.4p = 0.5
GhostNet47.09%48.50%57.82%61.42%71.03%
ResNeXt40.86%46.06%52.08%54.84%60.19%
MCMS-STM61.73%63.79%73.20%79.39%81.65%
Table 9. The accuracies of the MCMS-STM and deep learning methods for dataset 2 under different numbers of training samples.
Table 9. The accuracies of the MCMS-STM and deep learning methods for dataset 2 under different numbers of training samples.
p = 0.1p = 0.2p = 0.3p = 0.4p = 0.5
GhostNet35.41%35.03%46.43%50.66%56.31%
ResNeXt30.37%34.97%44.64%47.74%54.23%
MCMS-STM62.21%70.07%70.29%74.76%79.53%
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Gao, T.; Chen, H.; Chen, W. MCMS-STM: An Extension of Support Tensor Machine for Multiclass Multiscale Object Recognition in Remote Sensing Images. Remote Sens. 2022, 14, 196. https://doi.org/10.3390/rs14010196

AMA Style

Gao T, Chen H, Chen W. MCMS-STM: An Extension of Support Tensor Machine for Multiclass Multiscale Object Recognition in Remote Sensing Images. Remote Sensing. 2022; 14(1):196. https://doi.org/10.3390/rs14010196

Chicago/Turabian Style

Gao, Tong, Hao Chen, and Wen Chen. 2022. "MCMS-STM: An Extension of Support Tensor Machine for Multiclass Multiscale Object Recognition in Remote Sensing Images" Remote Sensing 14, no. 1: 196. https://doi.org/10.3390/rs14010196

APA Style

Gao, T., Chen, H., & Chen, W. (2022). MCMS-STM: An Extension of Support Tensor Machine for Multiclass Multiscale Object Recognition in Remote Sensing Images. Remote Sensing, 14(1), 196. https://doi.org/10.3390/rs14010196

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop