Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
On the τ Decomposition Method for the Stability and Bifurcation of the TCP/AQM Networks versus Time Delay
Next Article in Special Issue
Constructing Dixon Matrix for Sparse Polynomial Equations Based on Hybrid and Heuristics Scheme
Previous Article in Journal
Another Perspective on Automatic Construction of Integral Distinguishers for ARX Ciphers
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Efficient Algorithm for Solving the Matrix Optimization Problem in the Unsupervised Feature Selection

School of Mathematics and Computational Science, Guangxi Colleges and Universities Key Laboratory of Data Analysis and Computation, Guilin University of Electronic Technology, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Symmetry 2022, 14(3), 462; https://doi.org/10.3390/sym14030462
Submission received: 25 December 2021 / Revised: 11 January 2022 / Accepted: 17 January 2022 / Published: 25 February 2022

Abstract

:
In this paper, we consider the symmetric matrix optimization problem arising in the process of unsupervised feature selection. By relaxing the orthogonal constraint, this problem is transformed into a constrained symmetric nonnegative matrix optimization problem, and an efficient algorithm is designed to solve it. The convergence theorem of the new algorithm is derived. Finally, some numerical examples show that the new method is feasible. Notably, some simulation experiments in unsupervised feature selection illustrate that our algorithm is more effective than the existing algorithms.

1. Introduction

Throughout this paper, we use R n × m to denote the set of m × n real matrices. We write B 0 if the matrix B is nonnegative. The symbols T r ( B ) , B T stand for the trace and transpose of the matrix B, respectively. The symbol α stands for the l 2 -norm of the vector α , i.e., α = ( α T α ) 1 2 . The symbol B F stands for the Frobenius norm of the matrix B. The symbol I p stands for the p × p identity matrix. For the matrices A and B, A B denotes the Hadamard product of A and B. The symbol m a x { x , y } represents the greater of x and y.
In this paper, we consider the following symmetric matrix optimization problem in unsupervised feature selection.
Problem 1.
Given a matrix A R n × m , consider the symmetric matrix optimization problem
min X , Y 1 2 A A X Y F 2 , s . t . X 0 , Y 0 , X T X = I p .
Here A = [ f 1 , f 2 , , f m ] R n × m is the data matrix, X R m × p is the indicator matrix (feature weight matrix) and Y R p × m is the coefficient matrix.
Problem 1 arises in unsupervised feature selection, which is an important part of machine learning. This can be stated as follows. Data from image processing, pattern recognition and machine learning are usually high-dimensional data. If we deal with these data directly, this may increase the computational complexity and the memory of the algorithm. In particular, it may lead to the overfitting phenomenon for the machine learning model. Feature selection is a common dimension reduction method, the goal of which is to find the most representative feature subset from the original features, that is to say, for a given original high-dimensional data matrix A, we must find out the relationship between the original feature space and the subspace generated by the selected feature. Feature selection can be formalized as follows.
min I d ( S p a n ( A ) , S p a n ( A c ) ) = A I c A I Y F , s . t . | I | = k ,
where I denotes the index set of the selected features and Y is the coefficient matrix of the initial feature space in the selected features. From the viewpoint of matrix factorization, feature selection is expressed as follows
min X , Y 1 2 A A X Y , s . t . X 0 , X T X = I p .
Considering that the data in practical problems are often nonnegative, we add a constraint to guarantee that any feature is described as the positive linear combination of the selected features, so the problem in (3) can be rewritten as in (1).
In the last few years, many numerical methods have been proposed for solving optimization problems with nonnegative constraints, and these methods can be broadly classified into two categories: alternating gradient descent methods and alternating nonnegative least squares methods. The most commonly used alternating gradient descent method is the multiplicative update algorithm [1,2]. Although the multiplicative update algorithm is simple to implement, it lacks a convergence guarantee. The alternating nonnegative least squares method is used to solve nonnegative subproblems. Many numerical methods, such as the active method [3], the projected gradient method [4,5], the projected Barzilai–Borwein method [6,7], the projected Newton method [8] and the projected quasi-Newton method [9,10], have been designed to solve these subproblems.
For optimization problems with orthogonal constraints, which are also known as optimization problems on manifolds, there are many algorithms to solve this type of problem. In general, these can be divided into two categories: the feasible method and the infeasible method. The feasible method means that the variance obtained after each iteration must satisfy the orthogonal constraint. Many traditional optimization algorithms, such as the gradient method [11], the conjugate method [12], the trust-region method [13], Newton method [8] and the quasi-Newton method [14], can be used to deal with optimization problems on manifolds. Wen and Yin [15] proposed the CMBSS algorithm, which combined the Cayley transform and the curvilinear search approach with BB steps. However, the computational complexity increases when the number of variables or the amount of data increases, resulting in the low efficiency of this algorithm. Infeasible methods can overcome this disadvantage when facing high-dimensional data. In 2013, Lai and Osher [16] proposed a splitting method based on Bregman iteration and the ADMM method for orthogonality constraint problems. The SOC method is a valid and efficient method for solving the convex optimization problems but the proof of its convergence is still uncertain. Thus, Chen et al. [17] put forward a proximal alternating augmented Lagrangian method to solve such optimization problems with a non-smooth objective function and non-convex constraint.
Some unsupervised feature selection algorithms based on matrix decomposition have been proposed and have achieved good performance, such as SOCFS, MFFS, RUFSM, OPMF and so on. Based on the orthogonal basis clustering algorithm, SOCFS [18] does not explicitly use the pre-computed local structure information for data points represented as additional terms of their objective functions, but directly computes latent cluster information by means of the target matrix, conducting orthogonal basis clustering in a single unified term of the objective function. In 2017, Du S et al. [19] proposed RUFSM, in which robust discriminative feature selection and robust clustering are performed simultaneously under the l 2 , 1 -norm, while the local manifold structures of the data are preserved. MFFS was developed from the viewpoint of subspace learning. That is, it treats feature selection as a matrix factorization problem and introduces an orthogonal constraint into its objective function to select the most informative features from high-dimensional data.
OPMF [20] incorporates matrix factorization, ordinal locality structure preservation and inner-product regularization into a unified framework, which can not only preserve the ordinal locality structure of the original data, but also achieve sparsity and low redundancy among features.
However, research into Problem 1 is very scarce as far as we know. The greatest difficulty is how to deal with the nonnegative and orthogonal constraints, because the problem has highly structured constraints. In this paper, we first use the penalty technique to deal with the orthogonal constraint and reformulate Problem 1 as a minimization problem with nonnegative constraints. Then, we design a new method for solving this problem. Based on the auxiliary function, the convergence theorem of the new method is derived. Finally, some numerical examples show that the new method is feasible. In particular, some simulation experiments in unsupervised feature selection illustrate that our algorithm is more efficient than the existing algorithms.
The rest of this paper is organized as follows. A new algorithm is proposed to solve Problem 1 in Section 2 and the convergence analysis is given in Section 3. In Section 4, some numerical examples are reported. Numerical tests on the proposed algorithm applied to unsupervised feature selection are also reported in that section.

2. A New Algorithm for Solving Problem 1

In this section we first design a new algorithm for solving Problem 1; then, we present the properties of this algorithm.
Problem 1 is difficult to solve due to the orthogonal constraint X T X = I p . Fortunately, this difficulty can be overcome by adding a penalty term for the constraint. Therefore, Problem 1 can be transformed into the following form
min X , Y F ( X , Y ) = 1 2 A A X Y F 2 + ρ 4 X T X I p F 2 , s . t . X 0 , Y 0 ,
where ρ is a penalty coefficient. This extra term is used to penalize the divergence between X T X and I p . The parameter ρ is chosen by users to make a trade-off between making 1 2 A A X Y F 2 small, while ensuring that X T X I p F 2 is not excessively large.
Let the Lagrange function of (4) be
L ( X , Y ) = F ( X , Y ) T r ( α X T ) T r ( β Y T ) = 1 2 T r [ ( A A X Y ) T ( A A X Y ) ] T r ( α X T ) T r ( β Y T ) = 1 2 T r ( A T A ) T r ( Y T X T A T A ) + 1 2 T r ( Y T X T A T A X Y ) T r ( α X T ) T r ( β Y T ) ,
where α R m × p and β R p × m are the Lagrangian multipliers of X and Y. It is straightforward to obtain its gradient functions as follows
X F ( X , Y ) = A T A Y T + A T A X Y Y T + ρ ( X X T X X ) ,
Y F ( X , Y ) = X T A T A + X T A T A X Y .
Setting the partial derivatives of X and Y to zero, we obtain
L X = A T A Y T + A T A X Y Y T + ρ ( X X T X X ) α = 0 ,
L Y = X T A T A + X T A T A X Y β = 0 ,
which implies that
α = A T A Y T + A T A X Y Y T + ρ ( X X T X X ) = X F ( X , Y ) ,
β = X T A T A + X T A T A X Y = Y F ( X , Y ) .
Noting that ( X , Y ) is the stationary point of (4), if it satisfies the KKT conditions
X 0 , Y 0 , X F ( X , Y ) = α 0 , Y F ( X , Y ) = β 0 , α X = 0 , β Y = 0 ,
which implies
X F ( X , Y ) X = 0 , Y F ( X , Y ) Y = 0 ,
or
( A T A Y T + A T A X Y Y T + ρ ( X X T X X ) ) i j · X i j = 0 ,
( X T A T A + X T A T A X Y ) i j · Y i j = 0 .
According to (6) and (7) we can obtain the following iterations
X i j ( k + 1 ) X i j ( k ) [ A T A ( Y ( k ) ) T + ρ X ( k ) ] i j [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j ,
Y i j ( k + 1 ) Y i j ( k ) [ ( X ( k + 1 ) ) T A T A ] i j [ ( X ( k + 1 ) ) T A T A X ( k + 1 ) Y ( k ) ] i j ,
which are equivalent to the following update formulations
X i j ( k + 1 ) X i j ( k ) X i j ( k ) [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j [ X F ( X ( k ) , Y ( k ) ) ] i j ,
Y i j ( k + 1 ) Y i j ( k ) Y i j ( k ) [ ( X ( k + 1 ) ) T A T A X ( k + 1 ) Y ( k ) ] i j [ Y F ( X ( k + 1 ) , Y ( k ) ) ] i j .
However, the iterative formulae (10) and (11) have two drawbacks, as follows.
(1)
The denominator of (10) or (11) may be zero, which violates the rule of fractional operation.
(2)
When X i j ( k ) = 0 and [ X F ( X ( k ) , Y ( k ) ) ] i j < 0 or Y i j ( k ) = 0 and [ Y F ( X ( k ) , Y ( k ) ) ] i j < 0 , the convergence cannot be guaranteed under the updating rule of (10) and (11).
In order to overcome these difficulties, we designed the following iterative methods
X i j ( k + 1 ) X i j ( k ) X ¯ i j ( k ) [ X F ( X ( k ) , Y ( k ) ) ] i j [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j + δ ,
Y i j ( k + 1 ) Y i j ( k ) Y ¯ i j ( k ) [ Y F ( X ( k + 1 ) , Y ( k ) ) ] i j [ ( X ( k + 1 ) ) T A T A X ( k + 1 ) Y ( k ) ] i j + δ ,
where
X ¯ i j ( k ) = X i j ( k ) , i f [ X F ( X ( k ) , Y ( k ) ) ] i j 0 . m a x { X i j ( k ) , σ } , i f [ X F ( X ( k ) , Y ( k ) ) ] i j < 0 .
Y ¯ i j ( k ) = Y i j ( k ) , i f [ Y F ( X ( k ) , Y ( k ) ) ] i j 0 . m a x { Y i j ( k ) , σ } , i f [ Y F ( X ( k ) , Y ( k ) ) ] i j < 0 .
Here, σ is a small positive number which can guarantee the nonnegativity of every element of X and Y. So we can establish a new algorithm for solving Problem 1 as follows (Algorithm 1).
Algorithm 1: This Algorithm attempts to Solve Problem 1.
Input Data matrix A R n × m , the number of selected features p, parameters σ , ρ and δ .
Output An index set of selected features I { 1 , 2 , , m } and | I | = p .
1. Initialize matrix X ( 0 ) 0 and Y ( 0 ) 0 .
2. Set k : = 0 .
3. Repeat
4. Fix Y and update X by (2.9);
5. Fix X and update Y by (2.10);
6. Until convergence condition has been satisfied, otherwise set k : = k + 1 and turn to step 3.
7. End for
8. X = ( x 1 , x 2 , , x n ) T . Compute x i and sort them in a descending order to choose
   the top p features.
The sequences { X ( k ) } and { Y ( k ) } generated by Algorithm 1 have the following property.
Theorem 1.
If X ( 0 ) > 0 and Y ( 0 ) > 0 , then for arbitrary k 0 , we have X ( k ) > 0 and Y ( k ) > 0 . If X ( 0 ) 0 and Y ( 0 ) 0 , then for arbitrary k 0 , we have X ( k ) 0 and Y ( k ) 0 .
Proof. 
It is obvious that the conclusion is true when k = 0 . Now we will consider the case k > 0 . □
Case I.
From [ X F ( X , Y ) ] i j 0 it follows that X ¯ i j = X i j and
X i j ( k + 1 ) = X i j ( k ) X ¯ i j ( k ) [ X F ( X ( k ) , Y ( k ) ) ] i j [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j + δ = [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j X i j ( k ) + δ X i j ( k ) [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j + δ [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j X i j ( k ) [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j + δ + ( A T A Y ( k ) + ρ X ( k ) ) i j X i j ( k ) [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j + δ = [ ( A T A ( Y ( k ) ) T + ρ X ( k ) ) i j + δ ] X i j ( k ) [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j + δ .
Hence, if X i j ( k ) > 0 then X i j ( k + 1 ) > 0 and if X i j ( k ) 0 then X i j ( k + 1 ) 0 .
Case II.
From [ X F ( X , Y ) ] i j < 0 , it follows that X ¯ i j X i j and
X i j ( k + 1 ) = X i j ( k ) m a x ( X i j ( k ) , σ ) [ X F ( X ( k + 1 ) , Y ( k + 1 ) ) ] i j [ A T A X ( k ) Y ( k ) ( Y ( k ) ) T + ρ X ( k ) ( X ( k ) ) T X ( k ) ] i j + δ .
Noting that m a x { X i j ( k ) , σ } > 0 and [ X F ( X ( k ) , Y ( k ) ) ] i j < 0 , we can easily conclude that if X i j ( k ) > 0 then X i j ( k + 1 ) > 0 and if X i j ( k ) 0 then X i j ( k + 1 ) 0 .
Case III.
From [ Y F ( X , Y ) ] i j 0 it follows that Y ¯ i j = Y i j and
Y i j ( k + 1 ) = Y i j ( k ) Y ¯ i j ( k ) [ Y F ( X ( k + 1 ) , Y ( k ) ) ] i j [ ( X ( k + 1 ) ) T A T A X ( k + 1 ) Y ( k ) ] i j + δ = [ ( X ( k + 1 ) ) T A T A X ( k + 1 ) Y ( k ) ] i j Y i j ( k ) + δ Y i j ( k ) [ ( X ( k + 1 ) ) T A T A X ( k + 1 ) Y ( k ) ] i j + δ [ ( X ( k + 1 ) ) T A T A X ( k + 1 ) Y ( k ) ( X ( k + 1 ) ) T A T A ] i j Y i j ( k ) [ ( X ( k + 1 ) ) T A T A X ( k + 1 ) Y ( k ) ] i j + δ = [ ( ( X ( k + 1 ) ) T A T A ) i j + δ ] Y i j ( k ) [ ( X ( k + 1 ) ) T A T A X ( k + 1 ) Y ( k ) ] i j + δ .
Thus, if Y i j ( k ) > 0 then Y i j ( k + 1 ) > 0 and if Y i j ( k ) 0 then Y i j ( k + 1 ) 0 .
Case IV.
From [ Y F ( X , Y ) ] i j < 0 , it follows that Y ¯ i j Y i j and
Y i j ( k + 1 ) = Y i j ( k ) m a x ( Y i j ( k ) , σ ) [ Y F ( X , Y ) ] i j [ ( X ( k + 1 ) ) T A T A X ( k + 1 ) Y ( k ) ] i j + δ .
Based on the fact that m a x { Y i j ( k ) , σ } > 0 and [ Y F ( X , Y ) ] i j < 0 , we can conclude that if Y i j ( k ) > 0 then Y i j ( k + 1 ) > 0 and if Y i j ( k ) 0 then Y i j ( k + 1 ) 0 .  □

3. Convergence Analysis

In this section, we will give the convergence theorem for Algorithm 1. For the objective function
F ( X , Y ) = 1 2 A A X Y F 2 + ρ 4 X T X I p F 2
of Problem (4), we first prove that
F ( X ( k + 1 ) , Y ( k + 1 ) ) F ( X ( k ) , Y ( k ) ) ,
where X ( k ) and Y ( k ) are the k-th iteration of Algorithm 1, then obtain the limit point as the stationary point of Problem (4). In order to develop this section, we need a lemma.
Lemma 1
([21]). If there exists a function G ( u , u ) of H ( u ) satisfying
G ( u , u ) H ( u ) , G ( u , u ) = H ( u ) ,
then H ( u ) is non-increasing under the update rule
u ( k + 1 ) = a r g min u G ( u , u ( k ) ) .
Here G ( u , u ) is called an auxiliary function of H ( u ) if it satisfies (16).
Theorem 2.
Fixing X, the objective function F ( X , Y ) is non-increasing, that is
F ( X ( k ) , Y ( k + 1 ) ) F ( X ( k ) , Y ( k ) ) .
Proof. 
Set A = [ f 1 , f 2 , , f m ] and Y = ( y 1 , y 2 , , y m ) , then
A A X Y F 2 = i = 1 m f i A X y i ( k ) 2 .
Noting that
F ( X ( k ) , Y ( k + 1 ) ) = 1 2 A A X ( k ) Y ( k + 1 ) F 2 + ρ 4 ( X ( k ) ) T X ( k ) I p F 2 ,
F ( X ( k ) , Y ( k ) ) = 1 2 A A X ( k ) Y ( k ) F 2 + ρ 4 ( X ( k ) ) T X ( k ) I p F 2 ,
and when X is fixed we can ignore the constant term ρ 4 X T X I p F 2 , then
F ( X ( k ) , Y ( k + 1 ) ) F ( X ( k ) , Y ( k ) ) 1 2 A A X ( k ) Y ( k + 1 ) F 2 1 2 A A X ( k ) Y ( k ) F 2
i = 1 m f i A X ( k ) y i ( k + 1 ) i = 1 m f i A X ( k ) y i ( k ) f i A X ( k ) y i ( k + 1 ) f i A X ( k ) y i ( k ) .
If we need to prove F ( X ( k ) , Y ( k + 1 ) ) F ( X ( k ) , Y ( k ) ) , we must prove that
H ( y ) = 1 / 2 f A X y 2
is a nonincreasing function. Noting that H ( y ) is a quadratic function and its second-order Taylor approximation at y ( k ) is as follows
H ( y ) = H ( y ( k ) ) + ( y y ( k ) ) T H ( y ( k ) ) + 1 / 2 ( y y ( k ) ) T 2 H ( y ( k ) ) ( y y ( k ) ) ,
where
H ( y ( k ) ) = X T A f + X T A T A X y ( k ) ,
and
2 H ( y ( k ) ) = X T A T A X .
Now we will construct a function
G ( y , y ( k ) ) = H ( y ( k ) ) + ( y y ( k ) ) T H ( y ( k ) ) + 1 / 2 ( y y ( k ) ) T P ¯ ( y y ( k ) ) ,
where P ¯ is a diagonal matrix with
P ¯ i i = ( X T A T A X y ¯ ( k ) ) i + δ y ¯ i ( k ) , i f i I , 0 , i f i I .
I = { i y i ( k ) > 0 , H ( y ( k ) ) i 0 or y i ( k ) = 0 , H ( y ( k ) ) i < 0 } { i y ¯ i ( k ) > 0 , H ( y ( k ) ) i 0 } .
We begin to prove that G ( y , y ( k ) ) is an auxiliary function of H ( y ) . It is obvious that G ( y , y ) = H ( y ) is satisfied; now we prove that the inequality G ( y , y ( k ) ) H ( y ) holds. Noting that
G ( y , y ( k ) ) H ( y ) = 1 / 2 ( y y ( k ) ) ( P ¯ 2 H ( y ( k ) ) ) ( y y ( k ) ) = 1 / 2 ( y y ( k ) ) ( P ¯ X T A T A X ) ( y y ( k ) ) .
In fact, we can prove that the matrix P ¯ X T A T A X is a positive semi-definite matrix.
Case I.
When H ( y ( k ) ) > 0 or H ( y ( k ) ) < 0 but y i ( k ) > δ > 0 , we have y ¯ i ( k ) = y i ( k ) . For any nonzero vector z = ( z 1 , z 2 , , z p ) T R p , we have
z T ( P ¯ X T A T A X ) z = i = 1 p ( X T A T A X y ( k ) ) i ( Y ( k ) ) i z i 2 + i = 1 p δ ( y ( k ) ) i z i 2 i = 1 p j = 1 p ( X T A T A X ) i j z i z j > i = 1 p ( X T A T A X y ( k ) ) i ( y ( k ) ) i z i 2 i = 1 p j = 1 p ( X T A T A X ) i j z i z j = 1 / 2 i = 1 p j = 1 p ( X T A T A X ) i j ( y ( k ) ) i ( y ( k ) ) j z j 2 + 1 / 2 i = 1 p j = 1 p ( Z T A T A X ) j i ( y ( k ) ) j ( y ( k ) ) i z i 2 i = 1 p j = 1 p ( X T A T A X ) i j z i z j = 1 / 2 i = 1 p j = 1 p ( y i ( k ) y j t z j y j t y i ( k ) z i ) 2 ( X T A T A X ) i j 0 .
The last inequality is true due to the nonnegativity of X, and the data matrix A generally has practical significance, so the elements are usually nonnegative; therefore, we can obtain that the matrix X T A T A X is also a nonnegative matrix. Thus we obtain
G ( y , y ( k ) ) H ( y ) .
Case II.
When H ( y ( k ) ) < 0 and δ > y i ( k ) > 0 , we have y ¯ i ( k ) = δ , we can also use the same technique to verify that matrix P ¯ X T A T A X is positive semi-definite. According to Lemma 1, we obtain that H ( y ) is anon-increasing function so F ( X , Y ) is non-increasing when X is fixed. Therefore, we can obtain
F ( X ( k ) , Y ( k + 1 ) ) F ( X ( k ) , Y ( k ) ) .
This completes the proof. □
Similarly, we can use the same method to verify that when Y is fixed, the function F ( X , Y ) is also a non-increasing function. Thus, we have
F ( X ( k + 1 ) , Y ( k ) ) F ( X ( k ) , Y ( k ) ) .
Consequently, by (20) and (21), we obtain
F ( X ( k + 1 ) , Y ( k + 1 ) ) F ( X ( k + 1 ) , Y ( k ) ) F ( X ( k ) , Y ( k ) ) .
Theorem 3.
The sequence { ( X ( k ) , Y ( k ) ) } generated by Algorithm 1 converges to the stationary point of Problem 1.
Proof. 
Since F ( X ( k ) , Y ( k ) ) is a decreasing sequence and it is bounded with the lower bound zero and the upper bound F ( X ( 0 ) , Y ( 0 ) ) , and combining Theorem 1, there exist nonnegative matrices ( X , Y ) such that
lim k 0 , k F ( X ( k ) , Y ( k ) ) = F ( X , Y ) .
Because of the continuity and monotonicity of the function F, we can obtain
lim k 0 , k ( X ( k ) , Y ( k ) ) = ( X , Y ) .
Now we will prove the point { ( X , Y ) } is the stationary point of Problem 1, that is, we will prove that { ( X , Y ) } satisfies the KKT conditions (5). We first prove
i f Y i j > 0 , t h e n ( Y F ( X , Y ) ) i j = 0 ,
and
i f Y i j = 0 , t h e n ( Y F ( X , Y ) ) i j 0 .
Based on the definition of Y ¯ i j in (14)
Y ¯ i j ( k ) = m a x ( Y i j ( k ) , σ ) or Y i j ( k ) ,
so the sequence { Y ¯ i j ( k ) } may have two convergent points Y i j or σ . We set
Y ˜ = lim k 0 , k Y ¯ ( k ) .
Furthermore, according to (25), we have
lim k 0 , k ( Y i j ( k + 1 ) Y i j ( k ) ) = Y ˜ i j [ Y F ( X , Y ) ] i j [ ( X ) T A T A X Y ] i j + δ = 0 .
When Y i j > 0 , we have Y ˜ i j > 0 ; thereby it immediately implies Y F ( X , Y ) = 0 which is consistent with (24). Now we begin to prove (25). If the result is not true, there exists ( i , j ) such that
Y i j = 0 but [ Y F ( X , Y ) ] i j < 0 .
When k is large enough, we have [ Y F ( X ( k ) , Y ( k ) ) ] i j < 0 and
lim k 0 , k Y ¯ i j ( k ) = Y ˜ i j = σ .
Therefore
Y ˜ i j [ Y F ( X , Y ) ] i j [ ( X ) T A T A X Y ] i j + δ > 0 ,
which is a contradiction of (26); hence, (25) holds. (24) and (25) imply that Y satisfies the KKT conditions (5). In a similar way, we can prove that X satisfies the KKT conditions (5). Hence, ( X , Y ) is the stationary point of Problem 1. □

4. Numerical Experiments

In this section, we first present a simple example to illustrate that Algorithm 1 is feasible to solve Problem 1, and we apply Algorithm 1 to unsupervised feature selection. We also compare our algorithm with the MaxVar Algorithm [22], the UDFS Algorithm [23] and the MFFS Algorithm [24]. All experiments were performed in MATLAB R2014a on a PC with an Intel Core i5 processor at 2.50 GHz with a precision of ε = 2.22 × 10 16 . Set the gradient value
G V ( X , Y ) = X F ( X , Y ) X F 2 + Y F ( X , Y ) Y F 2 .
Due to the KKT conditions (5) we know that if G V ( X , Y ) = 0 then ( X , Y ) is the stationary point of Problem 1. So we use either G V ( X , Y ) 1.0 × 10 4 or the iteration step k has reached the upper limit 500 as the stopping criterion of Algorithm 1.

4.1. A Simple Example

Example 1.
Considering Problem 1 with n = 5, m = 4, p = 3 and
A = 0.6882 0.0113 0.6763 0.3245 0.4984 0.2828 0.5696 0.5210 0.0990 0.5896 0.5517 0.8649 0.2878 0.1720 0.9674 0.9941 0.5381 0.1701 0.6284 0.8385 .
Set the initial matrices
X ( 0 ) = 0.3474 0.4812 0.9596 0.7494 0.2862 0.4421 0.9394 0.5952 0.9620 0.6681 0.3364 0.6764 , Y ( 0 ) = 0.7061 0.8338 0.4641 0.8316 0.9577 0.1552 0.2987 0.5391 0.9399 0.8304 0.5233 0.2598 .
We use Algorithm 1 to solve this problem. After 99 iterations, we get the solution of Problem 1 as follows
X ( 99 ) 0 0 0.9478 0.9764 0 0 0 1.0000 0 0.0236 0 0.0522 , Y ( 99 ) 0.0005 0.9781 0.6839 1.3647 0 0 0 0 1.0662 0.0966 1.1570 0.8467 ,
and
G V ( X , Y ) = 9.8992 × 10 5 .
This example shows that Algorithm 1 is feasible to solve Problem 1.

4.2. Application to Unsupervised Feature Selection and Comparison with Existing Algorithms

4.2.1. Dataset

In the next stage of our study, we used standard databases to test the performance of our proposed algorithm. We first describe the four datasets use, the target image is shown in Figure 1, and the characteristics of which are summarized in Table 1.
1. COIL20 (http://www.cad.zju.edu.cn/home/dengcai/Data/MLData.html (accessed on 1 December 2021)): This dataset contains 20 objects. The images of each object were taken 5 degrees apart as the object was rotated on a turntable, and for each object there are 72 images. The size of each image is 32 × 32 pixels, with 256 grey levels per pixel. Thus, each image is represented by a 1024-dimensional vector.
2. PIE (http://archive.ics.uci.edu/ml/datasets.php (accessed on 1 December 2021)): This is a face image dataset with 53 different people; for each subject, 22 pictures were taken under different lighting conditions with different postures and expressions.
3. ORL (http://www.cad.zju.edu.cn/home/dengcai/Data/FaceData.html (accessed on 1 December 2021)): This dataset contains ten different images of each of 40 distinct subjects. For some subjects, the images were taken at different times, varying the lighting, facial expressions (open/closed eyes, smiling/not smiling) and facial details (glasses/no glasses). All the images were taken against a dark homogeneous background with the subjects in an upright, frontal position (with tolerance for some side movement).
4. Yale (http://www.cad.zju.edu.cn/home/dengcai/Data/FaceData.html (accessed on 1 December 2021)): This dataset contains 165 grayscale images of 15 individuals. There are 11 images per subject, and they have different facial expressions (happy, sad, surprised, sleepy, normal and wink) or lighting conditions (center-light, left-light, right-light). Using the above four datasets, we input these grayscale images as the initial value A, and then used the initial matrix X and Y so that we could obtain a series of function values through the iterative updating of the algorithm. Then, we were able to obtain four convergence curves for different databases, as shown in Figure 2.

4.2.2. Comparison Methods

1. MaxVar [22]: Selecting features according to the variance of features. The feature with the higher variance than others is more important.
2. UDFS [23]: l 2 , 1 -norm regularized discriminative feature selection method. This method selects the most distinctive feature through the local discrimination information of data and the correlation of features.
3. MFFS [24]: unsupervised feature selection via matrix factorization, in which the objective function originates from the distance between two subspaces.

4.2.3. Parameter Settings

In our proposed method, the parameter ρ was selected from the set { 10 , 10 2 , 10 9 } . In the following experiments, we set the value of ρ to be 10 4 , 10 9 , 10 8 and 10 7 for the COIL20, PIE, ORL and Yale databases. The numbers of selected features were taken from { 20 , 40 , 60 , , 200 } for all datasets. Then, we computed the average value of ACC and NMI when selecting different numbers of features. The maximum iteration number (maxiter) was set to 1000. For the sparsity parameters γ and λ in UDFS, we set γ = λ = 10 5 . The value of parameter ρ in MFFS was set as 10 8 . We set σ = δ = ε = 10 4 in our proposed algorithm. The results of the comparison of MaxVar, UDFS, MFFS and our proposed algorithm are presented in Figure 3 and Figure 4. The following two tables give the average accuracy and normalized mutual information calculated using our proposed algorithm.

4.2.4. Evaluation Metrics

There are two metrics to measure the results of clustering using the selected features. The accuracy of clustering (ACC) and normalized mutual information (NMI) can be calculated as follows. The value of ACC and NMI scales between 0 and 1, and a high value indicates an efficient clustering result. For every dataset, there are two parts, fea and gnd, the fea data are used to operate the selection, and after clustering one can obtain a clustering label, denoted by s i ; gnd is the true label of features denoted by t i , and the ACC can be computed using the clustering label and the true label.
ACC = i = 1 n δ ( t i , m a p ( s i ) ) n ,
where δ ( a , b ) = 1 if a = b ; δ ( a , b ) = 0 if a b . The m a p ( · ) indicates a mapping that permutes the label of clustering result to match the true label as well as possible using the Kuhn–Munkres Algorithm [25]. For two variables P and Q, the NMI is defined in the following form:
NMI ( P , Q ) = I ( P , Q ) H ( P ) H ( Q ) ,
where I ( P , Q ) is the mutual information of P and Q, and H ( P ) and H ( Q ) are the entropy of P and Q, respectively.

4.2.5. Experiments Results and Analysis

Figure 4 shows the curves of the iteration step and the values of the objective function when Algorithm 1 was run on four datasets. We can see that the objective function value decreases as the iteration step increases. After a finite number of iterations, the objective function value reaches the minimum and tends to be stable.
In Table 2 and Table 3, we report the best clustering accuracy and the best normalized mutual information, expressed as the number of selected feature changes. In Table 2 and Table 3, we can see that the performance of Algorithm 1 was more effective than that of the MaxVar Algorithm [22], the UDFS Algorithm [23] and the MFFS Algorithm [24] on all data sets, which shows the effectiveness and robustness of our proposed method.
In Figure 3 and Figure 4, we present the clustering accuracy and the normalized mutual information, expressed as the number of selected feature changes. We can see that the performance of MFFS was slightly better than that of Algorithm 1 on COIL20. However, on the other three datasets, Algorithm 1 was relatively more efficient compared with the MaxVar Algorithm [22], the UDFS Algorithm [23] and the MFFS Algorithm [24], especially when the number of selected features was large.

5. Conclusions

The symmetric matrix optimization problem in the area of unsupervised feature selection is considered in this paper. By relaxing the orthogonal constraint, this problem is converted into a constrained symmetric nonnegative matrix optimization problem. An efficient algorithm was designed to solve this problem and its convergence theorem was also derived. Finally, a simple example was given to verify the feasibility of the new method. Some simulation experiments in unsupervised feature selection showed that our algorithm was more effective than the existing methods.

Author Contributions

Conceptualization, C.L.; methodology, W.W. All authors have read and agreed to the published version of the manuscript.

Funding

The work was supported by the National Natural Science Foundation of China (No. 11761024), and the Natural Science Foundation of Guangxi Province (No. 2017GXNSFBA198082).

Institutional Review Board Statement

Institutional review board approval of our school was obtained for this study.

Informed Consent Statement

Written informed consent was obtained from all the participants prior to the enrollment of this study.

Data Availability Statement

The data used to support the findings of this study are available from the corresponding author upon request.

Conflicts of Interest

The author declares no conflict of interest.

References

  1. Lee, D.D.; Seung, H.S. Learning the parts of objects by non-negative matrix factorization. Nature 1999, 401, 788. [Google Scholar] [CrossRef] [PubMed]
  2. Smaragdis, P.; Brown, J.C. Non-negative matrix factorization for polyphonic music transcription. In Proceedings of the 2003 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics (IEEE Cat. No.03TH8684), New Paltz, NY, USA, 19–22 October 2003. [Google Scholar]
  3. Kim, H.; Park, H. Nonnegative Matrix Factorization Based on Alternating Nonnegativity Constrained Least Squares and Active Set Method. SIAM J. Matrix Anal. Appl. 2008, 30, 713–730. [Google Scholar] [CrossRef]
  4. Lin, C. Projected Gradient Methods for Nonnegative Matrix Factorization. Neural Comput. 2007, 19, 2756–2779. [Google Scholar] [CrossRef] [PubMed] [Green Version]
  5. Li, X.L.; Liu, H.W.; Zheng, X.Y. Non-monotone projection gradient method for non-negative matrix factorization. Comput. Optim. Appl. 2012, 51, 1163–1171. [Google Scholar] [CrossRef]
  6. Han, L.X.; Neumann, M.; Prasad, A.U. Alternating projected Barzilai-Borwein methods for Nonnegative Matrix Factorization. Electron. Trans. Numer. Anal. 2010, 36, 54–82. [Google Scholar]
  7. Huang, Y.K.; Liu, H.W.; Zhou, S. An efficient monotone projected Barzilai-Borwein method for nonnegative matrix factorization. Appl. Math. Lett. 2015, 45, 12–17. [Google Scholar] [CrossRef]
  8. Gong, P.H.; Zhang, C.S. Efficient Nonnegative Matrix Factorization via projected Newton method. Pattern Recognit. 2012, 45, 3557–3565. [Google Scholar] [CrossRef]
  9. Kim, D.; Sra, S.; Dhillon, I.S. Fast Newton-type Methods for the Least Squares Nonnegative Matrix Approximation Problem. In Proceedings of the SIAM International Conference on Data Mining, Minneapolis, MN, USA, 26–28 April 2007. [Google Scholar]
  10. Zdunek, R.; Cichocki, A. Non-negative matrix factorization with quasi-newton optimization. In Proceedings of the International Conference on Artificial Intelligence and Soft Computing, Zakopane, Poland, 25–29 June 2006. [Google Scholar]
  11. Abrudan, T.E.; Eriksson, J.; Koivunen, V. Steepest descent algorithms for optimization under unitary matrix constraint. IEEE Trans. Signal Process. 2008, 56, 1134–1147. [Google Scholar] [CrossRef]
  12. Abrudan, T.E.; Eriksson, J.; Koivunen, V. Conjugate gradient algorithm for optimization under unitary matrix constraint. Signal Process. 2009, 89, 1704–1714. [Google Scholar] [CrossRef]
  13. Absil, P.A.; Baker, C.G.; Gallivan, K.A. Trust-Region methods on riemannian manifolds. Found. Comput. Math. 2007, 7, 303–330. [Google Scholar] [CrossRef]
  14. Savas, B.; Lim, L.H. Quasi-Newton methods on grassmannians and multilinear approximations of tensors. SIAM J. Sci. Comput. 2010, 2, 3352–3393. [Google Scholar] [CrossRef]
  15. Wen, Z.W.; Yin, W.T. A feasible method for optimization with orthogonality constraints. Math. Program. 2013, 142, 397–434. [Google Scholar] [CrossRef] [Green Version]
  16. Lai, R.J.; Osher, S. A splitting method for orthogonality constrained problems. J. Sci. Comput. 2014, 58, 431–449. [Google Scholar] [CrossRef]
  17. Chen, W.Q.; Ji, H.; You, Y.F. An augmented lagrangian method for l1-regularized optimization problems with orthogonality constraints. SIAM J. Sci. Comput. 2016, 38, 570–592. [Google Scholar] [CrossRef] [Green Version]
  18. Han, D.; Kim, J. Unsupervised Simultaneous Orthogonal basis Clustering Feature Selection. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, Boston, MA, USA, 7–15 June 2015. [Google Scholar]
  19. Du, S.Q.; Ma, Y.D.; Li, S.L. Robust unsupervised feature selection via matrix factorization. Neurocomputing 2017, 241, 115–127. [Google Scholar] [CrossRef]
  20. Yi, Y.G.; Zhou, W.; Liu, Q.H. Ordinal preserving matrix factorization for unsupervised feature selection. Signal Process. Image Commun. 2018, 67, 118–131. [Google Scholar] [CrossRef]
  21. Lee, D.D.; Seung, H.S. Algorithms for non-negative matrix factorization. In Proceedings of the International Conference on Neural Information Processing Systems, Vancouver, BC, Canada, 3–8 December 2001. [Google Scholar]
  22. Lu, Y.J.; Cohen, I.; Zhou, X.S.; Tian, Q. Feature selection using principal feature analysis. In Proceedings of the 15th ACM International Conference on Multimedia, Augsburg, Germany, 24–29 September 2007. [Google Scholar]
  23. Yang, Y.; Shen, H.T.; Ma, Z.G. l21-norm regularized discriminative feature selection for unsupervised learning. In Proceedings of the Twenty-Second International Joint Conference on Artificial Intelligence, Barcelona, Spain, 16–22 July 2011. [Google Scholar]
  24. Wang, S.P.; Witold, P.; Zhu, Q.X. Subspace learning for unsupervised feature selection via matrix factorization. Pattern Recognit. 2015, 48, 10–19. [Google Scholar] [CrossRef]
  25. Lovász, L.; Plummer, M.D. Matching Theory, 1st ed.; Elsevier Science Ltd.: Amsterdam, The Netherlands, 1986. [Google Scholar]
Figure 1. Some images from different databases. (a) COIL20, (b) PIE, (c) ORL, (d) Yale.
Figure 1. Some images from different databases. (a) COIL20, (b) PIE, (c) ORL, (d) Yale.
Symmetry 14 00462 g001
Figure 2. Clustering results (ACC) of different feature selection algorithms. (a) COIL20, (b) PIE, (c) ORL, (d) Yale.
Figure 2. Clustering results (ACC) of different feature selection algorithms. (a) COIL20, (b) PIE, (c) ORL, (d) Yale.
Symmetry 14 00462 g002
Figure 3. Clustering results (NMI) of different feature selection algorithms. (a) COIL20, (b) PIE, (c) ORL, (d) Yale.
Figure 3. Clustering results (NMI) of different feature selection algorithms. (a) COIL20, (b) PIE, (c) ORL, (d) Yale.
Symmetry 14 00462 g003
Figure 4. The convergence curves of the proposed approach on four different databases. (a) ρ = 10 4 , (b) ρ = 10 9 , (c) ρ = 10 8 , (d) ρ = 10 7 .
Figure 4. The convergence curves of the proposed approach on four different databases. (a) ρ = 10 4 , (b) ρ = 10 9 , (c) ρ = 10 8 , (d) ρ = 10 7 .
Symmetry 14 00462 g004
Table 1. Database Description.
Table 1. Database Description.
DatasetSizeFeaturesClassesData Types
COIL201440102420Object images
PIE1166102453Face images
ORL400102440Face images
Yale165102415Face images
Table 2. Clustering results (average ACC) of different algorithms on different databases.
Table 2. Clustering results (average ACC) of different algorithms on different databases.
DatasetMaxvarMFFSUDFSAlgorithm 1
COIL200.48810.53350.51180.5404
PIE0.38650.46300.20210.4799
ORL0.37700.39030.40400.4798
Yale0.30490.38120.29940.4000
Table 3. Clustering results (average NMI) of different algorithms on different databases.
Table 3. Clustering results (average NMI) of different algorithms on different databases.
DatasetMaxvarMFFSUDFSAlgorithm 1
COIL200.60200.64950.60000.6518
PIE0.65270.70980.46220.7322
ORL0.60520.61610.61140.6790
Yale0.35570.44890.36280.4563
Publisher’s Note: MDPI stays neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Share and Cite

MDPI and ACS Style

Li, C.; Wu, W. An Efficient Algorithm for Solving the Matrix Optimization Problem in the Unsupervised Feature Selection. Symmetry 2022, 14, 462. https://doi.org/10.3390/sym14030462

AMA Style

Li C, Wu W. An Efficient Algorithm for Solving the Matrix Optimization Problem in the Unsupervised Feature Selection. Symmetry. 2022; 14(3):462. https://doi.org/10.3390/sym14030462

Chicago/Turabian Style

Li, Chunmei, and Wen Wu. 2022. "An Efficient Algorithm for Solving the Matrix Optimization Problem in the Unsupervised Feature Selection" Symmetry 14, no. 3: 462. https://doi.org/10.3390/sym14030462

APA Style

Li, C., & Wu, W. (2022). An Efficient Algorithm for Solving the Matrix Optimization Problem in the Unsupervised Feature Selection. Symmetry, 14(3), 462. https://doi.org/10.3390/sym14030462

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop