Nothing Special   »   [go: up one dir, main page]

Next Article in Journal
Short-Term Electricity Load Forecasting Based on Complete Ensemble Empirical Mode Decomposition with Adaptive Noise and Improved Sparrow Search Algorithm–Convolutional Neural Network–Bidirectional Long Short-Term Memory Model
Previous Article in Journal
A Differential Privacy Framework with Adjustable Efficiency–Utility Trade-Offs for Data Collection
You seem to have javascript disabled. Please note that many of the page functionalities won't work as expected without javascript enabled.
 
 
Font Type:
Arial Georgia Verdana
Font Size:
Aa Aa Aa
Line Spacing:
Column Width:
Background:
Article

An Improvement of the Alternating Direction Method of Multipliers to Solve the Convex Optimization Problem

College of Mathematics and Computational Science, Guilin University of Electronic Technology, Guilin 541004, China
*
Author to whom correspondence should be addressed.
Mathematics 2025, 13(5), 811; https://doi.org/10.3390/math13050811
Submission received: 18 December 2024 / Revised: 17 February 2025 / Accepted: 25 February 2025 / Published: 28 February 2025

Abstract

:
The alternating direction method is one of the attractive approaches for solving convex optimization problems with linear constraints and separable objective functions. Experience with applications has shown that the number of iterations depends significantly on the penalty parameter for the linear constraint. The penalty parameters in the classical alternating direction method are a constant. In this paper, an improved alternating direction method is proposed, which not only adaptively adjusts the penalty parameters per iteration based on the iteration message but also adds relaxation factors to the Lagrange multiplier update steps. Preliminary numerical experiments show that the technique of adaptive adjusting of the penalty parameters per iteration and attaching relaxation factors in the Lagrange multiplier updating steps are effective in practical applications.

1. Introduction

Many problems in the fields of signal and image processing, machine learning, medical image reconstruction, computer vision, and network communications [1,2,3,4,5,6,7,8] can be reduced to solving the following convex optimization problem with linear constraints:
min f ( x ) + g ( y ) | A x + B y = b , x X , y Y  
where A m × m 1 and B m × m 2 are the given matrices; b m is the given vector; f : m 1 and g : m 2 are continuous closed convex functions; and X and Y are nonempty closed convex subsets in m 1 and m 2 , respectively. The augmented Lagrangian function of the convex optimization problem (Equation (1)) is the following Equation (2).
L x , y , λ = f x + g y λ T A x + B y b + β 2 A x + B y b 2 2 ,
where β > 0 is the penalty parameter and λ m is the Lagrange multiplier vector. Based on the objective function with separable structural properties, the alternating direction method of multipliers (ADMM) has been proposed in the literature [9,10] to solve the problem (Equation (1)) as follows:
x k + 1 = arg min x X f x λ k T A x + B y k b + β 2 A x + B y k b 2 2 , y k + 1 = arg min y Y g y λ k T A x k + 1 + B y b + β 2 A x k + 1 + B y b 2 2 , λ k + 1 = λ k β A x k + 1 + B y k + 1 b .
The ADMM algorithm can be viewed as the Gauss Seidel form of the augmented Lagrangian multiplier method, and it can be interpreted as the Douglas Rachford splitting algorithm (DRSM) from a dual perspective [11]. In addition, the DRSM can be further interpreted as a proximal point algorithm (PPA) [12]. The ADMM algorithm is an important method for solving separable convex optimization problems. Due to its fast processing speed and good convergence performance, the ADMM algorithm is widely used in statistical learning, machine learning, and other fields. However, the ADMM algorithm has slow convergence speed under the requirement of a large scale and high precision. Therefore, many experts and scholars are working on improving the ADMM algorithm, such as Fortin and Glowinski, who suggested in [13] attaching a relaxation factor to the Lagrange multiplier updating step in Equation (3), which this results in the following scheme:
x k + 1 = arg min x X f x λ k T A x + B y k b + β 2 A x + B y k b 2 2 , y k + 1 = arg min y Y g y λ k T A x k + 1 + B y b + β 2 A x k + 1 + B y b 2 2 , λ k + 1 = λ k γ β A x k + 1 + B y k + 1 b ,
where the parameter γ can be chosen in the interval 0 , 1 + 5 2 , and thus it becomes possible to enlarge the step size for updating the Lagrange multiplier. An advantage of this larger step size in Equation (4) is that it can numerically lead to faster convergence empirically. This scheme (Equation (4)) differs from the original ADMM scheme (Equation (3)) only in the fact that the step size for updating the Lagrange multiplier can be larger than 1. But technically they are actually two distinct families of ADMM algorithms, in which one is derived from the operator splitting framework and the other is derived from Lagrangian splitting. Thus, despite the similarity in notation, the ADMM scheme (Equation (4)) with Fortin et al.’s larger step size and the original ADMM scheme (Equation (3)) are completely different in nature. On the other hand, Glowinski et al. [14] applied the Douglas Rachford splitting method (DRSM) [11] to the dual of Equation (1) and obtained the following scheme:
x k + 1 = arg min x X f x λ k T A x + B y k b + β 2 A x + B y k b 2 2 , λ k + 1 2 = λ k β A x k + 1 + B y k b , y k + 1 = arg min y Y g y λ k + 1 2 T A x k + 1 + B y b + β 2 A x k + 1 + B y b 2 2 , λ k + 1 = λ k + 1 2 β A x k + 1 + B y k + 1 b .
This scheme (Equation (5)) can be regarded as a symmetric version of the ADMM scheme (Equation (3)) in the sense that the variables x and y are treated equally, each of which is followed consequently by an update of the Lagrange multiplier. However, as shown in [15], the sequence generated by the symmetric ADMM (Equation (5)) is not necessarily strictly contractive with respect to the solution set of Equation (1), while this property can be ensured by the sequence generated by the ADMM (Equation (3)). Because of these deficiencies, Bingsheng He et al. [15] proposed the following scheme:
x k + 1 = arg min x X f x λ k T A x + B y k b + β 2 A x + B y k b 2 2 , λ k + 1 2 = λ k α β A x k + 1 + B y k b , y k + 1 = arg min y Y g y λ k + 1 2 T A x k + 1 + B y b + β 2 A x k + 1 + B y b 2 2 , λ k + 1 = λ k + 1 2 α β A x k + 1 + B y k + 1 b ,  
where the parameter α∈ (0, 1) is for shrinking the step sizes in Equation (5). The sequence generated by Equation (6) is strictly contractive with respect to the solution set of Equation (1). Thus, we called it the strictly contractive symmetric version of the ADMM. Limiting α ∈ (0, 1) in the strictly contractive symmetric version of the ADMM (Equation (6)) makes the updating of Lagrange multipliers more conservative and with smaller steps, but smaller steps should be strongly avoided in practical applications. Instead, one wants to seek larger steps where possible to speed up numerical performance. For this purpose, Bingsheng He et al. [16] proposed the following scheme:
x k + 1 = arg min x X f x λ k T A x + B y k b + β 2 A x + B y k b 2 2 , λ k + 1 2 = λ k r β A x k + 1 + B y k b , y k + 1 = arg min y Y s g y λ k + 1 2 T A x k + 1 + B y b + β 2 A x k + 1 + B y b 2 2 , λ k + 1 = λ k + 1 2 s β A x k + 1 + B y k + 1 b ,
where r and s are limited to the following region
D = s , r | s 0 , 1 + 5 2 , r 1 , 1 & r + s > 0 , | r | < 1 + s s 2 .
As for more symmetric alternation method improvements of the ADMM algorithm, we provide the reader with references [17,18,19,20,21,22,23,24,25,26].
In addition, the convergence speed of the ADMM algorithm is directly related to the choice of the penalty parameter β . Therefore, improvements of the ADMM algorithm with a change to penalty parameter β k instead of the fixed penalty parameter β are proposed. Such as He, Yang, and Wang [27] proposed an adaptive penalty parameter scheme to solve linearly constrained monotone variational inequalities. In this paper, we, based on the idea of references [16,27], propose an adaptive penalty parameter alternate direction method of multipliers with attach relaxation factors and with the Lagrange multiplier updated twice at each iteration. The iteration format is shown in Algorithm 1.
Algorithm 1: Improved alternate direction multiplier algorithm to solve convex optimization problem (1)
Given error tolerance ε > 0 , controls parameter μ 0 , 1 , and relaxation factors ( r , s ) D defined in (Equation (8)). Choose parameter sequence τ k , which satisfies τ k 0 and k = 0 τ k < + . Given initial penalty parameter β 0 > 0 and initial approximation x 0 , y 0 , λ 0 X , Y , r . Set k = 0 .
Step 1. Computing
x k + 1 = arg min x X f ( x ) λ k T A x + B y k b + β k 2 A x + B y k b 2 2 ,(9)
λ k + 1 2 = λ k r β k A x k + 1 + B y k b ,(10)
y k + 1 = arg min y Y g ( y ) λ k + 1 2 T A x k + 1 + B y b + β k 2 A x k + 1 + B y b 2 2 , (11)
λ k + 1 = λ k + 1 2 s β k A x k + 1 + B y k + 1 b .(12)
Step 2. If max A x k + 1 + B y k + 1 b 2 , B y k y k + 1 2 ε , stop. Otherwise, go to Step 3.
Step 3. Updating penalty parameter
β k + 1 = 1 + τ k β k , i f   x k P X x k f x k A T λ k 2 < μ A x k + B y k b 2 , β k / 1 + τ k , i f   μ x k P X x k f x k A T λ k 2 > A x k + B y k b 2 , β k , otherwise . (13)
Step 4. Set k = k + 1 , and go to Step 1.
We choose the algorithm described in [28] with the necessary modifications to solve the x-subproblem and y-subproblem in Algorithm 1. The iteration method to solve the x-subproblem can be described as follows in Algorithm 2.
Algorithm 2: Algorithm to solve x-subproblem (9) or y-subproblem (11)
Given constants θ ( 0 , 1 ) , ν ( 0 , 1 ) , l ( 1 , ) , error tolerance ε , and initial approximation x 0 X . Set η 0 = 0.9 and k = 0 .
Step 1. Computing x ¯ k + 1 = P X x k η k F x k , where F x = f x A T λ k + β k A T A x + B y k b .
Step 2. If ψ k = η k F x k F x ¯ k + 1 2 / x k x ¯ k + 1 2 ν , computing
x k + 1 = P X x k ρ k F x ¯ k + 1 ,
  where ρ k = l η k e k T d k / d k 2 2 , e k = x k x ¯ k + 1 d k = e k η k [ F ( x k ) F ( x ¯ k + 1 ) ] .
  If x k + 1 P X x k + 1 F x k + 1 2 ε , stop (in this case, x k + 1 is an approximate solution of the x-subproblem in Algorithm 1). Otherwise, define
     η k + 1 3 / 2 η k , i f   ψ k θ η k , otherwise and set k = k + 1 . Go to Step 1.
  If all of the above are not true, go to Step 3.
Step 3. Define η k 2 / 3 η k min 1 , 1 / ψ k and set x ¯ k + 1 = P X x k η k F x k . Go to Step 2.
The rest of this paper is organized as follows: In Section 2, the global convergence of the proposed algorithm is proven. In Section 3, some numerical comparison experiments with existing algorithms are given. Finally, we draw some conclusions.

2. Convergence of Algorithm 1

To conveniently prove the global convergence of Algorithm 1, we first give the following Lemmas 1–3.
Lemma 1.
([29]). Let Ω be a nonempty closed convex subset on n , and F ( u ) is a continuous closed convex function on Ω ; then, a sufficient necessary condition for u * to be a solution of the optimization problem
  min u Ω F ( u )
is that u * is a solution of the variational inequality
u u * T F u * 0 ,   u Ω
Lemma 2.
([26,27]). A sufficient necessary condition for the vector u * to be a solution of the variational inequality (Equation (14)) is that u * is also a solution of the projection equation
  u = P Ω u F u ,
where P Ω u is the projection of u onto Ω ; that is, P Ω u = arg min y Ω y u 2 .
Lemma 3.
([30]). Let Ω be a nonempty closed convex subset on n , then we have
P Ω v P Ω u 2 v u 2 ,   v , u Ω .
Let
w = x y λ ,   Q w = f ( x ) A T λ g ( y ) B T λ A x + B y b ( x , y , λ ) Ω = X × Y × r ,
then the problem (Equation (1)) is equivalent to finding w * such that
w w * Q T w * 0 ,   ( x , y , λ ) Ω .
It is known from Lemma 2 that w * is a solution of the variational inequality (Equation (15)) if and only if w * is the zero point of
e w : = e x w e y w e λ w = x P X x f ( x ) A T λ y P Y y g ( y ) B T λ A x + B y b .
Lemma 4.
Assume that the sequence { w k } is generated by Algorithm 1; then, the sufficient necessary condition for w k + 1 to be a solution of the variational inequality (Equation (15)) is
A x k + 1 + B y k + 1 b 2 = B y k B y k + 1 2 = 0 .
Proof. 
It follows from Lemma 1 that solving Equations (9) and (11) is equivalent to finding x k + 1 X and y k + 1 Y such that
x x k + 1 T f x k + 1 A T λ k β k A x k + 1 + B y k b 0 ,   x X ,
y y k + 1 T g y k + 1 B T λ k + 1 2 β k A x k + 1 + B y k + 1 b 0 ,   y Y .
Thus, it follows from Lemma 2 that
x k + 1 = P X x k + 1 f x k + 1 A T λ k β k A x k + 1 + B y k b ,
y k + 1 = P Y y k + 1 g ( y k + 1 ) B T λ k + 1 2 β k A x k + 1 + B y k + 1 b .
Noting that Equations (10) and (12) hold, we have
λ k + 1 = λ k r β k A x k + 1 + B y k b s β k A x k + 1 + B y k + 1 b .
Therefore, we have Lemma 3, that
e ( w k + 1 ) 2 2 = x k + 1 P X x k + 1 f x k + 1 A T λ k + 1 y k + 1 P Y y k + 1 g y k + 1 B T λ k + 1 A x k + 1 + B y k + 1 b 2 2 A T λ k λ k + 1 β k A x k + 1 + B y k b B T λ k + 1 2 λ k + 1 β k A x k + 1 + B y k + 1 b A x k + 1 + B y k + 1 b 2 2 = A T ( r + s 1 ) β k A x k + 1 + B y k + 1 b ( 1 r ) β k B y k B y k + 1 B T ( s 1 ) β k A x k + 1 + B y k + 1 b A x k + 1 + B y k + 1 b 2 2 [ ( r + s 1 ) 2 β k 2 A F 2 + ( s 1 ) 2 β k 2 B F 2 + 1 ] A x k + 1 + B y k + 1 b 2 2 + ( 1 r ) 2 β k 2 A F 2 B y k B y k + 1 2 2 .
From above inequality, and noting that r and s are fixed constants and { β k } is a bounded sequence of positive numbers, we know that w k + 1 to be the solution of the variational inequality (Equation (15)) if and only if Equation (16) holds. □
Lemma 5.
Assume that the sequence { w k } is generated by Algorithm 1 and w * is a solution of the variational inequality (Equation (14)), and we have
λ k λ * T A x k + 1 + B y k + 1 b β k A x k + 1 + B y k b T A x k + 1 A x * + r β k A x k + 1 + B y k b T B y k + 1 B y * + β k A x k + 1 + B y k + 1 b T B y k + 1 B y * .
Proof. 
Since w * is a solution of the variational inequality (Equation (15)), we have A x * + B y * b = 0 and
x k + 1 x * T f ( x * ) A T λ * 0 ,
y k + 1 y * T g ( y * ) B T λ * 0 .
On the other hand, we have by Equations (17) and (18) that
x * x k + 1 T f x k + 1 A T λ k β k A x k + 1 + B y k b 0 ,
y * y k + 1 T g y k + 1 B T λ k + 1 2 β k A x k + 1 + B y k + 1 b 0 .
Noting that the gradient of convex function is a monotone function, we have from Equations (20)–(22) that
x k + 1 x * T A T λ k λ * β k A x k + 1 + B y k b 0
Noting that the gradient of convex function is a monotone function, we have from Equations (21)–(23) that
y k + 1 y * T B T λ k + 1 2 λ * β k A x k + 1 + B y k + 1 b 0 .
Noting that A x * + B y * b = 0 , we know from Equation (11), Equations (24) and (25) that Equation (19) holds. □
Lemma 6.
For k 1 , we have
β k A x k + 1 + B y k + 1 b T B y k B y k + 1 1 s 1 + r β k 1 A x k + B y k b T B y k B y k + 1 r 1 + r β k B y k B y k + 1 2 2 .
Proof. 
From Equation (18) we have
y k y k + 1 T g y k + 1 B T λ k + 1 2 β k A x k + 1 + B y k + 1 b 0 ,
and on the other hand, by setting k = k 1 and y = y k + 1 in Equation (18), we have
y k + 1 y k T g y k B T λ k 1 2 β k 1 A x k + B y k b 0 .
Noting that Equation (6) holds, we have
λ k = λ k 1 2 s β k 1 A x k + B y k b .
Combining Equation (3) and Equation (29), we obtain
λ k + 1 2 = λ k 1 2 r β k A x k + 1 + B y k b s β k 1 A x k + B y k b .
From Equations (27), (28), and (30) and the monotonicity of the gradient of the convex function, we know that Equation (26) holds. □
Let
H k = r + s r s β k B T 2 B r β k B T r β k B I m , ( r , s ) D ,
then H k is a symmetric semi-definite matrix, and defining
x H k 2 = x T H k x ,
we have the following Lemma 7.
Lemma 7.
Assume that the sequence { w k } is generated by Algorithm 1 and w * is the solution of the variational inequality (Equation (14)), and we have
v k + 1 v * H k 2 v k v * H k 2 φ w k ,
where v k = y k λ k , v * = y * λ * and
φ w k = r + s 2 r + s β k 2 A x k + 1 + B y k + 1 b 2 2 + ( 1 r ) 2 ( r + s ) 1 + r β k 2 B y k B y k + 1 2 2 2 ( 1 r ) ( 1 s ) ( r + s ) 1 + r β k 1 β k A x k + B y k b T B y k B y k + 1 .
Proof. 
By Equations (3) and (4) and the definition of H k and x H k , we have
y k + 1 y k λ k + 1 λ k H k 2 = y k y k + 1 r + s β k A x k + 1 + B y k + 1 b + r β k B y k B y k + 1 H k 2 = r + s r s β k 2 B y k B y k + 1 2 2 + r + s β k A x k + 1 + B y k + 1 b + r β k B y k B y k + 1 2 2 2 r r + s β k 2 A x k + 1 + B y k + 1 b T B y k B y k + 1 2 r 2 β k 2 B y k B y k + 1 T B y k B y k + 1 = r + s 2 β k 2 A x k + 1 + B y k + 1 b 2 2 + 1 r r + s β k 2 B y k B y k + 1 2 2 .
By the definition of H k , Lemmas 5 and 6, and noting that A x * + B y * b = 0 , we have
2 y k y * λ k λ * T H k y k + 1 y k λ k + 1 λ k = 2 r + s r s β k 2 B y k B y * T B y k B y k + 1 + 2 r β k λ k λ * T B y k B y k + 1 + 2 r r + s β k 2 B y k B y * T A x k + 1 + B y k + 1 b + 2 r 2 β k 2 B y k B y * T B y k B y k + 1 2 r + s β k λ k λ * T A x k + 1 + B y k + 1 b 2 r β k λ k λ * T B y k B y k + 1 = 2 1 r r + s β k 2 B y k + 1 B y * T B y k B y k + 1 2 1 r r + s β k 2 B y k B y k + 1 2 2 + 2 r r + s β k 2 B y k B y * T A x k + 1 + B y k + 1 b 2 r + s β k λ k λ * T A x k + 1 + B y k + 1 b 2 1 r r + s β k 2 B y k + 1 B y * T B y k B y k + 1 2 1 r r + s β k 2 B y k B y k + 1 2 2 + 2 r r + s β k 2 B y k B y * T A x k + 1 + B y k + 1 b 2 r r + s β k 2 A x k + 1 + B y k b T B y k + 1 B y * 2 r + s β k 2 A x k + 1 + B y k b T A x k + 1 A x * 2 r + s β k 2 A x k + 1 + B y k + 1 b T B y k + 1 B y * = 2 1 r r + s β k 2 B y k B y k + 1 2 2 2 r + s β k 2 A x k + 1 + B y k + 1 b 2 2 2 r + s β k 2 B y k B y k + 1 T B y k + 1 B y * 2 r + s β k 2 B y k B y k + 1 T A x k + 1 A x * + 2 r r + s β k 2 A x k + 1 + B y k + 1 b T B y k B y k + 1 = 2 1 r r + s β k 2 B y k B y k + 1 2 2 2 r + s β k 2 A x k + 1 + B y k + 1 b 2 2 2 r + s β k 2 A x k + 1 + B y k + 1 b T B y k B y k + 1 + 2 r r + s β k 2 A x k + 1 + B y k + 1 b T B y k B y k + 1 = 2 1 r r + s β k 2 B y k B y k + 1 2 2 2 r + s β k 2 A x k + 1 + B y k + 1 b 2 2 2 1 r r + s β k 2 A x k + 1 + B y k + 1 b T B y k B y k + 1 2 ( r + s ) β k 2 A x k + 1 + B y k + 1 b 2 2 2 ( 1 r ) ( r + s ) 1 + r β k 2 B y k B y k + 1 2 2 2 ( 1 r ) ( 1 s ) ( r + s ) 1 + r β k 1 β k A x k + B y k b T B y k B y k + 1 2 ( r + s ) β k 2 A x k + 1 + B y k + 1 b 2 2 2 ( 1 r ) ( r + s ) 1 + r β k 2 B y k B y k + 1 2 2 + 2 ( 1 r ) ( 1 s ) ( r + s ) 1 + r β k 1 β k A x k + B y k b T B y k B y k + 1 .
Therefore, we have
v k + 1 v * H k 2 = y k y * λ k λ * H k 2 + y k + 1 y k λ k + 1 λ k H k 2 + 2 y k y * λ k λ * T H k y k + 1 y k λ k + 1 λ k v k v * H k 2 ( r + s ) ( 2 r s ) A x k + 1 + B y k + 1 b 2 2 ( 1 r ) 2 ( r + s ) 1 + r B y k B y k + 1 2 2 + 2 ( 1 r ) ( 1 s ) ( r + s ) 1 + r β k 1 β k ( A x k + 1 + B y k + 1 b ) T ( B y k B y k + 1 ) = v k v * H k 2 φ ( w k ) .
Hence, the inequality Equation (31) holds. □
To prove the global convergence of Algorithm 1, we divided the domain D defined in Equation (8) into the following five parts:
D 1 = r , s | r 1 , 1 , s 0 , 1 , r + s > 0 , D 2 = r , s | r 1 , 1 , s = 1 , D 3 = r , s | r 1 , 0 , s 1 , 1 + 5 2 , r < 1 + s s 2 , D 4 = r , s | r = 0 , s 1 , 1 + 5 2 , D 5 = r , s | r 0 , 1 , s 1 , 1 + 5 2 , r < 1 + s s 2 .
Obviously,
D = n = 1 5 D n   and   D i D j = ,   i , j 1 , 2 , 3 , 4 , 5 , i j
Lemma 8.
Assume that the sequence { w k } is generated by Algorithm 1; then, for any r , s D , there exist constants C 0 , C 1 , C 2 > 0 and ξ , such that φ ( w k ) defined in Equation (32) satisfies
φ w k ξ C 0 β k 2 A x k + 1 + B y k + 1 b 2 2 β k 1 2 A x k + B y k b 2 2 + C 1 β k 2 A x k + 1 + B y k + 1 b 2 2 + C 2 β k 2 B y k y k + 1 2 2 ,
where  ξ = 0 , r , s D 2 , 1 , r , s D \ D 2 .
Proof. 
In the case ( r , s ) D 1 , we have by the Cauchy–Schwarz inequality for the last term of φ w k in Lemma 7 that
2 ( 1 r ) ( 1 s ) ( r + s ) 1 + r β k 1 β k A x k + B y k b T B y k B y k + 1 1 s r + s β k 1 2 A x k + B y k b 2 2 + ( 1 r ) 2 ( 1 s ) ( r + s ) ( 1 + r ) 2 β k 2 B y k B y k + 1 2 2 .
Thus, we have
φ w k r + s 2 r + s β k A x k + 1 + B y k + 1 b 2 2 2 + ( 1 r ) 2 ( r + s ) 1 + r β k B y k B y k + 1 2 2 2 1 s r + s β k 1 2 A x k + B y k b 2 2 ( 1 r ) 2 ( 1 s ) ( r + s ) ( 1 + r ) 2 β k B y k B y k + 1 2 2 2 = 1 s r + s β k A x k + 1 + B y k + 1 b 2 2 2 β k 1 2 A x k + B y k b 2 2 + 1 r r + s β k A x k + 1 + B y k + 1 b 2 2 2 + ( 1 r ) ( r + s ) 1 + r 2 β k B y k B y k + 1 2 2 2 .
Let C 0 , C 1 , C 2 in Equation (33) as
C 0 = r + s 1 s > 0 ,   C 1 = r + s 1 r > 0 , C 2 = ( 1 r ) ( r + s ) 1 + r 2 > 0 ,
then the inequality Equation (33) holds.
In the case r , s D 2 , that is, r ( 1 , 1 ) ,   s = 1 , we have by Lemma 7 that
φ w k = 1 + r 1 r β k 2 A x k + 1 + B y k + 1 b 2 2 + 1 r 2 β k 2 B y k B y k + 1 2 2 .
Let   C 1 = 1 + r 1 r > 0 , C 2 = 1 r 2 > 0 , and we have
φ w k C 1 β k 2 A x k + 1 + B y k + 1 b 2 2 + C 2 β k 2 B y k y k + 1 2 2 .
In the case ( r , s ) D 3 , we have by the Cauchy–Schwarz inequality for the last term of φ w k in Lemma 7 that
2 ( 1 r ) ( 1 s ) ( r + s ) 1 + r β k 1 β k A x k + B y k b T B y k B y k + 1 r + s T 1 r + s β k 1 2 A x k + B y k b 2 2 + ( 1 r ) 2 ( 1 s ) 2 ( r + s ) ( 1 + r ) 2 [ T 1 ( r + s ) ] β k B y k B y k + 1 2 2 2 ,
where T 1 = r + s + ( s 2 s ) ( 2 s ) 1 + r > 0 . Thus, we have
φ w k r + s T 1 r + s β k A x k + 1 + B y k + 1 b 2 2 2 β k 1 2 A x k + B y k b 2 2 + r + s 2 T 1 β k A x k + 1 + B y k + 1 b 2 2 2 + ( 1 r ) 2 ( r + s ) 1 + r ( 1 r ) 2 ( 1 s ) 2 ( r + s ) ( 1 + r ) 2 [ T 1 ( r + s ) ] β k B y k B y k + 1 2 2 2 .
And let C 0 , C 1 , C 2 in Equation (33) as
C 0 = r + s T 1 r + s = ( r + s ) ( s 2 s ) ( 2 s ) 1 + r > 0 ,   C 1 = r + s 2 T 1 > r r + s > 0 , C 2 = ( 1 r ) 2 ( r + s ) 1 + r ( 1 r ) 2 ( 1 s ) 2 ( r + s ) ( 1 + r ) 2 [ T 1 ( r + s ) ] = ( 1 r ) 2 ( r + s ) ( 1 + s s 2 ) s ( 1 + r ) ( 2 s ) > 0 ,
then the inequality Equation (33) holds.
In the case ( r , s ) D 4 , we have by the Cauchy–Schwarz inequality for the last term of φ w k in Lemma 7 that
2 s 1 s β k 1 β k A x k + B y k b T B y k B y k + 1 s T 2 s β k 1 2 A x k + B y k b 2 2 + s ( 1 s ) 2 T 2 s B y k B y k + 1 2 2 ,
where T 2 = 1 3 s 2 s + 5 . Thus, we have
φ w k s T 2 s β k A x k + 1 + B y k + 1 b 2 2 2 β k 1 2 A x k + B y k b 2 2 + s 2 T 2 β k A x k + 1 + B y k + 1 b 2 2 2 + s 1 ( 1 s ) 2 T 2 s B y k B y k + 1 2 2 .
And let C 0 , C 1 , C 2 in Equation (33) as
C 0 = s T 2 s = 1 3 s s 2 4 s + 5 > 0 ,   C 1 = s 2 T 2 = 1 3 s 1 + s s 2 > 0 , C 2 = s 1 ( 1 s ) 2 T 2 s = 2 s ( 1 + s s 2 ) 1 + ( s 2 ) 2 > 0 ,
then the inequality Equation (33) holds.
In the case ( r , s ) D 5 , we have by the Cauchy–Schwarz inequality for the last term of φ w k in Lemma 7 that
2 ( 1 r ) ( 1 s ) ( r + s ) 1 + r β k 1 β k A x k + B y k b T B y k B y k + 1 r + s T 3 r + s β k 1 2 A x k + B y k b 2 2 + ( 1 r ) 2 ( 1 s ) 2 ( r + s ) ( 1 + r ) 2 [ T 2 ( r + s ) ] β k B y k B y k + 1 2 2 2 ,
where T 3 = r + s + 1 s 2 . Thus, we have
φ w k r + s T 3 r + s β k A x k + 1 + B y k + 1 b 2 2 2 β k 1 2 A x k + B y k b 2 2 + r + s 2 T 3 β k A x k + 1 + B y k + 1 b 2 2 2 + ( 1 r ) 2 ( r + s ) 1 + r ( 1 r ) 2 ( 1 s ) 2 ( r + s ) ( 1 + r ) 2 [ T 3 ( r + s ) ] β k B y k B y k + 1 2 2 2 .
And let C 0 , C 1 , C 2 in Equation (33) as
C 0 = r + s T 3 r + s = r + s 1 s 2 > 0 , C 1 = r + s 2 T 3 > 0 , C 2 = ( 1 r ) 2 ( r + s ) 1 + r ( 1 r ) 2 ( 1 s ) 2 ( r + s ) ( 1 + r ) 2 [ T 3 ( r + s ) ] = r ( 1 r ) 2 ( r + s ) ( 1 + r ) 2 > 0 ,
then the inequality Equation (33) holds. □
Lemma 9.
Assume that the sequence { w k } is generated by Algorithm 1; then, we have that there exist constants t 1 , t 2 0 , ( 1 r ) ( r + s ) r + s r s such that
t 1 r + s r s β k 1 2 B y k B y * 2 2 v k v * H k 1 2   and   t 2 λ k λ * 2 2 v k v * H k 1 2 ,
where  ( r , s ) D .
Proof. 
By the definition of H k and x H k , we have
v k v * H k 1 2 t 1 r + s r s β k 1 2 B y k B y * 2 2 = y k y * λ k λ * T 1 t 1 r + s r s β k 1 2 B T B r β k 1 B T r β k 1 B I m y k y * λ k λ * .
Since t 1 0 , ( 1 r ) ( r + s ) r + s r s , we know that matrix
1 t 1 r + s r s β k 1 2 B T B r β k 1 B T r β k 1 B I m
is a positive semi-definite matrix. And thus v k v * H k 1 2 t 1 r + s r s β k 1 2 B y k B y * 2 2 0 . Analogously, we have t 2 λ k λ * 2 2 v k v * H k 1 2 .□
Lemma 10.
Assume that the sequence w k is generated by Algorithm 1; then, there exists a constant C ¯ > 0 such that v k v * H k 1 2 < C ¯ holds for all positive integer numbers k.
Proof. 
If β k > β k 1 , we have by Equation (26), Lemmas 7–9, and the definition of H k and x H k that
v k + 1 v * H k 2 + ξ C 0 β k A x k + 1 + B y k + 1 b 2 2 2 v k v * H k 2 + ξ C 0 β k 1 2 A x k + B y k b 2 2 C 1 β k A x k + 1 + B y k + 1 b 2 2 2 + C 2 β k B y k y k + 1 2 2 2 v k v * H k 2 + ξ C 0 β k 1 2 A x k + B y k b 2 2 = 1 + τ k 1 v k v * H k 1 2 + y k y * λ k λ * T r + s r s τ k 1 + τ k 1 2 β k 1 2 B T B 0 0 τ k 1 I m y k y * λ k λ * + ξ C 0 β k 1 2 A x k + B y k b 2 2 = 1 + τ k 1 v k v * H k 1 2 + τ k 1 + τ k 1 2 r + s r s β k 1 2 B y k B y * 2 2 τ k 1 λ k λ * 2 2 + ξ C 0 β k 1 2 A x k + B y k b 2 2 1 + τ k 1 v k v * H k 1 2 + τ k 1 t 1 1 + τ k 1 v k v * H k 1 2 + ξ C 0 β k 1 2 A x k + B y k b 2 2 ( 1 + τ k 1 ) ( 1 + τ k 1 t 1 ) v k v * H k 1 2 + ξ C 0 β k 1 2 A x k + B y k b 2 2 .
If β k < β k 1 , we have by Equation (26), Lemmas 7–9, and the definition of H k and x H k that
v k + 1 v * H k 2 + ξ C 0 β k A x k + 1 + B y k + 1 b 2 2 2 v k v * H k 2 + ξ C 0 β k 1 2 A x k + B y k b 2 2 = 1 1 + τ k 1 v k v * H k 1 2 + y k y * λ k λ * T r + s r s τ k 1 β k 1 2 1 + τ k 1 2 B T B 0 0 τ k 1 1 + τ k 1 I m y k y * λ k λ * + ξ C 0 β k 1 2 A x k + B y k b 2 2 = 1 1 + τ k 1 v k v * H k 1 2 + τ k 1 1 + τ k 1 λ k λ * 2 2 τ k 1 r + s r s 1 + τ k 1 2 β k 1 2 B y k B y * 2 2 + ξ C 0 β k 1 2 A x k + B y k b 2 2 t 2 + τ k 1 t 2 1 + τ k 1 v k v * H k 1 2 + ξ C 0 β k 1 2 A x k + B y k b 2 2 .
Let
> = k | β k > β k 1 ,   < = k | β k < β k 1 ,   = = k | β k = β k 1
with
> < = ,   > < = = 1 , 2 , , k ,
then
v k + 1 v * H k 2 i > ( 1 + τ i 1 ) ( 1 + τ i 1 t 1 ) j < t 2 + τ j 1 t 2 ( 1 + τ j 1 ) v 1 v * H 0 2 + ξ C 0 β 0 2 A x 1 + B y 1 b 2 2 .
Noting that i = 1 ( 1 + τ i 1 ) ( 1 + τ i 1 t 1 ) < and i = 1 t 2 + τ i 1 t 2 ( 1 + τ i 1 ) < , we know that there exists a constant C ¯ > 0 such that v k v * H k 1 2 < C ¯ holds for all positive integer numbers k. □
Theorem 1.
Assume that the sequence w k is generated by Algorithm 1; then, we have
lim k { A x k + 1 + B y k + 1 b 2 2 + B y k B y k + 1 2 2 } = 0 .
And so w ^ = lim k w k is a solution of the variational inequality (Equation (15)). Hence ( x ^ , y ^ ) , is a solution of the optimization problem (Equation (1)).
Proof. 
If β k > β k 1 , we have by Lemmas 7, 8, and 10 that
C 1 β k A x k + 1 + B y k + 1 b 2 2 2 + C 2 β k B y k y k + 1 2 2 2 v k v * H k 2 v k + 1 v * H k 2 + ξ C 0 ( β k 1 A x k + B y k b 2 2 2 β k A x k + 1 + B y k + 1 b 2 2 2 ) v k v * H k 1 2 v k + 1 v * H k 2 + ξ C 0 ( β k 1 A x k + B y k b 2 2 2 β k A x k + 1 + B y k + 1 b 2 2 2 ) + ( 1 + τ k 1 ) ( 1 + τ k 1 t 1 ) v k v * H k 1 2 .
Hence, we have
min ( C 1 , C 2 ) i = 1 β i ( 2 A x i + 1 + B y i + 1 b 2 2 + B y i y i + 1 2 2 ) i = 1 β i ( 2 C 1 A x i + 1 + B y i + 1 b 2 2 + C 2 B y i y i + 1 2 2 ) v 1 v * H 0 2 v v * H 2 + ξ C 0 β 0 A x 1 + B y 1 b 2 2 2 β 2 A x + B y b 2 2 + C ¯ i = 1 ( 1 + τ i 1 ) ( 1 + τ i 1 t 1 ) .
Noting that i = 1 ( 1 + τ i 1 ) ( 1 + τ i 1 t 1 ) < , the above equation is equivalent to
min ( C 1 , C 2 ) i = 1 β i ( 2 A x i + 1 + B y i + 1 b 2 2 + B y i y i + 1 2 2 ) + ξ C 0 β 2 A x + B y b 2 2 v 1 v * H 0 2 v v * H 2 + ξ C 0 β 0 A x 1 + B y 1 b 2 2 2 + C ¯ i = 1 ( 1 + τ i 1 ) ( 1 + τ i 1 t 1 ) < .
So, we have lim k { A x k + 1 + B y k + 1 b 2 2 + B y k B y k + 1 2 2 } = 0 . .
If β k < β k 1 , we have by Lemmas 7, 8, and 10 that
C 1 β k A x k + 1 + B y k + 1 b 2 2 2 + C 2 β k B y k y k + 1 2 2 2 v k v * H k 2 v k + 1 v * H k 2 + ξ C 0 ( β k 1 A x k + B y k b 2 2 2 β k A x k + 1 + B y k + 1 b 2 2 2 ) v k v * H k 1 2 v k + 1 v * H k 2 + ξ C 0 ( β k 1 A x k + B y k b 2 2 2 β k A x k + 1 + B y k + 1 b 2 2 2 ) + t 2 + τ k 1 t 2 1 + τ k 1 v k v * H k 1 2 .
Hence, we have
min C 1 , C 2 i = 1 β i 2 ( A x i + 1 + B y i + 1 b 2 2 + B y i y i + 1 2 2 ) i = 1 β i 2 ( C 1 A x i + 1 + B y i + 1 b 2 2 + C 2 B y i y i + 1 2 2 ) v 1 v * H 0 2 v v * H 2 + ξ C 0 β 0 2 A x 1 + B y 1 b 2 2 β 2 A x + B y b 2 2 + C ¯ i = 1 t 2 + τ i 1 t 2 1 + τ i 1 .
Noting that i = 1 t 2 + τ i 1 t 2 1 + τ i 1 < , the above equation is equivalent to
min C 1 , C 2 i = 1 β i 2 A x i + 1 + B y i + 1 b 2 2 + B y i y i + 1 2 2 + ξ C 0 β 2 A x + B y b 2 2 v 1 v * H 0 2 v v * H 2 + ξ C 0 β 0 2 A x 1 + B y 1 b 2 2 + C ¯ i = 1 t 2 + τ i 1 t 2 ( 1 + τ i 1 ) < .
So, we have lim k { A x k + 1 + B y k + 1 b 2 2 + B y k B y k + 1 2 2 } = 0 .
If β k = β k 1 = β , we have by Lemmas 7 and 8 that
C 1 β k A x k + 1 + B y k + 1 b 2 2 2 + C 2 β k B y k y k + 1 2 2 2 v k v * H k 2 v k + 1 v * H k 2 + ξ C 0 ( β k 1 A x k + B y k b 2 2 2 β k A x k + 1 + B y k + 1 b 2 2 2 ) .
Hence, we have
min C 1 , C 2 i = 1 β i 2 ( A x i + 1 + B y i + 1 b 2 2 + B y i y i + 1 2 2 ) i = 1 β i 2 ( C 1 A x i + 1 + B y i + 1 b 2 2 + C 2 B y i y i + 1 2 2 ) v 1 v * H k 2 v v * H k 2 + ξ C 0 β 0 2 A x 1 + B y 1 b 2 2 β 2 A x + B y b 2 2 .
with its equivalence to
min C 1 , C 2 i = 1 β i 2 A x i + 1 + B y i + 1 b 2 2 + B y i y i + 1 2 2 + ξ C 0 β 2 A x + B y b 2 2 v 1 v * H k 2 v v * H k 2 + ξ C 0 β 0 2 A x 1 + B y 1 b 2 2 < .
So, we have lim k { A x k + 1 + B y k + 1 b 2 2 + B y k B y k + 1 2 2 } = 0 . By that discussed above and Lemma 4, we know that w ^ = lim k w k is a solution of the variational inequality (Equation (19)). Hence, ( x ^ , y ^ ) , is a solution of the optimization problem (Equation (1)). □

3. Numerical Experiments

In this section, some numerical experiments will be tested to illustrate the efficiency of Algorithm 1. The experiments are divided into two parts. In the first part, we first analyze the relationship between the parameters r and s and the convergence of Algorithm 1 and give the selection range of the best parameters r and s. Then, we give a comparison of Algorithm 1 with the algorithm proposed in [15], the algorithm proposed in [16], and the algorithm proposed in [27] to solve a convex quadratic programming problem. In the second part, we give the numerical comparison between Algorithm 1 and the algorithm proposed in [16] in image restoration. All codes were written in MATLAB R2019a, and all experiments were performed on a personal computer with an Intel(R) Core i7-6567U processor (3.3 GHz) and 8 GB memory.

3.1. Convex Quadratic Programming Problem

To specify the problem (Equation (1)), we choose a separable convex quadratic programming problem proposed in [31], where
f x = 1 2 x T M 1 x + q 1 x T and   g y = 1 2 y T M 2 y + q 2 y T ,
and here the matrices M i = Q i T Q i   ( i = 1 , 2 ) , and Q 1 m × m , and Q 2 n × n are randomly generated square matrices with all entries are chosen in intervals [−5, 5]. The vectors q i   ( i = 1 , 2 ) are generated from a uniform distribution in the interval [−500, 500]. For the linear constraint, A = 1 1 × m and B = 1 1 × n (i.e., all entries of A and B are taken as 1), and b = m + n / 100 . The convex set X m is defined as a sphere with the origin at the center of the circle and radius r = 1 , and the convex set Y n is defined as the box in n with all entries are chosen in intervals [0, 5].
Since the convergence of the original ADMM algorithm is well proven, the algorithm proposed in [11] may not converge in some cases, the symmetric ADMM algorithm proposed in [16] has a larger step size relative to the algorithm in [16], and the adaptive penalty parameter has been less studied. Therefore, we only compare Algorithm 1 numerically with the algorithms presented in references [14], [16] and [27] denoted respectively by Algorithm 14, 16 and 27.
To ensure the fairness of the experiment, the parameter γ of Algorithm 14 and 27 is chosen as the optimal value γ = 1.6 ; the parameters r and s of the algorithm proposed in [18] are chosen as the optimal value r = 0.8, s = 1.17; and the initial penalty parameter β of all algorithms mentioned above is chosen as 1. The initial iterations for all tested methods are zero vectors, and the stopping criterion is chosen as
max A x k + 1 + B y k + 1 b 2 , B y k B y k + 1 2 ε = 10 6
For Algorithm 1, we take
τ k = 0.3 , i f   k k max , 0 , o t h e r w i s e .
The parameters of Algorithm 2, which solves the x-subproblem and the y-subproblem of Algorithm 1, are chosen as θ = 0.1 , ν = 0.8 , and l = 1.8 . Table 1 reports the iteration numbers (Iter) and computation time (Time) of Algorithm 1 with different parameters r and s, and the best estimate of the parameters r and s can be chosen as r = 0.95, s = 1.12. And so, in the following numerical comparison experiment, the parameters r and s of Algorithm 1 are chosen as r = 0.95, s = 1.12.
In Table 2, we show the iteration numbers (Iter), the computational time (Time) of Algorithm 1, and Algorithm14, 16 and 27 for solving the separable convex quadratic programming problem proposed in [31]. To investigate the stability and accuracy of the algorithms, we tested 8 different sets of m and n values throughout the experiment, running each set of data 10 times and averaging the final results. In Figure 1, we plot the comparison in terms of objective function values (Obj) and computing time with Algorithm 1, and Algorithm14, Algorithm16 and Algorithm27 for solving the separable convex quadratic programming problem proposed in [31].
Based on the tests reported in Table 2, Figure 1, and many other performed unreported tests that show similar patterns, we know that Algorithm 1 is more efficient than other existing algorithms.

3.2. Image Restoration Problem

In this subsection, we consider the total variational image deblurring model proposed in [16], whose discretized version can be written as
min y A y 1 + λ 2 B x z 2
where y R n represents a digital clean image, z R n is a corrupted input image, A : = 1 , 2 : R n R n × R n is the discrete gradient operator, and 1 : R n R n and 2 : R n R n are the discrete derivatives in the horizontal and vertical directions, respectively. B : R n R n is the matrix representation of a spatially invariant blurring operator, λ > 0 is a constant balancing the data fidelity and total variational regularization terms, and 1 defined on R n × R n is given by
x 1 = { i , j } = 1 n x i , j 1 2 + x i , j 2 2 , x = ( x 1 , x 2 ) R n × R n ,
This is a basic model for various more advanced image processing tasks, and it has been studied extensively in the literature. Introducing the auxiliary variable x, we can reformulate Equation (34) as
min x , y x 1 + λ 2 B y z 2   s . t   x A y = 0
which is a special case of the generic model (Equation (1)) under discussion, and thus Algorithm 1 is applicable.
We test two images lena.tif (512 × 512) and man.pgm (512 × 512). These images are first corrupted by the blur operator with kernel size of 21 × 21, and then the blurred images are further corrupted by zero-white Gaussian noise with a standard derivation 0.002. In Figure 2, we list the original and degraded images. The quality of restored images is measured by the value of the SNR given by
SNR = 20   log 10 y y y ¯
where y is the original image and y ¯ is the recovered image. A larger SNR value means a higher quality of the restored image. Table 3 and Figure 3 report the comparison experiment between Algorithm 1 and the algorithm proposed in [16] for the image restoration problem.

4. Conclusions

The alternating direction method is one of the most attractive approaches for solving convex optimization problems with linear constraints and separable objective functions. Experience with applications has shown that the iteration numbers and computing time depend significantly on the penalty parameter for the linear constraint. The penalty parameters in the classical alternating direction method are a constant. In this paper, we, based on the idea of references [16,28], propose an adaptive penalty parameter alternate direction method of multipliers with attach relaxation factors and with the Lagrange multiplier updated twice at each iteration (Algorithm 1), which not only adaptively adjusts the penalty parameters per iteration based on the iteration message but also adds relaxation factors to the Lagrange multiplier update steps. For the proposed algorithm, that is, Algorithm 1, the global convergence is proven (that is, Theorem 1). Preliminary numerical experiments show that the technique of adaptive adjusting of penalty parameters per iteration and attaching relaxation factors in Lagrange multiplier updating steps are effective in practical applications (see Table 2 and Figure 1).

Author Contributions

Writing—original draft preparation, J.P. and Z.W.; writing-review and editing, S.Y. and Z.T. All authors have read and agreed to the published version of the manuscript.

Funding

This work was supported by National Natural Science Foundation of China (grant number 11961012) and the Special Research Project for Guangxi Young Innovative Talents (grant number AD20297063).

Data Availability Statement

The original contributions presented in the study are included in the article; further inquiries can be directed to the corresponding author.

Acknowledgments

The authors thank the anonymous reviewer for valuable suggestions that helped them to improve this article.

Conflicts of Interest

The authors declare no conflict of interest.

References

  1. Yang, J.F.; Zhang, Y. Alternating direction algorithms for l1-problems in compressive sensing. SIAM J. Sci. Comput. 2011, 33, 250–278. [Google Scholar] [CrossRef]
  2. Li, J.C. A parameterized proximal point algorithm for separable convex optimization. Optim. Lett. 2018, 12, 1589–1608. [Google Scholar]
  3. Tao, M.; Yuan, X.M. Recovering low-rank and sparse components of matrices from incomplete and noisy observations. SIAM J. Optim. 2011, 21, 57–81. [Google Scholar] [CrossRef]
  4. Hager, W.W.; Zhang, H.C. Convergence rates for an inexact ADMM applied to separable convex optimization. Comput. Optim. Appl. 2020, 77, 729–754. [Google Scholar] [CrossRef]
  5. Liu, Z.S.; Li, J.C.; Liu, X.N. A new model for sparse and Low-Rank matrix decomposition. J. Appl. Anal. Comput. 2017, 7, 600–616. [Google Scholar]
  6. Jiang, F.; Wu, Z.M. An inexact symmetric ADMM algorithm with indefinite proximal term for sparse signal recovery and image restoration problems. J. Comput. Appl. Math. 2023, 417, 114628. [Google Scholar] [CrossRef]
  7. Bai, J.C.; Hager, W.W.; Zhan, H.C. An inexact accelerated stochastic ADMM for separable convex optimization. Comput. Optim. Appl. 2022, 81, 479–518. [Google Scholar] [CrossRef]
  8. Boyd, S.; Parikh, N.; Chu, E.; Peleato, B.; Eckstein, J. Distributed optimization and statistical learning via the alternation direction method of multipliers. Found. Trends Mach. Learn. 2011, 3, 1–122. [Google Scholar] [CrossRef]
  9. Chan, T.F.; Glowinski, R. Finite Element Approximation and Iterative Solution of a Class of Mildly Nonlinear Elliptic Equations; Technical Report STAN-CS-78-674; Computer Science Department, Stanford University: Stanford, CA, USA, 1978. [Google Scholar]
  10. Hestenes, M.R. Multiplier and gradient methods. J. Optim. Theory Appl. 1969, 4, 303–320. [Google Scholar] [CrossRef]
  11. Lions, P.L.; Mercier, B. Splitting algorithms for the sum of two nonlinear operators. SIAM J. Numer. Anal. 1979, 16, 964–979. [Google Scholar] [CrossRef]
  12. Cai; X. J.; Gu, G.Y.; He, B.S.; Yuan, X.M. A proximal point algorithm revisit on the alternating direction method of multipliers. Sci. China Math. 2013, 56, 2179–2186. [Google Scholar] [CrossRef]
  13. Fortin, M.; Glowinski, R. Augmented Lagrangian Methods: Applications to the Numerical Solution of Boundary-Value Problems; North-Holland: Amsterdam, The Netherlands, 1983. [Google Scholar]
  14. Glowinski, R.; Karkkainen, T.; Majava, K. On the Convergence of Operator-Splitting Methods, in Numerical Methods for Scientific Computing, Variational Problems and Applications; Heikkola, E., Kuznetsov, P.Y., Neittaanm, P., Pironneau, O., Eds.; CIMNE: Barcelona, Spain, 2003; pp. 67–79. [Google Scholar]
  15. He, B.S.; Liu, H.; Wang, Z.R.; Yuan, X.M. A strictly contractive Peaceman-Rachford splitting method for convex programming. SIAM J. Optim. 2014, 24, 1011–1040. [Google Scholar] [CrossRef] [PubMed]
  16. He, B.S.; Ma, F.; Yuan, X.M. Convergence study on the symmetric version of ADMM with larger step sizes. SIAM J. Imaging Sci. 2016, 9, 1467–1501. [Google Scholar] [CrossRef]
  17. Luo, G.; Yang, Q.Z. A fast symmetric alternating direction method of multipliers. Numer. Math. Theory Meth. Appl. 2020, 13, 200–219. [Google Scholar]
  18. Li, X.X.; Yuan, X.M. A proximal strictly contractive Peaceman-Rachford splitting method for convex programming with applications to imaging. SIAM J. Imaging Sci. 2015, 8, 1332–1365. [Google Scholar] [CrossRef]
  19. Bai, J.C.; Li, J.C.; Xu, F.M.; Zhang, H.C. Generalized symmetric ADMM for separable convex optimization. Comput. Optim. Appl. 2018, 70, 129–170. [Google Scholar] [CrossRef]
  20. Wu, Z.M.; Li, M. An LQP-based symmetric Alternating direction method of multipliers with larger step sizes. J. Oper. Res. Soc. China 2019, 7, 365–383. [Google Scholar] [CrossRef]
  21. Chang, X.K.; Bai, J.C.; Song, D.J.; Liu, S.Y. Linearized symmetric multi-block ADMM with indefinite proximal regularization and optimal proximal parameter. Calcolo 2020, 57, 38. [Google Scholar] [CrossRef]
  22. Han, D.R.; Sun, D.F.; Zhang, L.W. Linear rate convergence of the alternating direction method of multipliers for convex composite programming. Math. Oper. Res. 2018, 43, 622–637. [Google Scholar] [CrossRef]
  23. Gao, B.; Ma, F. Symmetric alternating direction method with indefinite proximal regularization for linearly constrained convex optimization. J. Optim. Theory Appl. 2018, 176, 178–204. [Google Scholar] [CrossRef]
  24. Shen, Y.; Zuo, Y.N.; Yu, A.L. A partially proximal S-ADMM for separable convex optimization with linear constraints. Appl. Numer. Math. 2021, 160, 65–83. [Google Scholar] [CrossRef]
  25. Adona, V.A.; Goncalves, M.L.N. An inexact version of the symmetric proximal ADMM for solving separable convex optimization. Numer. Algor. 2023, 94, 1–28. [Google Scholar] [CrossRef]
  26. He, B.S.; Yang, H. Some convergence properties of a method of multipliers for linearly constrained monotone variational inequalities. Oper. Res. Lett. 1998, 23, 151–161. [Google Scholar] [CrossRef]
  27. He, B.S.; Yang, H.; Wang, S.L. Alternating direction method with Self-Adaptive penalty parameters for monotone variational inequalities. J. Optim. Theory Appl. 2000, 106, 337–356. [Google Scholar] [CrossRef]
  28. He, B.S.; Liao, L.Z. Improvements of some projection methods for monotone nonlinear variational inequalities. J. Optim. Theory Appl. 2002, 112, 111–128. [Google Scholar] [CrossRef]
  29. Palomar, D.P.; Scutari, G. Variational inequality theory a mathematical framework for multiuser communication systems and signal processing. In Proceedings of the International Workshop on Mathematical of Issues in Information Sciences-MIIS, Xi’an, China, 7–13 July 2012. [Google Scholar]
  30. He, B.S. A new method for a class of linear variational inequalities. Math. Programm. 1994, 66, 137–144. [Google Scholar] [CrossRef]
  31. Han, D.R.; He, H.J.; Yang, H.; Yuan, X.M. A customized Douglas–Rachford splitting algorithm for separable convex minimization with linear constraints. Numer. Math. 2014, 127, 167–200. [Google Scholar] [CrossRef]
Figure 1. Objective function curve of Algorithm 1 and the other three algorithms.
Figure 1. Objective function curve of Algorithm 1 and the other three algorithms.
Mathematics 13 00811 g001
Figure 2. First column is original images; second column is corrupted images.
Figure 2. First column is original images; second column is corrupted images.
Mathematics 13 00811 g002
Figure 3. First column is blurred images; second column is recovered images by Algorithm 1; and third column is recovered images by the algorithm in [18].
Figure 3. First column is blurred images; second column is recovered images by Algorithm 1; and third column is recovered images by the algorithm in [18].
Mathematics 13 00811 g003
Table 1. The numerical comparison of Algorithm 1 with different parameters r and s.
Table 1. The numerical comparison of Algorithm 1 with different parameters r and s.
mnr = 0.95
s = 1.12
r = 0.9
s = 1
r =
s = 1
r = 1
s = 1.5
r = 1.54
s = 0.2
r = 0.1
s = 1.57
IterTimeIterTimeIterTimeIterTimeIterTimeIterTime
80020010.70.4111.40.4410.80.4411.10.4617.70.6519.10.71
60040010.70.2611.40.2911.00.2711.60.2916.60.4218.70.49
160040010.01.5110.31.669.91.6111.01.7518.52.8418.02.82
120080010.01.1310.71.269.91.1610.71.2717.82.0518.02.08
24006009.76.9510.17.239.76.9810.87.7219.013.1317.112.30
180012009.62.7310.33.069.82.9211.23.3018.25.2217.05.03
32008009.815.3210.015.5910.115.8110.316.0119.029.2216.826.24
240016009.78.5410.18.8310.28.8310.89.3417.114.3015.613.86
Table 2. The numerical comparison of Algorithm 1 with the other three algorithms.
Table 2. The numerical comparison of Algorithm 1 with the other three algorithms.
mnobjAlgorithm 1Algorithm 16Algorithm 27Algorithm 14
r = 0.95, s = 1.12r = 0.8, s = 1.17 γ = 1.6 γ = 1.6
IterTimeIterTimeIterTimeIterTime
800200−9478.3910.60.4139.11.4120.20.6850.11.79
600400−8939.2710.70.2642.41.1920.00.5154.11.47
1600400−12,473.2710.01.5129.15.1318.93.0637.56.62
1200800−12,038.1110.01.1330.13.7719.02.1638.84.84
2400600−14,789.519.56.9521.715.5617.612.6127.920.01
18001200−14,312.399.82.7322.97.1918.05.3030.19.51
3200800−16,191.259.815.3218.328.1817.327.3223.836.43
24001600−14,365.079.78.5419.717.0317.314.9925.521.98
Table 3. The numerical comparison of Algorithm 1 and the algorithm in [16].
Table 3. The numerical comparison of Algorithm 1 and the algorithm in [16].
Lena (512 × 512)man (512 × 512)
IterTimeSNRIterTimeSNR
Algorithm 1
r = 0.95, s = 1.12
263.9323.02304.5117.02
Algorithm 16
r =0.8, s = 1.17
375.5423.00406.0817.00
Disclaimer/Publisher’s Note: The statements, opinions and data contained in all publications are solely those of the individual author(s) and contributor(s) and not of MDPI and/or the editor(s). MDPI and/or the editor(s) disclaim responsibility for any injury to people or property resulting from any ideas, methods, instructions or products referred to in the content.

Share and Cite

MDPI and ACS Style

Peng, J.; Wang, Z.; Yu, S.; Tang, Z. An Improvement of the Alternating Direction Method of Multipliers to Solve the Convex Optimization Problem. Mathematics 2025, 13, 811. https://doi.org/10.3390/math13050811

AMA Style

Peng J, Wang Z, Yu S, Tang Z. An Improvement of the Alternating Direction Method of Multipliers to Solve the Convex Optimization Problem. Mathematics. 2025; 13(5):811. https://doi.org/10.3390/math13050811

Chicago/Turabian Style

Peng, Jingjing, Zhijie Wang, Siting Yu, and Zengao Tang. 2025. "An Improvement of the Alternating Direction Method of Multipliers to Solve the Convex Optimization Problem" Mathematics 13, no. 5: 811. https://doi.org/10.3390/math13050811

APA Style

Peng, J., Wang, Z., Yu, S., & Tang, Z. (2025). An Improvement of the Alternating Direction Method of Multipliers to Solve the Convex Optimization Problem. Mathematics, 13(5), 811. https://doi.org/10.3390/math13050811

Note that from the first issue of 2016, this journal uses article numbers instead of page numbers. See further details here.

Article Metrics

Back to TopTop