Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Accelerated Smoothing Hard Thresholding Algorithms for 0 Regularized Nonsmooth Convex Regression Problem

Published: 15 June 2023 Publication History

Abstract

We study a class of constrained sparse optimization problems with cardinality penalty, where the feasible set is defined by box constraint, and the loss function is convex but not necessarily smooth. First, we propose an accelerated smoothing hard thresholding (ASHT) algorithm for solving such problems, which combines smoothing approximation, extrapolation technique and iterative hard thresholding method. The extrapolation coefficients can be chosen to satisfy supkβk=1. We discuss the convergence of ASHT algorithm with different extrapolation coefficients, and give a sufficient condition to ensure that any accumulation point of the iterates is a local minimizer of the original problem. For a class of special updating schemes on the extrapolation coefficients, we obtain that the iterates are convergent to a local minimizer of the problem, and the convergence rate is o(lnσk/k) with σ(1/2,1] on the loss and objective function values. Second, we consider the case in which the loss function is Lipschitz continuously differentiable, and develop an accelerated hard thresholding (AHT) algorithm to solve it. We prove that the iterates of AHT algorithm converge to a local minimizer of the problem that satisfies a desirable lower bound property. Moreover, we show that the convergence rates of loss and objective function values are o(k-2). Finally, some numerical examples are presented to show the theoretical results.

References

[1]
Adly S and Attouch H Finite convergence of proximal-gradient inertial algorithms combining dry friction with Hessian-driven damping SIAM J. Optim. 2020 30 3 2134-2162
[2]
Alecsa CD, László SC, and Pinţa T An extension of the second order dynamical system that model Nesterov’s convex gradient method Appl. Math. Optim. 2021 84 2 1687-1716
[3]
Attouch, H., László, S.: Newton-like inertial dynamics and proximal algorithms governed by maximally monotone operators. SIAM J. Optim. 30(4), 3252–3283 (2020)
[4]
Attouch H and Peypouquet J The rate of convergence of Nesterov’s accelerated forward-backward method is actually faster than 1/k2 SIAM J. Optim. 2016 26 3 1824-1834
[5]
Beck A and Teboulle M A fast iterative shrinkage-thresholding algorithm for linear inverse problems SIAM J. Imaging Sci. 2009 2 1 183-202
[6]
Bertsekas DP Nonlinear Programming 1999 2 Belmont Athena Scientific
[7]
Bian W Smoothing accelerated algorithm for constrained nonsmooth convex optimization problems (in chinese) Sci. Sin. Math. 2020 50 1651-1666
[8]
Bian W and Chen X Optimality and complexity for constrained optimization problems with nonconvex regularization Math. Oper. Res. 2017 42 4 1063-1084
[9]
Bian W and Chen X A smoothing proximal gradient algorithm for nonsmooth convex regression with cardinality penalty SIAM J. Numer. Anal. 2020 58 1 858-883
[10]
Bian W, Chen X, and Ye YY Complexity analysis of interior point algorithms for non-Lipschitz and nonconvex minimization Math. Program. 2015 149 1–2 301-327
[11]
Blumensath T and Davies M Sparse and shift-invariant representations of music IEEE Trans. Audio Speech Lang. Process. 2006 14 1 50-57
[12]
Blumensath T and Davies M Iterative thresholding for sparse approximations J. Fourier Anal. Appl. 2008 14 5–6 629-654
[13]
Blumensath T and Davies M Iterative hard thresholding for compressed sensing Appl. Comput. Harmon. Anal. 2009 27 3 265-274
[14]
Boţ, R.I., Böhm, A.: Variable smoothing for convex optimization problems using stochastic gradients. J. Sci. Comput. (2020)
[15]
Bruckstein A, Donoho D, and Elad M From sparse solutions of systems of equations to sparse modeling of signals and images SIAM Rev. 2009 51 1 34-81
[16]
Candès E, Romberg J, and Tao T Robust uncertainty principles: exact signal reconstruction from highly incomplete frequency information IEEE Trans. Inf. Theory 2006 52 2 489-509
[17]
Chambolle A, DeVore R, Lee N, and Lucier B Nonlinear wavelet image processing: variational problems, compression, and noise removal through wavelet shrinkage IEEE Trans. Image Process. 1998 7 3 319-335
[18]
Chen X Smoothing methods for nonsmooth, nonconvex minimization Math. Program. 2012 134 1 71-99
[19]
Combettes P and Wajs V Signal recovery by proximal forward-backward splitting Multiscale Model. Simul. 2005 4 4 1168-1200
[20]
Dai W and Milenkovic O Subspace pursuit for compressive sensing signal reconstruction IEEE Trans. Inf. Theory 2009 55 5 2230-2249
[21]
Daubechies I, Defrise M, and De Mol C An iterative thresholding algorithm for linear inverse problems with a sparsity constraint Commun. Pure Appl. Math. 2004 57 11 1413-1457
[22]
Doikov N and Nesterov Y Contracting proximal methods for smooth convex optimization SIAM J. Optim. 2020 30 4 3146-3169
[23]
Donoho D Compressed sensing IEEE Trans. Inf. Theory 2006 52 4 1289-1306
[24]
Fan J and Li R Variable selection via nonconcave penalized likelihood and its oracle properties J. Am. Stat. Assoc. 2001 96 456 1348-1360
[25]
Hale E, Yin W, and Zhang Y Fixed-point continuation for 1-minimization: methodology and convergence SIAM J. Optim. 2008 19 3 1107-1130
[26]
Hoda S, Gilpin A, Pena J, and Sandholm T Smoothing techniques for computing Nash equilibria of sequential games Math. Oper. Res. 2010 35 2 494-512
[27]
Liu Y and Wu Y Variable selection via a combination of the 0 and 1 penalties J. Comput. Graph. Stat. 2007 16 4 782-798
[28]
Lu Z Iterative hard thresholding methods for 0 regularized convex cone programming Math. Program. 2014 147 1–2 125-154
[29]
Lu Z and Zhang Y Sparse approximation via penalty decomposition methods SIAM J. Optim. 2013 23 4 2448-2478
[30]
Mallat S and Zhang Z Matching pursuits with time-frequency dictionaries IEEE Trans. Signal Process. 1993 41 12 3397-3415
[31]
Nesterov Y Smooth minimization of non-smooth functions Math. Program. 2005 103 1 127-152
[32]
Nesterov Y Gradient methods for minimizing composite functions Math. Program. 2013 140 1 125-161
[33]
Nikolova M Local strong homogeneity of a regularized estimator SIAM J. Appl. Math. 2000 61 2 633-658
[34]
Pati, Y., Rezaiifar, R., Krishnaprasad, P.: Orthogonal matching pursuit-recursive function approximation with applications to wavelet decomposition. In: Conference Record of the Twenty-Seventh Asilomar Conference on Signal, Systems and Computers, vol. 1–2, pp. 40–44 (1993)
[35]
Peleg D and Meir R A bilinear formulation for vector sparsity optimization Signal Process. 2008 88 2 375-389
[36]
Soubies E, Blanc-Feraud L, and Aubert G A continuous exact 0 penalty (CEL0) for least squares regularized problem SIAM J. Imaging Sci. 2015 8 3 1607-1639
[37]
Su W, Boyd S, and Candès E A differential equation for modeling Nesterov’s accelerated gradient method: theory and insights J. Mach. Learn. Res. 2016 17 153 1-43
[38]
Tan CH, Qian YQ, Ma SQ, and Zhang T Accelerated dual-averaging primal-dual method for composite convex minimization Optim. Method Softw. 2020 35 4 741-766
[39]
Tibshirani R Regression shrinkage and selection via the Lasso J. R. Stat. Soc. Ser. B-Methodol. 1996 58 1 267-288
[40]
Wen B, Chen X, and Pong T Linear convergence of proximal gradient algorithm with extrapolation for a class of nonconvex nonsmooth minimization problems SIAM J. Optim. 2017 27 1 124-145
[41]
Wen B and Xue XP On the convergence of the iterates of proximal gradient algorithm with extrapolation for convex nonsmooth minimization problems J. Glob. Optim. 2019 75 3 767-787
[42]
Wu F and Bian W Accelerated iterative hard thresholding algorithm for 0 regularized regression problem J. Glob. Optim. 2020 76 4 819-840
[43]
Wu, F., Bian, W.: Accelerated forward-backward method with fast convergence rate for nonsmooth convex optimization beyond differentiability. arXiv:2110.01454v1 (2021)
[44]
Yu Q and Zhang XZ A smoothing proximal gradient algorithm for matrix rank minimization problem Comput. Optim. Appl. 2022 81 2 519-538
[45]
Zhang C and Chen X A smoothing active set method for linearly constrained non-Lipschitz nonconvex optimization SIAM J. Optim. 2020 30 1 1-30
[46]
Zheng Z, Fan Y, and Lv J High dimensional thresholded regression and shrinkage effect J. R. Stat. Soc. Ser. B-Stat. Methodol. 2014 76 3 627-649

Recommendations

Comments

Please enable JavaScript to view thecomments powered by Disqus.

Information & Contributors

Information

Published In

cover image Journal of Scientific Computing
Journal of Scientific Computing  Volume 96, Issue 2
Aug 2023
816 pages

Publisher

Plenum Press

United States

Publication History

Published: 15 June 2023
Accepted: 13 May 2023
Revision received: 11 May 2023
Received: 30 May 2022

Author Tags

  1. Nonsmooth optimization
  2. Smoothing method
  3. Cardinality penalty
  4. Accelerated algorithm with extrapolation
  5. Convergence rate

Author Tags

  1. 90C30
  2. 65K05
  3. 49J52
  4. 49M37

Qualifiers

  • Research-article

Funding Sources

Contributors

Other Metrics

Bibliometrics & Citations

Bibliometrics

Article Metrics

  • 0
    Total Citations
  • 0
    Total Downloads
  • Downloads (Last 12 months)0
  • Downloads (Last 6 weeks)0
Reflects downloads up to 07 Mar 2025

Other Metrics

Citations

View Options

View options

Figures

Tables

Media

Share

Share

Share this Publication link

Share on social media