Nothing Special   »   [go: up one dir, main page]

skip to main content
research-article

Perseus: a simple and optimal high-order method for variational inequalities: Perseus: a simple and optimal high-order method for...

Published: 13 March 2024 Publication History

Abstract

This paper settles an open and challenging question pertaining to the design of simple and optimal high-order methods for solving smooth and monotone variational inequalities (VIs). A VI involves finding xX such that F(x),x-x0 for all xX. We consider the setting in which F:RdRd is smooth with up to (p-1)th-order derivatives. For p=2, the cubic regularization of Newton’s method has been extended to VIs with a global rate of O(ϵ-1) (Nesterov in Cubic regularization of Newton’s method for convex problems with constraints, Tech. rep., Université catholique de Louvain, Center for Operations Research and Econometrics (CORE), 2006). An improved rate of O(ϵ-2/3loglog(1/ϵ)) can be obtained via an alternative second-order method, but this method requires a nontrivial line-search procedure as an inner loop. Similarly, the existing high-order methods based on line-search procedures have been shown to achieve a rate of O(ϵ-2/(p+1)loglog(1/ϵ)) (Bullins and Lai in SIAM J Optim 32(3):2208–2229, 2022; Jiang and Mokhtari in Generalized optimistic methods for convex–concave saddle point problems, 2022; Lin and Jordan in Math Oper Res 48(4):2353–2382, 2023). As emphasized by Nesterov (Lectures on convex optimization, vol 137, Springer, Berlin, 2018), however, such procedures do not necessarily imply the practical applicability in large-scale applications, and it is desirable to complement these results with a simple high-order VI method that retains the optimality of the more complex methods. We propose a pth-order method that does not require any line search procedure and provably converges to a weak solution at a rate of O(ϵ-2/(p+1)). We prove that our pth-order method is optimal in the monotone setting by establishing a lower bound of Ω(ϵ-2/(p+1)) under a generalized linear span assumption. A restarted version of our pth-order method attains a linear rate for smooth and pth-order uniformly monotone VIs and another restarted version of our pth-order method attains a local superlinear rate for smooth and strongly monotone VIs. Further, the similar pth-order method achieves a global rate of O(ϵ-2/p) for solving smooth and nonmonotone VIs satisfying the Minty condition. Two restarted versions attain a global linear rate under additional pth-order uniform Minty condition and a local superlinear rate under additional strong Minty condition.

References

[1]
Adil, D., Bullins, B., Jambulapati, A., Sachdeva, S.: Optimal methods for higher-order smooth monotone variational inequalities. ArXiv Preprint: arXiv:2205.06167 (2022)
[2]
Antipin AS Method of convex programming using a symmetric modification of Lagrange function Matekon 1978 14 2 23-38
[3]
Arjevani Y, Shamir O, and Shiff R Oracle complexity of second-order methods for smooth convex optimization Math. Program. 2019 178 1 327-360
[4]
Baes M Estimate Sequence Methods: Extensions and Approximations 2009 Zürich Institute for Operations Research, ETH
[5]
Bauschke HH and Combettes PL Convex Analysis and Monotone Operator Theory in Hilbert Spaces 2017 Berlin Springer
[6]
Birgin EG, Gardenghi JL, Martinez JM, Santos SA, and Toint PL Evaluation complexity for nonlinear constrained optimization using unscaled KKT conditions and high-order models SIAM J. Optim. 2016 26 2 951-967
[7]
Birgin EG, Gardenghi JL, Martínez JM, Santos SA, and Toint PL Worst-case evaluation complexity for unconstrained nonlinear optimization using high-order regularized models Math. Program. 2017 163 1–2 359-368
[8]
Brighi L and John R Characterizations of pseudomonotone maps and economic equilibrium J. Stat. Manag. Syst. 2002 5 1–3 253-273
[9]
Bullins, B.: Highly smooth minimization of nonsmooth problems. In: COLT, pp. 988–1030. PMLR (2020)
[10]
Bullins B and Lai KA Higher-order methods for convex-concave min-max optimization and monotone variational inequalities SIAM J. Optim. 2022 32 3 2208-2229
[11]
Carmon Y and Duchi J Gradient descent finds the cubic-regularized nonconvex Newton step SIAM J. Optim. 2019 29 3 2146-2178
[12]
Carmon Y, Duchi JC, Hinder O, and Sidford A Lower bounds for finding stationary points I Math. Program. 2020 184 1–2 71-120
[13]
Carmon, Y., Hausler, D., Jambulapati, A., Jin, Y., Sidford, A.: Optimal and adaptive Monteiro-Svaiter acceleration. In: NeurIPS, pp. 20338–20350 (2022)
[14]
Cartis C, Gould NI, and Toint PL Universal regularization methods: varying the power, the smoothness and the accuracy SIAM J. Optim. 2019 29 1 595-615
[15]
Cartis C, Gould NIM, and Toint PL On the complexity of steepest descent, Newton’s and regularized Newton’s methods for nonconvex unconstrained optimization problems SIAM J. Optim. 2010 20 6 2833-2852
[16]
Cartis C, Gould NIM, and Toint PL Adaptive cubic regularisation methods for unconstrained optimization. Part I: motivation, convergence and numerical results Math. Program. 2011 127 2 245-295
[17]
Cartis C, Gould NIM, and Toint PL Adaptive cubic regularisation methods for unconstrained optimization. Part II: worst-case function-and derivative-evaluation complexity Math. Program. 2011 130 2 295-319
[18]
Cartis C, Gould NIM, and Toint PL Evaluation Complexity of Algorithms for Nonconvex Optimization: Theory, Computation and Perspectives 2022 Philadelphia SIAM
[19]
Cesa-Bianchi N and Lugosi G Prediction, Learning, and Games 2006 Cambridge Cambridge University Press
[20]
Chen Y, Lan G, and Ouyang Y Accelerated schemes for a class of variational inequalities Math. Program. 2017 165 1 113-149
[21]
Choi SC, DeSarbo WS, and Harker PT Product positioning under price competition Manag. Sci. 1990 36 2 175-199
[22]
Cottle R, Giannessi F, and Lions JL Variational Inequalities and Complementarity Problems: Theory and Applications 1980 New York Wiley
[23]
Dang CD and Lan G On the convergence properties of non-Euclidean extragradient methods for variational inequalities with generalized monotone operators Comput. Optim. Appl. 2015 60 2 277-310
[24]
Daskalakis, C., Skoulakis, S., Zampetakis, M.: The complexity of constrained min-max optimization. In: STOC, pp. 1466–1478 (2021)
[25]
Diakonikolas, J.: Halpern iteration for near-optimal and parameter-free monotone inclusion and strong solutions to variational inequalities. In: COLT, pp. 1428–1451. PMLR (2020)
[26]
Diakonikolas, J., Daskalakis, C., Jordan, M.I.: Efficient methods for structured nonconvex-nonconcave min-max optimization. In: AISTATS, pp. 2746–2754. PMLR (2021)
[27]
Doikov N and Nesterov Y Local convergence of tensor methods Math. Program. 2022 193 1 315-336
[28]
Ewerhart C Cournot games with biconcave demand Games Econ. Behav. 2014 85 37-47
[29]
Facchinei F and Pang JS Finite-Dimensional Variational Inequalities and Complementarity Problems 2007 Berlin Springer
[30]
Fercoq O and Qu Z Adaptive restart of accelerated gradient methods under local quadratic growth condition IMA J. Numer. Anal. 2019 39 4 2069-2095
[31]
Freund RM and Lu H New computational guarantees for solving convex optimization problems with first order methods, via a function growth condition measure Math. Program. 2018 170 2 445-477
[32]
Fukushima M Equivalent differentiable optimization problems and descent methods for asymmetric variational inequality problems Math. Program. 1992 53 99-110
[33]
Gallego G and Hu M Dynamic pricing of perishable assets under competition Manag. Sci. 2014 60 5 1241-1259
[34]
Gasnikov, A., Dvurechensky, P., Gorbunov, E., Vorontsova, E., Selikhanovych, D., Uribe, C.A., Jiang, B., Wang, H., Zhang, S., Bubeck, S., Jiang, Q., Lee, Y.T., Li, Y., Sidford, A.: Near optimal methods for minimizing convex functions with Lipschitz p-th derivatives. In: COLT, pp. 1392–1393. PMLR (2019)
[35]
Ghadimi S and Lan G Optimal stochastic approximation algorithms for strongly convex stochastic composite optimization, II: shrinking procedures and optimal algorithms SIAM J. Optim. 2013 23 4 2061-2089
[36]
Giselsson, P., Boyd, S.: Monotonicity and restart in fast gradient methods. In: CDC, pp. 5058–5063. IEEE (2014)
[37]
Goodfellow, I., Pouget-Abadie, J., Mirza, M., Xu, B., Warde-Farley, D., Ozair, S., Courville, A., Bengio, Y.: Generative adversarial nets. In: NIPS, pp. 2672–2680 (2014)
[38]
Gould NIM, Lucidi S, Roma M, and Toint PL Solving the trust-region subproblem using the Lanczos method SIAM J. Optim. 1999 9 2 504-525
[39]
Gould NIM, Robinson DP, and Thorne HS On solving trust-region and other regularised subproblems in optimization Math. Program. Comput. 2010 2 1 21-57
[40]
Grapiglia GN and Nesterov Y Regularized Newton methods for minimizing functions with Hölder continuous Hessians SIAM J. Optim. 2017 27 1 478-506
[41]
Grapiglia GN and Nesterov Y Accelerated regularized Newton methods for minimizing composite convex functions SIAM J. Optim. 2019 29 1 77-99
[42]
Grapiglia GN and Nesterov Y Tensor methods for minimizing convex functions with Hölder continuous higher-order derivatives SIAM J. Optim. 2020 30 4 2750-2779
[43]
Grapiglia GN and Nesterov Y On inexact solution of auxiliary problems in tensor methods for convex optimization Optim. Methods Softw. 2021 36 1 145-170
[44]
Grapiglia GN and Nesterov Y Adaptive third-order methods for composite convex optimization SIAM J. Optim. 2023 33 3 1855-1883
[45]
Hammond JH and Magnanti TL Generalized descent methods for asymmetric systems of equations Math. Oper. Res. 1987 12 4 678-699
[46]
Harker PT and Pang JS Finite-dimensional variational inequality and nonlinear complementarity problems: a survey of theory, algorithms and applications Math. Program. 1990 48 1 161-220
[47]
Hartman P and Stampacchia G On some non-linear elliptic differential-functional equations Acta Math. 1966 115 271-310
[48]
Huang K, Zhang J, and Zhang S Cubic regularized Newton method for the saddle point models: a global and local convergence analysis J. Sci. Comput. 2022 91 2 1-31
[49]
Huang, K., Zhang, S.: An approximation-based regularized extra-gradient method for monotone variational inequalities. ArXiv Preprint: arXiv:2210.04440 (2022)
[50]
Huang, K., Zhang, S.: Beyond monotone variational inequalities: solution methods and iteration complexities. ArXiv Preprint: arXiv:2304.04153 (2023)
[51]
Iusem AN, Jofré A, Oliveira RI, and Thompson P Extragradient method with variance reduction for stochastic variational inequalities SIAM J. Optim. 2017 27 2 686-724
[52]
Jiang B, Lin T, and Zhang S A unified adaptive tensor approximation scheme to accelerate composite convex optimization SIAM J. Optim. 2020 30 4 2897-2926
[53]
Jiang, R., Mokhtari, A.: Generalized optimistic methods for convex–concave saddle point problems. ArXiv Preprint: arXiv:2202.09674 (2022)
[54]
Kannan A and Shanbhag UV Optimal stochastic extragradient schemes for pseudomonotone stochastic variational inequality problems and their variants Comput. Optim. Appl. 2019 74 3 779-820
[55]
Kinderlehrer D and Stampacchia G An Introduction to Variational Inequalities and Their Applications 2000 Philadelphia SIAM
[56]
Kleinberg, B., Li, Y., Yuan, Y.: An alternative view: when does SGD escape local minima? In: ICML, pp. 2698–2707. PMLR (2018)
[57]
Kornowski, G., Shamir, O.: High-order oracle complexity of smooth and strongly convex optimization. ArXiv Preprint: arXiv:2010.06642 (2020)
[58]
Korpelevich GM The extragradient method for finding saddle points and other problems Matecon 1976 12 747-756
[59]
Kotsalis G, Lan G, and Li T Simple and optimal methods for stochastic variational inequalities, I: operator extrapolation SIAM J. Optim. 2022 32 3 2041-2073
[60]
Kovalev, D., Gasnikov, A.: The first optimal acceleration of high-order methods in smooth convex optimization. In: NeurIPS, pp. 35339–35351 (2022)
[61]
Lan G and Zhou Y An optimal randomized incremental gradient method Math. Program. 2018 171 1 167-215
[62]
Lan G and Zhou Y Random gradient extrapolation for distributed and stochastic optimization SIAM J. Optim. 2018 28 4 2753-2782
[63]
Lemke CE and Howson JT Equilibrium points of bimatrix games J. Soc. Ind. Appl. Math. 1964 12 2 413-423
[64]
Li, Y., Yuan, Y.: Convergence analysis of two-layer neural networks with ReLU activation. In: NIPS, pp. 597–607 (2017)
[65]
Lin T and Jordan MI A control-theoretic perspective on optimal high-order optimization Math. Program. 2022 195 1 929-975
[66]
Lin T and Jordan MI Monotone inclusions, acceleration, and closed-loop control Math. Oper. Res. 2023 48 4 2353-2382
[67]
Lin, T., Mertikopoulos, P., Jordan, M.I.: Explicit second-order min-max optimization methods with optimal convergence guarantee. ArXiv Preprint: arXiv:2210.12860 (2022)
[68]
Liu M, Rafique H, Lin Q, and Yang T First-order convergence theory for weakly-convex–weakly-concave min–max problems J. Mach. Learn. Res. 2021 22 169 1-34
[69]
Madry, A., Makelov, A., Schmidt, L., Tsipras, D., Vladu, A.: Towards deep learning models resistant to adversarial attacks. In: ICLR (2018). https://openreview.net/forum?id=rJzIBfZAb
[70]
Magnanti TL and Perakis G A unifying geometric solution framework and complexity analysis for variational inequalities Math. Program. 1995 71 3 327-351
[71]
Magnanti TL and Perakis G Averaging schemes for variational inequalities and systems of equations Math. Oper. Res. 1997 22 3 568-587
[72]
Magnanti TL and Perakis G The orthogonality theorem and the strong-f-monotonicity condition for variational inequality algorithms SIAM J. Optim. 1997 7 1 248-273
[73]
Magnanti TL and Perakis G Solving variational inequality and fixed point problems by line searches and potential optimization Math. Program. 2004 101 3 435-461
[74]
Marques Alves M Variants of the A-HPE and large-step A-HPE algorithms for strongly convex problems with applications to accelerated high-order tensor methods Optim. Methods Softw. 2022 37 6 2021-2051
[75]
Martínez J On high-order model regularization for constrained optimization SIAM J. Optim. 2017 27 4 2447-2458
[76]
Mertikopoulos P and Zhou Z Learning in games with continuous action sets and unknown payoff functions Math. Program. 2019 173 1 465-507
[77]
Minty GJ Monotone (nonlinear) operators in Hilbert space Duke Math. J. 1962 29 3 341-346
[78]
Mokhtari A, Ozdaglar AE, and Pattathil S Convergence rate of o(1/k) for optimistic gradient and extragradient methods in smooth convex-concave saddle point problems SIAM J. Optim. 2020 30 4 3230-3251
[79]
Monteiro RDC and Svaiter BF On the complexity of the hybrid proximal extragradient method for the iterates and the ergodic mean SIAM J. Optim. 2010 20 6 2755-2787
[80]
Monteiro RDC and Svaiter BF Complexity of variants of Tseng’s modified FB splitting and Korpelevich’s methods for hemivariational inequalities with applications to saddle-point and convex optimization problems SIAM J. Optim. 2011 21 4 1688-1720
[81]
Monteiro RDC and Svaiter BF Iteration-complexity of a Newton proximal extragradient method for monotone variational inequalities and inclusion problems SIAM J. Optim. 2012 22 3 914-935
[82]
Monteiro RDC and Svaiter BF An accelerated hybrid proximal extragradient method for convex optimization and its implications to second-order methods SIAM J. Optim. 2013 23 2 1092-1125
[83]
Necoara I, Nesterov Y, and Glineur F Linear convergence of first order methods for non-strongly convex optimization Math. Program. 2019 175 1 69-107
[84]
Nemirovski A Prox-method with rate of convergence o(1/t) for variational inequalities with Lipschitz continuous monotone operators and smooth convex-concave saddle point problems SIAM J. Optim. 2004 15 1 229-251
[85]
Nemirovski AS and Nesterov YE Optimal methods of smooth convex minimization USSR Comput. Math. Math. Phys. 1985 25 2 21-30
[86]
Nesterov, Y.: Cubic regularization of Newton’s method for convex problems with constraints. Tech. rep., Université catholique de Louvain, Center for Operations Research and Econometrics (CORE) (2006)
[87]
Nesterov Y Dual extrapolation and its applications to solving variational inequalities and related problems Math. Program. 2007 109 2 319-344
[88]
Nesterov Y Accelerating the cubic regularization of Newton’s method on convex problems Math. Program. 2008 112 1 159-181
[89]
Nesterov Y Gradient methods for minimizing composite functions Math. Program. 2013 140 1 125-161
[90]
Nesterov Y Lectures on Convex Optimization 2018 Berlin Springer 137
[91]
Nesterov Y Implementable tensor methods in unconstrained convex optimization Math. Program. 2021 186 1 157-183
[92]
Nesterov, Y.: Inexact accelerated high-order proximal-point methods. Mathematical Programming, pp. 1–26 (2021)
[93]
Nesterov Y Inexact high-order proximal-point methods with auxiliary search procedure SIAM J. Optim. 2021 31 4 2807-2828
[94]
Nesterov Y Superfast second-order methods for unconstrained convex optimization J. Optim. Theory Appl. 2021 191 1 1-30
[95]
Nesterov Y and Polyak BT Cubic regularization of Newton method and its global performance Math. Program. 2006 108 1 177-205
[96]
Nesterov, Y.E.: A method of solving a convex programming problem with convergence rate o(k2). In: Doklady Akademii Nauk, vol. 269, pp. 543–547. Russian Academy of Sciences (1983)
[97]
Ostroukhov, P., Kamalov, R., Dvurechensky, P., Gasnikov, A.: Tensor methods for strongly convex strongly concave saddle point problems and strongly monotone variational inequalities. ArXiv Preprint: arXiv:2012.15595 (2020)
[98]
Ouyang Y and Xu Y Lower complexity bounds of first-order methods for convex–concave bilinear saddle-point problems Math. Program. 2021 185 1 1-35
[99]
O’donoghue B and Candes E Adaptive restart for accelerated gradient schemes Found. Comput. Math. 2015 15 3 715-732
[100]
Popov LD A modification of the Arrow–Hurwicz method for search of saddle points Math. Notes Acad. Sci. USSR 1980 28 5 845-848
[101]
Ralph, D., Wright, S.J.: Superlinear convergence of an interior-point method for monotone variational inequalities. Complementarity and Variational Problems: State of the Art pp. 345–385 (1997)
[102]
Renegar J and Grimmer B A simple nearly optimal restart scheme for speeding up first-order methods Found. Comput. Math. 2022 22 1 211-256
[103]
Rockafellar RT and Wets RJB Variational Analysis 2009 Berlin Springer 317
[104]
Roulet, V., d’Aspremont, A.: Sharpness, restart and acceleration. In: NIPS, pp. 1119–1129 (2017)
[105]
Scarf H The approximation of fixed points of a continuous mapping SIAM J. Appl. Math. 1967 15 5 1328-1343
[106]
Sinha, A., Namkoong, H., Duchi, J.: Certifiable distributional robustness with principled adversarial training. In: ICLR (2018). https://openreview.net/forum?id=Hk6kPgZA-
[107]
Solodov MV and Svaiter BF A new projection method for variational inequality problems SIAM J. Control. Optim. 1999 37 3 765-776
[108]
Song C, Jiang Y, and Ma Y Unified acceleration of high-order algorithms under general Hölder continuity SIAM J. Optim. 2021 31 3 1797-1826
[109]
Song, C., Zhou, Z., Zhou, Y., Jiang, Y., Ma, Y.: Optimistic dual extrapolation for coherent non-monotone variational inequalities. In: NeurIPS, pp. 14303–14314 (2020)
[110]
Titov, A.A., Ablaev, S.S., Alkousa, M.S., Stonyakin, F.S., Gasnikov, A.V.: Some adaptive first-order methods for variational inequalities with relatively strongly monotone operators and generalized smoothness. In: ICOPTA, pp. 135–150. Springer (2022)
[111]
Todd MJ The Computation of Fixed Points and Applications 2013 Berlin Springer
[112]
Trémolières R, Lions JL, and Glowinski R Numerical Analysis of Variational Inequalities 2011 Amsterdam Elsevier
[113]
Tseng P A modified forward-backward splitting method for maximal monotone mappings SIAM J. Control. Optim. 2000 38 2 431-446
[114]
Wibisono A, Wilson AC, and Jordan MI A variational perspective on accelerated methods in optimization Proc. Natl. Acad. Sci. 2016 113 47 E7351-E7358
[115]
Zhang J, Hong M, and Zhang S On lower iteration complexity bounds for the convex–concave saddle point problems Math. Program. 2022 194 1 901-935

Index Terms

  1. Perseus: a simple and optimal high-order method for variational inequalities: Perseus: a simple and optimal high-order method for...
        Index terms have been assigned to the content through auto-classification.

        Recommendations

        Comments

        Please enable JavaScript to view thecomments powered by Disqus.

        Information & Contributors

        Information

        Published In

        cover image Mathematical Programming: Series A and B
        Mathematical Programming: Series A and B  Volume 209, Issue 1
        Jan 2025
        925 pages

        Publisher

        Springer-Verlag

        Berlin, Heidelberg

        Publication History

        Published: 13 March 2024
        Accepted: 14 February 2024
        Received: 07 May 2022

        Author Tags

        1. Variational inequality
        2. Simple and optimal high-order method
        3. Acceleration
        4. Iteration complexity
        5. Local superlinear convergence

        Author Tags

        1. 49M15
        2. 49J40
        3. 65K15
        4. 68Q25
        5. 90C33
        6. 90C60

        Qualifiers

        • Research-article

        Funding Sources

        Contributors

        Other Metrics

        Bibliometrics & Citations

        Bibliometrics

        Article Metrics

        • 0
          Total Citations
        • 0
          Total Downloads
        • Downloads (Last 12 months)0
        • Downloads (Last 6 weeks)0
        Reflects downloads up to 05 Mar 2025

        Other Metrics

        Citations

        View Options

        View options

        Figures

        Tables

        Media

        Share

        Share

        Share this Publication link

        Share on social media