Skip to main content
Log in

An Adaptive Analog of Nesterov’s Method for Variational Inequalities with a Strongly Monotone Operator

  • Published:
Numerical Analysis and Applications Aims and scope Submit manuscript

Abstract

An adaptive analog of Nesterov’s method for variational inequalities with a strongly monotone operator is proposed. The main idea of the method is an adaptive choice of constants in the maximized concave functionals at each iteration. In this case there is no need in specifying exact values of the constants, since this method makes it possible to find suitable constants at each iteration. Some estimates for the parameters determining the quality of the solution to the variational inequality are obtained as functions of the number of iterations.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Similar content being viewed by others

References

  1. Facchinei, F. and Pang, J.S., Finite-Dimensional Variational Inequality and Complementarity Problems, vols. 1, 2, New York: Springer-Verlag, 2003.

  2. Antipin, A.S., Jacimovic, V., and Jacimovic, M., Dynamics and Variational Inequalities, Zh. Vych. Mat. Mat. Fiz., 2017, vol. 57, no. 5, pp. 784–801.

    MathSciNet  MATH  Google Scholar 

  3. Korpelevich, G.M., The Extragradient Method for Finding Saddle Points and Other Problems, Ekon. Mat. Met., 1976, vol. 12, no. 4, pp. 747–756.

    MathSciNet  MATH  Google Scholar 

  4. Nemirovski, A., Prox-Method with Rate of Convergence O(1/T) for Variational Inequalities with Lipschitz Continuous Monotone Operators and Smooth Convex-Concave Saddle Point Problems, SIAM J. Optim., 2004, vol. 15, pp. 229–251.

    Article  MathSciNet  MATH  Google Scholar 

  5. Nesterov, Yu., UniversalGradientMethods for Convex Optimization Problems, Math. Program. Ser. A and B Archive, 2015, vol. 152, iss. 1/2. pp. 381–404.

    Article  MATH  Google Scholar 

  6. Gasnikov, A.V., Sovremennye chislennye metody optimizatsii. Metod universal’nogo gradientnogo spuska (Modern Numerical Methods of Optimization: Universal Gradient Descent Method), Moscow: Moscow Institute of Physics and Technology, 2018.

    Google Scholar 

  7. Nesterov, Yu., Dual Extrapolation and Its Application for Solving Variational Inequalities and Related Problems, Math. Program. Ser. B, 2007, vol. 109, iss. 2/3, pp. 319–344.

    Article  MathSciNet  MATH  Google Scholar 

  8. Nesterov, Yu.E., Vvedenie v vypukluyu optimizatsiyu (Introduction to Convex Optimization), Moscow: Moscow Center for ContinuousMathematical Education, 2010.

    Google Scholar 

  9. Nesterov, Yu. and Scrimali, L., Solving Strongly Monotone Variational and Quasi-Variational Inequalities, Discr. Cont. Dyn. Sys., 2007, 31.10.2139/ssrn.970903.

    Google Scholar 

  10. Nesterov, Yu.E., Algorithmic Convex Optimization, Doct. Sci. (Phys.-Math.) Dissertation, Moscow, 2013.

    Google Scholar 

Download references

Acknowledgements

The author would like to thank A.V.Gasnikov, Yu.E.Nesterov, and the anonymous reviewers for their useful discussions and comments.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to F. S. Stonyakin.

Additional information

Russian Text © The Author(s), 2019, published in Sibirskii Zhurnal Vychislitel’noi Matematiki, 2019, Vol. 22, No. 2, pp. 201–221.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Stonyakin, F.S. An Adaptive Analog of Nesterov’s Method for Variational Inequalities with a Strongly Monotone Operator. Numer. Analys. Appl. 12, 166–175 (2019). https://doi.org/10.1134/S199542391902006X

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1134/S199542391902006X

Keywords

Navigation