Skip to main content
Log in

Accelerated Gradient Sliding for Minimizing a Sum of Functions

  • COMPUTER SCIENCE
  • Published:
Doklady Mathematics Aims and scope Submit manuscript

Abstract

We propose a new way of justifying the accelerated gradient sliding of G. Lan, which allows one to extend the sliding technique to a combination of an accelerated gradient method with an accelerated variance reduction method. New optimal estimates for the solution of the problem of minimizing a sum of smooth strongly convex functions with a smooth regularizer are obtained.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Similar content being viewed by others

REFERENCES

  1. A. V. Gasnikov, Modern Numerical Optimization Methods: Universal Gradient Descent Method (Mosk. Fiz.-Tekh. Inst., Moscow, 2018) [in Russian].

    Google Scholar 

  2. Z. Allen-Zhu and E. Hazan, Adv. Neural Inf. Process. Syst. 29, 1614–1622 (2016).

    Google Scholar 

  3. A. Beznosikov, E. Gorbunov, and A. Gasnikov, “Derivative-free method for decentralized distributed non-smooth optimization,” The 21st World Congress of the International Federation of Automatic Control (IFAC, 2020).

  4. L. Bottou, F. E. Curtis, and J. Nocedal, SIAM Rev. 60 (2), 223–311 (2018).

    Article  MathSciNet  Google Scholar 

  5. P. Dvurechensky, A. Gasnikov, and A. Tiurin, Randomized similar triangles method: A unifying framework for accelerated randomized optimization methods (coordinate descent, directional search, derivative-free method). arXiv:1707.08486

  6. E. Hazan, Lecture Notes: Optimization for Machine Learning. arXiv:1909.03550

  7. A. Ivanova, A. Gasnikov, P. Dvurechensky, D. Dvinskikh, A. Tyurin, E. Vorontsova, and D. Pasechnyuk, Oracle Complexity Separation in Convex Optimization. arXiv:2002.02706

  8. A. Ivanova, D. Grishchenko, A. Gasnikov, and E. Shulgin, Adaptive Catalyst for Smooth Convex Optimization. arXiv:1911.11271

  9. G. Lan, Lectures on Optimization: Methods for Machine Learning. https://pwp.gatech.edu/guanghui-lan/publications/

  10. G. Lan, Z. Li, and Y. Zhou, Adv. Neural Inf. Process. Syst. 32, 10462–10472 (2019).

    Google Scholar 

  11. H. Lin, J. Mairal, and Z. Harchaoui, J. Machine Learning Res. 18 (1), 7854–7907 (2017).

    Google Scholar 

  12. Yu. Nesterov, Math. Program. 140 (1), 125–161 (2013).

    Article  MathSciNet  Google Scholar 

  13. Yu. Nesterov, Math. Program. 103 (1), 127–152 (2005).

    Article  MathSciNet  Google Scholar 

  14. S. Shalev-Shwartz and S. Ben-David, Understanding Machine Learning: From Theory to Algorithms (Cambridge Univ. Press, Cambridge, 2014).

    Book  Google Scholar 

  15. S. Shalev-Shwartz, O. Shamir, N. Srebro, and K. Sridharan, “Stochastic convex optimization,” COLT (2009).

    Google Scholar 

Download references

Funding

This work was supported by the Russian Foundation for Basic Research, project no. 18-31-20005 mol_a_ved (Section 1) and project no. 19-31-90062 Graduate students (Section 2).

Dvinskikh acknowledges the support of the Ministry of Science and Higher Education of the Russian Federation, state assignment no. 075-00337-20-03.

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to S. S. Omelchenko or A. V. Gasnikov.

Additional information

Translated by I. Ruzanova

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Dvinskikh, D.M., Omelchenko, S.S., Gasnikov, A.V. et al. Accelerated Gradient Sliding for Minimizing a Sum of Functions. Dokl. Math. 101, 244–246 (2020). https://doi.org/10.1134/S1064562420030084

Download citation

  • Received:

  • Revised:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1134/S1064562420030084

Keywords:

Navigation