IDEAS home Printed from https://ideas.repec.org/a/spr/coopap/v81y2022i3d10.1007_s10589-022-00349-z.html
   My bibliography  Save this article

On the acceleration of the Barzilai–Borwein method

Author

Listed:
  • Yakui Huang

    (Hebei University of Technology)

  • Yu-Hong Dai

    (Chinese Academy of Sciences)

  • Xin-Wei Liu

    (Hebei University of Technology)

  • Hongchao Zhang

    (Louisiana State University)

Abstract

The Barzilai–Borwein (BB) gradient method is efficient for solving large-scale unconstrained problems to modest accuracy due to its ingenious stepsize which generally yields nonmonotone behavior. In this paper, we propose a new stepsize to accelerate the BB method by requiring finite termination for minimizing the two-dimensional strongly convex quadratic function. Based on this new stepsize, we develop an efficient gradient method for quadratic optimization which adaptively takes the nonmonotone BB stepsizes and certain monotone stepsizes. Two variants using retard stepsizes associated with the new stepsize are also presented. Numerical experiments show that our strategies of properly inserting monotone gradient steps into the nonmonotone BB method could significantly improve its performance and our new methods are competitive with the most successful gradient descent methods developed in the recent literature.

Suggested Citation

  • Yakui Huang & Yu-Hong Dai & Xin-Wei Liu & Hongchao Zhang, 2022. "On the acceleration of the Barzilai–Borwein method," Computational Optimization and Applications, Springer, vol. 81(3), pages 717-740, April.
  • Handle: RePEc:spr:coopap:v:81:y:2022:i:3:d:10.1007_s10589-022-00349-z
    DOI: 10.1007/s10589-022-00349-z
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s10589-022-00349-z
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s10589-022-00349-z?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Yu-Hong Dai & Yakui Huang & Xin-Wei Liu, 2019. "A family of spectral gradient methods for optimization," Computational Optimization and Applications, Springer, vol. 74(1), pages 43-65, September.
    2. Roberta De Asmundis & Daniela di Serafino & William Hager & Gerardo Toraldo & Hongchao Zhang, 2014. "An efficient gradient method using the Yuan steplength," Computational Optimization and Applications, Springer, vol. 59(3), pages 541-563, December.
    3. di Serafino, Daniela & Ruggiero, Valeria & Toraldo, Gerardo & Zanni, Luca, 2018. "On the steplength selection in gradient methods for unconstrained optimization," Applied Mathematics and Computation, Elsevier, vol. 318(C), pages 176-195.
    4. Clóvis Gonzaga & Ruana Schneider, 2016. "On the steepest descent algorithm for quadratic functions," Computational Optimization and Applications, Springer, vol. 63(2), pages 523-542, March.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Giulia Ferrandi & Michiel E. Hochstenbach & Nataša Krejić, 2023. "A harmonic framework for stepsize selection in gradient methods," Computational Optimization and Applications, Springer, vol. 85(1), pages 75-106, May.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Roberto Andreani & Marcos Raydan, 2021. "Properties of the delayed weighted gradient method," Computational Optimization and Applications, Springer, vol. 78(1), pages 167-180, January.
    2. Masoud Fatemi, 2022. "On initial point selection of the steepest descent algorithm for general quadratic functions," Computational Optimization and Applications, Springer, vol. 82(2), pages 329-360, June.
    3. Crisci, Serena & Ruggiero, Valeria & Zanni, Luca, 2019. "Steplength selection in gradient projection methods for box-constrained quadratic programs," Applied Mathematics and Computation, Elsevier, vol. 356(C), pages 312-327.
    4. E. Loli Piccolomini & V. L. Coli & E. Morotti & L. Zanni, 2018. "Reconstruction of 3D X-ray CT images from reduced sampling by a scaled gradient projection algorithm," Computational Optimization and Applications, Springer, vol. 71(1), pages 171-191, September.
    5. Yu-Hong Dai & Yakui Huang & Xin-Wei Liu, 2019. "A family of spectral gradient methods for optimization," Computational Optimization and Applications, Springer, vol. 74(1), pages 43-65, September.
    6. Bonettini, Silvia & Prato, Marco & Rebegoldi, Simone, 2016. "A cyclic block coordinate descent method with generalized gradient projections," Applied Mathematics and Computation, Elsevier, vol. 286(C), pages 288-300.
    7. Serena Crisci & Federica Porta & Valeria Ruggiero & Luca Zanni, 2023. "Hybrid limited memory gradient projection methods for box-constrained optimization problems," Computational Optimization and Applications, Springer, vol. 84(1), pages 151-189, January.
    8. Giulia Ferrandi & Michiel E. Hochstenbach & Nataša Krejić, 2023. "A harmonic framework for stepsize selection in gradient methods," Computational Optimization and Applications, Springer, vol. 85(1), pages 75-106, May.
    9. Corsaro, Stefania & De Simone, Valentina & Marino, Zelda, 2021. "Split Bregman iteration for multi-period mean variance portfolio optimization," Applied Mathematics and Computation, Elsevier, vol. 392(C).
    10. di Serafino, Daniela & Ruggiero, Valeria & Toraldo, Gerardo & Zanni, Luca, 2018. "On the steplength selection in gradient methods for unconstrained optimization," Applied Mathematics and Computation, Elsevier, vol. 318(C), pages 176-195.
    11. Na Huang, 2022. "On R-linear convergence analysis for a class of gradient methods," Computational Optimization and Applications, Springer, vol. 81(1), pages 161-177, January.
    12. Harry Fernando Oviedo Leon, 2019. "A delayed weighted gradient method for strictly convex quadratic minimization," Computational Optimization and Applications, Springer, vol. 74(3), pages 729-746, December.
    13. Behzad Azmi & Karl Kunisch, 2020. "Analysis of the Barzilai-Borwein Step-Sizes for Problems in Hilbert Spaces," Journal of Optimization Theory and Applications, Springer, vol. 185(3), pages 819-844, June.
    14. Yutao Zheng & Bing Zheng, 2017. "A New Modified Barzilai–Borwein Gradient Method for the Quadratic Minimization Problem," Journal of Optimization Theory and Applications, Springer, vol. 172(1), pages 179-186, January.
    15. Stefania Corsaro & Valentina Simone, 2019. "Adaptive $$l_1$$ l 1 -regularization for short-selling control in portfolio selection," Computational Optimization and Applications, Springer, vol. 72(2), pages 457-478, March.
    16. Mina Torabi & Mohammad-Mehdi Hosseini, 2018. "A New Descent Algorithm Using the Three-Step Discretization Method for Solving Unconstrained Optimization Problems," Mathematics, MDPI, vol. 6(4), pages 1-18, April.
    17. Marco Viola & Mara Sangiovanni & Gerardo Toraldo & Mario R. Guarracino, 2019. "Semi-supervised generalized eigenvalues classification," Annals of Operations Research, Springer, vol. 276(1), pages 249-266, May.
    18. Clóvis Gonzaga & Ruana Schneider, 2016. "On the steepest descent algorithm for quadratic functions," Computational Optimization and Applications, Springer, vol. 63(2), pages 523-542, March.
    19. Stefania Corsaro & Valentina De Simone & Zelda Marino, 2021. "Fused Lasso approach in portfolio selection," Annals of Operations Research, Springer, vol. 299(1), pages 47-59, April.
    20. di Serafino, Daniela & Toraldo, Gerardo & Viola, Marco, 2021. "Using gradient directions to get global convergence of Newton-type methods," Applied Mathematics and Computation, Elsevier, vol. 409(C).

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:coopap:v:81:y:2022:i:3:d:10.1007_s10589-022-00349-z. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.