IDEAS home Printed from https://ideas.repec.org/a/spr/joptap/v188y2021i2d10.1007_s10957-020-01787-7.html
   My bibliography  Save this article

On the Linear Convergence of Forward–Backward Splitting Method: Part I—Convergence Analysis

Author

Listed:
  • Yunier Bello-Cruz

    (Northern Illinois University)

  • Guoyin Li

    (University of New South Wales)

  • Tran T. A. Nghia

    (Oakland University)

Abstract

In this paper, we study the complexity of the forward–backward splitting method with Beck–Teboulle’s line search for solving convex optimization problems, where the objective function can be split into the sum of a differentiable function and a nonsmooth function. We show that the method converges weakly to an optimal solution in Hilbert spaces, under mild standing assumptions without the global Lipschitz continuity of the gradient of the differentiable function involved. Our standing assumptions is weaker than the corresponding conditions in the paper of Salzo (SIAM J Optim 27:2153–2181, 2017). The conventional complexity of sublinear convergence for the functional value is also obtained under the local Lipschitz continuity of the gradient of the differentiable function. Our main results are about the linear convergence of this method (in the quotient type), in terms of both the function value sequence and the iterative sequence, under only the quadratic growth condition. Our proof technique is direct from the quadratic growth conditions and some properties of the forward–backward splitting method without using error bounds or Kurdya-Łojasiewicz inequality as in other publications in this direction.

Suggested Citation

  • Yunier Bello-Cruz & Guoyin Li & Tran T. A. Nghia, 2021. "On the Linear Convergence of Forward–Backward Splitting Method: Part I—Convergence Analysis," Journal of Optimization Theory and Applications, Springer, vol. 188(2), pages 378-401, February.
  • Handle: RePEc:spr:joptap:v:188:y:2021:i:2:d:10.1007_s10957-020-01787-7
    DOI: 10.1007/s10957-020-01787-7
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s10957-020-01787-7
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s10957-020-01787-7?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Dmitriy Drusvyatskiy & Adrian S. Lewis, 2018. "Error Bounds, Quadratic Growth, and Linear Convergence of Proximal Methods," Mathematics of Operations Research, INFORMS, vol. 43(3), pages 919-948, August.
    2. Ion Necoara & Yurii Nesterov & François Glineur, 2019. "Linear convergence of first order methods for non-strongly convex optimization," LIDAM Reprints CORE 3000, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    3. Regina S. Burachik & Alfredo N. Iusem, 2008. "Set-Valued Mappings and Enlargements of Monotone Operators," Springer Optimization and Its Applications, Springer, number 978-0-387-69757-4, September.
    4. Regina S. Burachik & Alfredo N. Iusem, 2008. "Enlargements of Monotone Operators," Springer Optimization and Its Applications, in: Set-Valued Mappings and Enlargements of Monotone Operators, chapter 0, pages 161-220, Springer.
    5. Patrick L. Combettes & Jean-Christophe Pesquet, 2011. "Proximal Splitting Methods in Signal Processing," Springer Optimization and Its Applications, in: Heinz H. Bauschke & Regina S. Burachik & Patrick L. Combettes & Veit Elser & D. Russell Luke & Henry (ed.), Fixed-Point Algorithms for Inverse Problems in Science and Engineering, chapter 0, pages 185-212, Springer.
    6. Heinz H. Bauschke & Jérôme Bolte & Marc Teboulle, 2017. "A Descent Lemma Beyond Lipschitz Gradient Continuity: First-Order Methods Revisited and Applications," Mathematics of Operations Research, INFORMS, vol. 42(2), pages 330-348, May.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Yunier Bello-Cruz & Guoyin Li & Tran Thai An Nghia, 2022. "Quadratic Growth Conditions and Uniqueness of Optimal Solution to Lasso," Journal of Optimization Theory and Applications, Springer, vol. 194(1), pages 167-190, July.
    2. Nguyen Huy Chieu & Nguyen Thi Quynh Trang & Ha Anh Tuan, 2022. "Quadratic Growth and Strong Metric Subregularity of the Subdifferential for a Class of Non-prox-regular Functions," Journal of Optimization Theory and Applications, Springer, vol. 194(3), pages 1081-1106, September.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Yunier Bello-Cruz & Guoyin Li & Tran Thai An Nghia, 2022. "Quadratic Growth Conditions and Uniqueness of Optimal Solution to Lasso," Journal of Optimization Theory and Applications, Springer, vol. 194(1), pages 167-190, July.
    2. Warunun Inthakon & Suthep Suantai & Panitarn Sarnmeta & Dawan Chumpungam, 2020. "A New Machine Learning Algorithm Based on Optimization Method for Regression and Classification Problems," Mathematics, MDPI, vol. 8(6), pages 1-17, June.
    3. Huynh Van Ngai & Nguyen Huu Tron & Michel Théra, 2014. "Metric Regularity of the Sum of Multifunctions and Applications," Journal of Optimization Theory and Applications, Springer, vol. 160(2), pages 355-390, February.
    4. Dawan Chumpungam & Panitarn Sarnmeta & Suthep Suantai, 2021. "A New Forward–Backward Algorithm with Line Searchand Inertial Techniques for Convex Minimization Problems with Applications," Mathematics, MDPI, vol. 9(13), pages 1-20, July.
    5. Walaa M. Moursi & Lieven Vandenberghe, 2019. "Douglas–Rachford Splitting for the Sum of a Lipschitz Continuous and a Strongly Monotone Operator," Journal of Optimization Theory and Applications, Springer, vol. 183(1), pages 179-198, October.
    6. Sedi Bartz & Minh N. Dao & Hung M. Phan, 2022. "Conical averagedness and convergence analysis of fixed point algorithms," Journal of Global Optimization, Springer, vol. 82(2), pages 351-373, February.
    7. Bello Cruz, J.Y. & Iusem, A.N., 2015. "Full convergence of an approximate projection method for nonsmooth variational inequalities," Mathematics and Computers in Simulation (MATCOM), Elsevier, vol. 114(C), pages 2-13.
    8. Regina S. Burachik & Minh N. Dao & Scott B. Lindstrom, 2021. "Generalized Bregman Envelopes and Proximity Operators," Journal of Optimization Theory and Applications, Springer, vol. 190(3), pages 744-778, September.
    9. Juan Pablo Luna & Claudia Sagastizábal & Mikhail Solodov, 2020. "A class of Benders decomposition methods for variational inequalities," Computational Optimization and Applications, Springer, vol. 76(3), pages 935-959, July.
    10. Heinz H. Bauschke & Warren L. Hare & Walaa M. Moursi, 2016. "On the Range of the Douglas–Rachford Operator," Mathematics of Operations Research, INFORMS, vol. 41(3), pages 884-897, August.
    11. Hsien-Chung Wu, 2018. "Near Fixed Point Theorems in Hyperspaces," Mathematics, MDPI, vol. 6(6), pages 1-15, May.
    12. Walaa M. Moursi, 2018. "The Forward–Backward Algorithm and the Normal Problem," Journal of Optimization Theory and Applications, Springer, vol. 176(3), pages 605-624, March.
    13. Dawan Chumpungam & Panitarn Sarnmeta & Suthep Suantai, 2022. "An Accelerated Convex Optimization Algorithm with Line Search and Applications in Machine Learning," Mathematics, MDPI, vol. 10(9), pages 1-20, April.
    14. Regina S. Burachik & Alfredo N. Iusem & Jefferson G. Melo, 2013. "An Inexact Modified Subgradient Algorithm for Primal-Dual Problems via Augmented Lagrangians," Journal of Optimization Theory and Applications, Springer, vol. 157(1), pages 108-131, April.
    15. L. C. Ceng & B. S. Mordukhovich & J. C. Yao, 2010. "Hybrid Approximate Proximal Method with Auxiliary Variational Inequality for Vector Optimization," Journal of Optimization Theory and Applications, Springer, vol. 146(2), pages 267-303, August.
    16. J. Bello Cruz & A. Iusem, 2010. "Convergence of direct methods for paramonotone variational inequalities," Computational Optimization and Applications, Springer, vol. 46(2), pages 247-263, June.
    17. Xin Jiang & Lieven Vandenberghe, 2023. "Bregman Three-Operator Splitting Methods," Journal of Optimization Theory and Applications, Springer, vol. 196(3), pages 936-972, March.
    18. Zamani, Moslem & Abbaszadehpeivasti, Hadi & de Klerk, Etienne, 2023. "The exact worst-case convergence rate of the alternating direction method of multipliers," Other publications TiSEM f30ae9e6-ed19-423f-bd1e-0, Tilburg University, School of Economics and Management.
    19. S. Bonettini & M. Prato & S. Rebegoldi, 2018. "A block coordinate variable metric linesearch based proximal gradient method," Computational Optimization and Applications, Springer, vol. 71(1), pages 5-52, September.
    20. Adrien B. Taylor & Julien M. Hendrickx & François Glineur, 2018. "Exact Worst-Case Convergence Rates of the Proximal Gradient Method for Composite Convex Minimization," Journal of Optimization Theory and Applications, Springer, vol. 178(2), pages 455-476, August.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:joptap:v:188:y:2021:i:2:d:10.1007_s10957-020-01787-7. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.