IDEAS home Printed from https://ideas.repec.org/a/spr/coopap/v67y2017i2d10.1007_s10589-017-9896-7.html
   My bibliography  Save this article

Local and global convergence of a general inertial proximal splitting scheme for minimizing composite functions

Author

Listed:
  • Patrick R. Johnstone

    (University of Illinois)

  • Pierre Moulin

    (University of Illinois)

Abstract

This paper is concerned with convex composite minimization problems in a Hilbert space. In these problems, the objective is the sum of two closed, proper, and convex functions where one is smooth and the other admits a computationally inexpensive proximal operator. We analyze a family of generalized inertial proximal splitting algorithms (GIPSA) for solving such problems. We establish weak convergence of the generated sequence when the minimum is attained. Our analysis unifies and extends several previous results. We then focus on $$\ell _1$$ ℓ 1 -regularized optimization, which is the ubiquitous special case where the nonsmooth term is the $$\ell _1$$ ℓ 1 -norm. For certain parameter choices, GIPSA is amenable to a local analysis for this problem. For these choices we show that GIPSA achieves finite “active manifold identification”, i.e. convergence in a finite number of iterations to the optimal support and sign, after which GIPSA reduces to minimizing a local smooth function. We prove local linear convergence under either restricted strong convexity or a strict complementarity condition. We determine the rate in terms of the inertia, stepsize, and local curvature. Our local analysis is applicable to certain recent variants of the Fast Iterative Shrinkage–Thresholding Algorithm (FISTA), for which we establish active manifold identification and local linear convergence. Based on our analysis we propose a momentum restart scheme in these FISTA variants to obtain the optimal local linear convergence rate while maintaining desirable global properties.

Suggested Citation

  • Patrick R. Johnstone & Pierre Moulin, 2017. "Local and global convergence of a general inertial proximal splitting scheme for minimizing composite functions," Computational Optimization and Applications, Springer, vol. 67(2), pages 259-292, June.
  • Handle: RePEc:spr:coopap:v:67:y:2017:i:2:d:10.1007_s10589-017-9896-7
    DOI: 10.1007/s10589-017-9896-7
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s10589-017-9896-7
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s10589-017-9896-7?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. NESTEROV, Yurii, 2013. "Gradient methods for minimizing composite functions," LIDAM Reprints CORE 2510, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    2. Renato D. C. Monteiro & Camilo Ortiz & Benar F. Svaiter, 2016. "An adaptive accelerated first-order method for convex optimization," Computational Optimization and Applications, Springer, vol. 64(1), pages 31-73, May.
    3. Friedman, Jerome H. & Hastie, Trevor & Tibshirani, Rob, 2010. "Regularization Paths for Generalized Linear Models via Coordinate Descent," Journal of Statistical Software, Foundation for Open Access Statistics, vol. 33(i01).
    4. Laurent Condat, 2013. "A Primal–Dual Splitting Method for Convex Optimization Involving Lipschitzian, Proximable and Linear Composite Terms," Journal of Optimization Theory and Applications, Springer, vol. 158(2), pages 460-479, August.
    5. Boţ, Radu Ioan & Csetnek, Ernö Robert & Hendrich, Christopher, 2015. "Inertial Douglas–Rachford splitting for monotone inclusion problems," Applied Mathematics and Computation, Elsevier, vol. 256(C), pages 472-487.
    6. Patrick L. Combettes & Jean-Christophe Pesquet, 2011. "Proximal Splitting Methods in Signal Processing," Springer Optimization and Its Applications, in: Heinz H. Bauschke & Regina S. Burachik & Patrick L. Combettes & Veit Elser & D. Russell Luke & Henry (ed.), Fixed-Point Algorithms for Inverse Problems in Science and Engineering, chapter 0, pages 185-212, Springer.
    7. Unknown, 2005. "Forward," 2005 Conference: Slovenia in the EU - Challenges for Agriculture, Food Science and Rural Affairs, November 10-11, 2005, Moravske Toplice, Slovenia 183804, Slovenian Association of Agricultural Economists (DAES).
    8. Hui Zhang & Wotao Yin & Lizhi Cheng, 2015. "Necessary and Sufficient Conditions of Solution Uniqueness in 1-Norm Minimization," Journal of Optimization Theory and Applications, Springer, vol. 164(1), pages 109-122, January.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Zhongming Wu & Min Li, 2019. "General inertial proximal gradient method for a class of nonconvex nonsmooth optimization problems," Computational Optimization and Applications, Springer, vol. 73(1), pages 129-158, May.
    2. E. M. Bednarczuk & A. Jezierska & K. E. Rutkowski, 2018. "Proximal primal–dual best approximation algorithm with memory," Computational Optimization and Applications, Springer, vol. 71(3), pages 767-794, December.
    3. Szilárd Csaba László, 2023. "A Forward–Backward Algorithm With Different Inertial Terms for Structured Non-Convex Minimization Problems," Journal of Optimization Theory and Applications, Springer, vol. 198(1), pages 387-427, July.
    4. Xiaoqi Yang & Chenchen Zu, 2022. "Convergence of Inexact Quasisubgradient Methods with Extrapolation," Journal of Optimization Theory and Applications, Springer, vol. 193(1), pages 676-703, June.
    5. Zhongming Wu & Chongshou Li & Min Li & Andrew Lim, 2021. "Inertial proximal gradient methods with Bregman regularization for a class of nonconvex optimization problems," Journal of Global Optimization, Springer, vol. 79(3), pages 617-644, March.
    6. Fan Wu & Wei Bian, 2020. "Accelerated iterative hard thresholding algorithm for $$l_0$$l0 regularized regression problem," Journal of Global Optimization, Springer, vol. 76(4), pages 819-840, April.
    7. Xiaoya Zhang & Wei Peng & Hui Zhang, 2022. "Inertial proximal incremental aggregated gradient method with linear convergence guarantees," Mathematical Methods of Operations Research, Springer;Gesellschaft für Operations Research (GOR);Nederlands Genootschap voor Besliskunde (NGB), vol. 96(2), pages 187-213, October.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Sun, Shilin & Wang, Tianyang & Yang, Hongxing & Chu, Fulei, 2022. "Damage identification of wind turbine blades using an adaptive method for compressive beamforming based on the generalized minimax-concave penalty function," Renewable Energy, Elsevier, vol. 181(C), pages 59-70.
    2. Zhongming Wu & Min Li, 2019. "General inertial proximal gradient method for a class of nonconvex nonsmooth optimization problems," Computational Optimization and Applications, Springer, vol. 73(1), pages 129-158, May.
    3. Luis Briceño-Arias & Julio Deride & Cristian Vega, 2022. "Random Activations in Primal-Dual Splittings for Monotone Inclusions with a Priori Information," Journal of Optimization Theory and Applications, Springer, vol. 192(1), pages 56-81, January.
    4. Donghwan Kim & Jeffrey A. Fessler, 2018. "Adaptive Restart of the Optimized Gradient Method for Convex Optimization," Journal of Optimization Theory and Applications, Springer, vol. 178(1), pages 240-263, July.
    5. Puya Latafat & Panagiotis Patrinos, 2017. "Asymmetric forward–backward–adjoint splitting for solving monotone inclusions involving three operators," Computational Optimization and Applications, Springer, vol. 68(1), pages 57-93, September.
    6. TAYLOR, Adrien B. & HENDRICKX, Julien M. & François GLINEUR, 2016. "Exact worst-case performance of first-order methods for composite convex optimization," LIDAM Discussion Papers CORE 2016052, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    7. Weibin Mo & Yufeng Liu, 2022. "Efficient learning of optimal individualized treatment rules for heteroscedastic or misspecified treatment‐free effect models," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 84(2), pages 440-472, April.
    8. David Degras, 2021. "Sparse group fused lasso for model segmentation: a hybrid approach," Advances in Data Analysis and Classification, Springer;German Classification Society - Gesellschaft für Klassifikation (GfKl);Japanese Classification Society (JCS);Classification and Data Analysis Group of the Italian Statistical Society (CLADAG);International Federation of Classification Societies (IFCS), vol. 15(3), pages 625-671, September.
    9. Christian Kanzow & Theresa Lechner, 2021. "Globalized inexact proximal Newton-type methods for nonconvex composite functions," Computational Optimization and Applications, Springer, vol. 78(2), pages 377-410, March.
    10. Weiwei Kong & Renato D. C. Monteiro, 2022. "Accelerated inexact composite gradient methods for nonconvex spectral optimization problems," Computational Optimization and Applications, Springer, vol. 82(3), pages 673-715, July.
    11. Bo Wen & Xiaojun Chen & Ting Kei Pong, 2018. "A proximal difference-of-convex algorithm with extrapolation," Computational Optimization and Applications, Springer, vol. 69(2), pages 297-324, March.
    12. Hedy Attouch & Zaki Chbani & Jalal Fadili & Hassan Riahi, 2022. "Fast Convergence of Dynamical ADMM via Time Scaling of Damped Inertial Dynamics," Journal of Optimization Theory and Applications, Springer, vol. 193(1), pages 704-736, June.
    13. Lorenzo Stella & Andreas Themelis & Panagiotis Patrinos, 2017. "Forward–backward quasi-Newton methods for nonsmooth optimization problems," Computational Optimization and Applications, Springer, vol. 67(3), pages 443-487, July.
    14. Chadarat Thongphaen & Warunun Inthakon & Suthep Suantai & Narawadee Phudolsitthiphat, 2022. "Common Attractive Point Results for Two Generalized Nonexpansive Mappings in Uniformly Convex Banach Spaces," Mathematics, MDPI, vol. 10(8), pages 1-18, April.
    15. Cheik Traoré & Saverio Salzo & Silvia Villa, 2023. "Convergence of an asynchronous block-coordinate forward-backward algorithm for convex composite optimization," Computational Optimization and Applications, Springer, vol. 86(1), pages 303-344, September.
    16. Hongwei Liu & Ting Wang & Zexian Liu, 2022. "Some modified fast iterative shrinkage thresholding algorithms with a new adaptive non-monotone stepsize strategy for nonsmooth and convex minimization problems," Computational Optimization and Applications, Springer, vol. 83(2), pages 651-691, November.
    17. Yunier Bello-Cruz & Guoyin Li & Tran Thai An Nghia, 2022. "Quadratic Growth Conditions and Uniqueness of Optimal Solution to Lasso," Journal of Optimization Theory and Applications, Springer, vol. 194(1), pages 167-190, July.
    18. Yekini Shehu & Aviv Gibali, 2020. "Inertial Krasnoselskii–Mann Method in Banach Spaces," Mathematics, MDPI, vol. 8(4), pages 1-13, April.
    19. Jian Huang & Yuling Jiao & Lican Kang & Jin Liu & Yanyan Liu & Xiliang Lu, 2022. "GSDAR: a fast Newton algorithm for $$\ell _0$$ ℓ 0 regularized generalized linear models with statistical guarantee," Computational Statistics, Springer, vol. 37(1), pages 507-533, March.
    20. Masoud Ahookhosh & Arnold Neumaier, 2018. "Solving structured nonsmooth convex optimization with complexity $$\mathcal {O}(\varepsilon ^{-1/2})$$ O ( ε - 1 / 2 )," TOP: An Official Journal of the Spanish Society of Statistics and Operations Research, Springer;Sociedad de Estadística e Investigación Operativa, vol. 26(1), pages 110-145, April.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:coopap:v:67:y:2017:i:2:d:10.1007_s10589-017-9896-7. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.