IDEAS home Printed from https://ideas.repec.org/a/spr/joptap/v193y2022i1d10.1007_s10957-021-01957-1.html
   My bibliography  Save this article

An Accelerated Coordinate Gradient Descent Algorithm for Non-separable Composite Optimization

Author

Listed:
  • Aviad Aberdam

    (Technion - Israel Institute of Technology)

  • Amir Beck

    (Tel Aviv University)

Abstract

Coordinate descent algorithms are popular in machine learning and large-scale data analysis problems due to their low computational cost iterative schemes and their improved performances. In this work, we define a monotone accelerated coordinate gradient descent-type method for problems consisting of minimizing $$f+g$$ f + g , where f is quadratic and g is nonsmooth and non-separable and has a low-complexity proximal mapping. The algorithm is enabled by employing the forward–backward envelope, a composite envelope that possess an exact smooth reformulation of $$f+g$$ f + g . We prove the algorithm achieves a convergence rate of $$O(1/k^{1.5})$$ O ( 1 / k 1.5 ) in terms of the original objective function, improving current coordinate descent-type algorithms. In addition, we describe an adaptive variant of the algorithm that backtracks the spectral information and coordinate Lipschitz constants of the problem. We numerically examine our algorithms on various settings, including two-dimensional total-variation-based image inpainting problems, showing a clear advantage in performance over current coordinate descent-type methods.

Suggested Citation

  • Aviad Aberdam & Amir Beck, 2022. "An Accelerated Coordinate Gradient Descent Algorithm for Non-separable Composite Optimization," Journal of Optimization Theory and Applications, Springer, vol. 193(1), pages 219-246, June.
  • Handle: RePEc:spr:joptap:v:193:y:2022:i:1:d:10.1007_s10957-021-01957-1
    DOI: 10.1007/s10957-021-01957-1
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s10957-021-01957-1
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s10957-021-01957-1?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Lorenzo Stella & Andreas Themelis & Panagiotis Patrinos, 2017. "Forward–backward quasi-Newton methods for nonsmooth optimization problems," Computational Optimization and Applications, Springer, vol. 67(3), pages 443-487, July.
    2. NESTEROV, Yurii, 2013. "Gradient methods for minimizing composite functions," LIDAM Reprints CORE 2510, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    3. NESTEROV, Yurii, 2012. "Efficiency of coordinate descent methods on huge-scale optimization problems," LIDAM Reprints CORE 2511, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    4. Friedman, Jerome H. & Hastie, Trevor & Tibshirani, Rob, 2010. "Regularization Paths for Generalized Linear Models via Coordinate Descent," Journal of Statistical Software, Foundation for Open Access Statistics, vol. 33(i01).
    5. N. Maculan & C.P. Santiago & E.M. Macambira & M.H.C. Jardim, 2003. "An O(n) Algorithm for Projecting a Vector on the Intersection of a Hyperplane and a Box in Rn," Journal of Optimization Theory and Applications, Springer, vol. 117(3), pages 553-574, June.
    6. Pontus Giselsson & Mattias Fält, 2018. "Envelope Functions: Unifications and Further Properties," Journal of Optimization Theory and Applications, Springer, vol. 178(3), pages 673-698, September.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. TAYLOR, Adrien B. & HENDRICKX, Julien M. & François GLINEUR, 2016. "Exact worst-case performance of first-order methods for composite convex optimization," LIDAM Discussion Papers CORE 2016052, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    2. Weibin Mo & Yufeng Liu, 2022. "Efficient learning of optimal individualized treatment rules for heteroscedastic or misspecified treatment‐free effect models," Journal of the Royal Statistical Society Series B, Royal Statistical Society, vol. 84(2), pages 440-472, April.
    3. Reza Eghbali & Maryam Fazel, 2017. "Decomposable norm minimization with proximal-gradient homotopy algorithm," Computational Optimization and Applications, Springer, vol. 66(2), pages 345-381, March.
    4. Christian Kanzow & Theresa Lechner, 2021. "Globalized inexact proximal Newton-type methods for nonconvex composite functions," Computational Optimization and Applications, Springer, vol. 78(2), pages 377-410, March.
    5. Masoud Ahookhosh & Le Thi Khanh Hien & Nicolas Gillis & Panagiotis Patrinos, 2021. "Multi-block Bregman proximal alternating linearized minimization and its application to orthogonal nonnegative matrix factorization," Computational Optimization and Applications, Springer, vol. 79(3), pages 681-715, July.
    6. Yangyang Xu & Shuzhong Zhang, 2018. "Accelerated primal–dual proximal block coordinate updating methods for constrained convex optimization," Computational Optimization and Applications, Springer, vol. 70(1), pages 91-128, May.
    7. Cheik Traoré & Saverio Salzo & Silvia Villa, 2023. "Convergence of an asynchronous block-coordinate forward-backward algorithm for convex composite optimization," Computational Optimization and Applications, Springer, vol. 86(1), pages 303-344, September.
    8. Jian Huang & Yuling Jiao & Lican Kang & Jin Liu & Yanyan Liu & Xiliang Lu, 2022. "GSDAR: a fast Newton algorithm for $$\ell _0$$ ℓ 0 regularized generalized linear models with statistical guarantee," Computational Statistics, Springer, vol. 37(1), pages 507-533, March.
    9. Md Sarowar Morshed & Md Saiful Islam & Md. Noor-E-Alam, 2020. "Accelerated sampling Kaczmarz Motzkin algorithm for the linear feasibility problem," Journal of Global Optimization, Springer, vol. 77(2), pages 361-382, June.
    10. Jean-Charles Richard & Thierry Roncalli, 2019. "Constrained Risk Budgeting Portfolios: Theory, Algorithms, Applications & Puzzles," Papers 1902.05710, arXiv.org.
    11. Yanli Liu & Wotao Yin, 2019. "An Envelope for Davis–Yin Splitting and Strict Saddle-Point Avoidance," Journal of Optimization Theory and Applications, Springer, vol. 181(2), pages 567-587, May.
    12. Patrick R. Johnstone & Pierre Moulin, 2017. "Local and global convergence of a general inertial proximal splitting scheme for minimizing composite functions," Computational Optimization and Applications, Springer, vol. 67(2), pages 259-292, June.
    13. Adrien B. Taylor & Julien M. Hendrickx & François Glineur, 2018. "Exact Worst-Case Convergence Rates of the Proximal Gradient Method for Composite Convex Minimization," Journal of Optimization Theory and Applications, Springer, vol. 178(2), pages 455-476, August.
    14. Majid Jahani & Naga Venkata C. Gudapati & Chenxin Ma & Rachael Tappenden & Martin Takáč, 2021. "Fast and safe: accelerated gradient methods with optimality certificates and underestimate sequences," Computational Optimization and Applications, Springer, vol. 79(2), pages 369-404, June.
    15. Olivier Fercoq & Zheng Qu, 2020. "Restarting the accelerated coordinate descent method with a rough strong convexity estimate," Computational Optimization and Applications, Springer, vol. 75(1), pages 63-91, January.
    16. Quoc Tran-Dinh, 2019. "Proximal alternating penalty algorithms for nonsmooth constrained convex optimization," Computational Optimization and Applications, Springer, vol. 72(1), pages 1-43, January.
    17. Sarah Perrin & Thierry Roncalli, 2019. "Machine Learning Optimization Algorithms & Portfolio Allocation," Papers 1909.10233, arXiv.org.
    18. R. Lopes & S. A. Santos & P. J. S. Silva, 2019. "Accelerating block coordinate descent methods with identification strategies," Computational Optimization and Applications, Springer, vol. 72(3), pages 609-640, April.
    19. Mingyi Hong & Tsung-Hui Chang & Xiangfeng Wang & Meisam Razaviyayn & Shiqian Ma & Zhi-Quan Luo, 2020. "A Block Successive Upper-Bound Minimization Method of Multipliers for Linearly Constrained Convex Optimization," Mathematics of Operations Research, INFORMS, vol. 45(3), pages 833-861, August.
    20. Tutz, Gerhard & Pößnecker, Wolfgang & Uhlmann, Lorenz, 2015. "Variable selection in general multinomial logit models," Computational Statistics & Data Analysis, Elsevier, vol. 82(C), pages 207-222.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:joptap:v:193:y:2022:i:1:d:10.1007_s10957-021-01957-1. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.