IDEAS home Printed from https://ideas.repec.org/a/spr/joptap/v171y2016i1d10.1007_s10957-016-0991-1.html
   My bibliography  Save this article

New Augmented Lagrangian-Based Proximal Point Algorithm for Convex Optimization with Equality Constraints

Author

Listed:
  • Yuan Shen

    (Nanjing University of Finance and Economics)

  • Hongyong Wang

    (Nanjing University of Finance and Economics)

Abstract

The augmented Lagrangian method is a classic and efficient method for solving constrained optimization problems. However, its efficiency is still, to a large extent, dependent on how efficient the subproblem be solved. When an accurate solution to the subproblem is computationally expensive, it is more practical to relax the subproblem. Specifically, when the objective function has a certain favorable structure, the relaxed subproblem yields a closed-form solution that can be solved efficiently. However, the resulting algorithm usually suffers from a slower convergence rate than the augmented Lagrangian method. In this paper, based on the relaxed subproblem, we propose a new algorithm with a faster convergence rate. Numerical results using the proposed approach are reported for three specific applications. The output is compared with the corresponding results from state-of-the-art algorithms, and it is shown that the efficiency of the proposed method is superior to that of existing approaches.

Suggested Citation

  • Yuan Shen & Hongyong Wang, 2016. "New Augmented Lagrangian-Based Proximal Point Algorithm for Convex Optimization with Equality Constraints," Journal of Optimization Theory and Applications, Springer, vol. 171(1), pages 251-261, October.
  • Handle: RePEc:spr:joptap:v:171:y:2016:i:1:d:10.1007_s10957-016-0991-1
    DOI: 10.1007/s10957-016-0991-1
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s10957-016-0991-1
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s10957-016-0991-1?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to

    for a different version of it.

    References listed on IDEAS

    as
    1. R. S. Burachik & S. Scheimberg & B. F. Svaiter, 2001. "Robustness of the Hybrid Extragradient Proximal-Point Algorithm," Journal of Optimization Theory and Applications, Springer, vol. 111(1), pages 117-136, October.
    2. NESTEROV, Yu., 2005. "Smooth minimization of non-smooth functions," LIDAM Reprints CORE 1819, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    3. R. T. Rockafellar, 1976. "Augmented Lagrangians and Applications of the Proximal Point Algorithm in Convex Programming," Mathematics of Operations Research, INFORMS, vol. 1(2), pages 97-116, May.
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Feng Ma, 2019. "On relaxation of some customized proximal point algorithms for convex minimization: from variational inequality perspective," Computational Optimization and Applications, Springer, vol. 73(3), pages 871-901, July.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Ya-Feng Liu & Xin Liu & Shiqian Ma, 2019. "On the Nonergodic Convergence Rate of an Inexact Augmented Lagrangian Framework for Composite Convex Programming," Mathematics of Operations Research, INFORMS, vol. 44(2), pages 632-650, May.
    2. Guoqiang Wang & Bo Yu, 2019. "PAL-Hom method for QP and an application to LP," Computational Optimization and Applications, Springer, vol. 73(1), pages 311-352, May.
    3. Mauricio Romero Sicre, 2020. "On the complexity of a hybrid proximal extragradient projective method for solving monotone inclusion problems," Computational Optimization and Applications, Springer, vol. 76(3), pages 991-1019, July.
    4. Jean-Pierre Crouzeix & Abdelhak Hassouni & Eladio Ocaña, 2023. "A Short Note on the Twice Differentiability of the Marginal Function of a Convex Function," Journal of Optimization Theory and Applications, Springer, vol. 198(2), pages 857-867, August.
    5. Liang Chen & Anping Liao, 2020. "On the Convergence Properties of a Second-Order Augmented Lagrangian Method for Nonlinear Programming Problems with Inequality Constraints," Journal of Optimization Theory and Applications, Springer, vol. 187(1), pages 248-265, October.
    6. Stefano Cipolla & Jacek Gondzio, 2023. "Proximal Stabilized Interior Point Methods and Low-Frequency-Update Preconditioning Techniques," Journal of Optimization Theory and Applications, Springer, vol. 197(3), pages 1061-1103, June.
    7. Bingsheng He & Li-Zhi Liao & Xiang Wang, 2012. "Proximal-like contraction methods for monotone variational inequalities in a unified framework I: Effective quadruplet and primary methods," Computational Optimization and Applications, Springer, vol. 51(2), pages 649-679, March.
    8. Marwan A. Kutbi & Abdul Latif & Xiaolong Qin, 2019. "Convergence of Two Splitting Projection Algorithms in Hilbert Spaces," Mathematics, MDPI, vol. 7(10), pages 1-13, October.
    9. Brecht Evens & Puya Latafat & Panagiotis Patrinos, 2025. "Convergence of the Chambolle–Pock Algorithm in the Absence of Monotonicity," Journal of Optimization Theory and Applications, Springer, vol. 206(1), pages 1-45, July.
    10. Darinka Dentcheva & Gabriela Martinez & Eli Wolfhagen, 2016. "Augmented Lagrangian Methods for Solving Optimization Problems with Stochastic-Order Constraints," Operations Research, INFORMS, vol. 64(6), pages 1451-1465, December.
    11. Gui-Hua Lin & Zhen-Ping Yang & Hai-An Yin & Jin Zhang, 2023. "A dual-based stochastic inexact algorithm for a class of stochastic nonsmooth convex composite problems," Computational Optimization and Applications, Springer, vol. 86(2), pages 669-710, November.
    12. Dirk Lorenz & Marc Pfetsch & Andreas Tillmann, 2014. "An infeasible-point subgradient method using adaptive approximate projections," Computational Optimization and Applications, Springer, vol. 57(2), pages 271-306, March.
    13. Xiaoming Yuan, 2011. "An improved proximal alternating direction method for monotone variational inequalities with separable structure," Computational Optimization and Applications, Springer, vol. 49(1), pages 17-29, May.
    14. Zhu, Daoli & Marcotte, Patrice, 1995. "Coupling the auxiliary problem principle with descent methods of pseudoconvex programming," European Journal of Operational Research, Elsevier, vol. 83(3), pages 670-685, June.
    15. Jueyou Li & Zhiyou Wu & Changzhi Wu & Qiang Long & Xiangyu Wang, 2016. "An Inexact Dual Fast Gradient-Projection Method for Separable Convex Optimization with Linear Coupled Constraints," Journal of Optimization Theory and Applications, Springer, vol. 168(1), pages 153-171, January.
    16. Guoyin Li & Alfred Ma & Ting Pong, 2014. "Robust least square semidefinite programming with applications," Computational Optimization and Applications, Springer, vol. 58(2), pages 347-379, June.
    17. Masaru Ito, 2016. "New results on subgradient methods for strongly convex optimization problems with a unified analysis," Computational Optimization and Applications, Springer, vol. 65(1), pages 127-172, September.
    18. repec:cdl:itsdav:qt89s5s8hn is not listed on IDEAS
    19. TAYLOR, Adrien B. & HENDRICKX, Julien M. & François GLINEUR, 2016. "Exact worst-case performance of first-order methods for composite convex optimization," LIDAM Discussion Papers CORE 2016052, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    20. Dimitris Bertsimas & Nishanth Mundru, 2021. "Sparse Convex Regression," INFORMS Journal on Computing, INFORMS, vol. 33(1), pages 262-279, January.
    21. Amir Beck & Shoham Sabach, 2015. "Weiszfeld’s Method: Old and New Results," Journal of Optimization Theory and Applications, Springer, vol. 164(1), pages 1-40, January.

    More about this item

    Keywords

    ;
    ;
    ;
    ;
    ;

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:joptap:v:171:y:2016:i:1:d:10.1007_s10957-016-0991-1. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.