IDEAS home Printed from https://ideas.repec.org/a/inm/orijoc/v34y2022i6p2989-3006.html
   My bibliography  Save this article

Solving Stochastic Optimization with Expectation Constraints Efficiently by a Stochastic Augmented Lagrangian-Type Algorithm

Author

Listed:
  • Liwei Zhang

    (School of Mathematical Sciences, Dalian University of Technology, 116024 Dalian, China)

  • Yule Zhang

    (School of Science, Dalian Maritime University, 116026 Dalian, China)

  • Jia Wu

    (School of Mathematical Sciences, Dalian University of Technology, 116024 Dalian, China)

  • Xiantao Xiao

    (School of Mathematical Sciences, Dalian University of Technology, 116024 Dalian, China)

Abstract

This paper considers the problem of minimizing a convex expectation function with a set of inequality convex expectation constraints. We propose a stochastic augmented Lagrangian-type algorithm—namely, the stochastic linearized proximal method of multipliers—to solve this convex stochastic optimization problem. This algorithm can be roughly viewed as a hybrid of stochastic approximation and the traditional proximal method of multipliers. Under mild conditions, we show that this algorithm exhibits O ( K − 1 / 2 ) expected convergence rates for both objective reduction and constraint violation if parameters in the algorithm are properly chosen, where K denotes the number of iterations. Moreover, we show that, with high probability, the algorithm has a O ( log ( K ) K − 1 / 2 ) constraint violation bound and O ( log 3 / 2 ( K ) K − 1 / 2 ) objective bound. Numerical results demonstrate that the proposed algorithm is efficient.

Suggested Citation

  • Liwei Zhang & Yule Zhang & Jia Wu & Xiantao Xiao, 2022. "Solving Stochastic Optimization with Expectation Constraints Efficiently by a Stochastic Augmented Lagrangian-Type Algorithm," INFORMS Journal on Computing, INFORMS, vol. 34(6), pages 2989-3006, November.
  • Handle: RePEc:inm:orijoc:v:34:y:2022:i:6:p:2989-3006
    DOI: 10.1287/ijoc.2022.1228
    as

    Download full text from publisher

    File URL: http://dx.doi.org/10.1287/ijoc.2022.1228
    Download Restriction: no

    File URL: https://libkey.io/10.1287/ijoc.2022.1228?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    References listed on IDEAS

    as
    1. Panos Parpas & Berç Rustem, 2007. "Computational Assessment of Nested Benders and Augmented Lagrangian Decomposition for Mean-Variance Multistage Stochastic Problems," INFORMS Journal on Computing, INFORMS, vol. 19(2), pages 239-247, May.
    2. Neslihan Fidan Keçeci & Viktor Kuzmenko & Stan Uryasev, 2016. "Portfolios Dominating Indices: Optimization with Second-Order Stochastic Dominance Constraints vs. Minimum and Mean Variance Portfolios," JRFM, MDPI, vol. 9(4), pages 1-14, October.
    3. R. T. Rockafellar, 1976. "Augmented Lagrangians and Applications of the Proximal Point Algorithm in Convex Programming," Mathematics of Operations Research, INFORMS, vol. 1(2), pages 97-116, May.
    4. Darinka Dentcheva & Gabriela Martinez & Eli Wolfhagen, 2016. "Augmented Lagrangian Methods for Solving Optimization Problems with Stochastic-Order Constraints," Operations Research, INFORMS, vol. 64(6), pages 1451-1465, December.
    5. Xiaodi Bai & Jie Sun & Xiaojin Zheng, 2021. "An Augmented Lagrangian Decomposition Method for Chance-Constrained Optimization Problems," INFORMS Journal on Computing, INFORMS, vol. 33(3), pages 1056-1069, July.
    6. Guanghui Lan & Zhiqiang Zhou, 2020. "Algorithms for stochastic optimization with function or expectation constraints," Computational Optimization and Applications, Springer, vol. 76(2), pages 461-498, June.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Mauricio Romero Sicre, 2020. "On the complexity of a hybrid proximal extragradient projective method for solving monotone inclusion problems," Computational Optimization and Applications, Springer, vol. 76(3), pages 991-1019, July.
    2. Qiuxia Yang, 2020. "Fiscal Transparency and Public Service Quality Association: Evidence from 12 Coastal Provinces and Cities of China," JRFM, MDPI, vol. 14(1), pages 1-14, December.
    3. Jean-Pierre Crouzeix & Abdelhak Hassouni & Eladio Ocaña, 2023. "A Short Note on the Twice Differentiability of the Marginal Function of a Convex Function," Journal of Optimization Theory and Applications, Springer, vol. 198(2), pages 857-867, August.
    4. Liang Chen & Anping Liao, 2020. "On the Convergence Properties of a Second-Order Augmented Lagrangian Method for Nonlinear Programming Problems with Inequality Constraints," Journal of Optimization Theory and Applications, Springer, vol. 187(1), pages 248-265, October.
    5. Li Xia, 2020. "Risk‐Sensitive Markov Decision Processes with Combined Metrics of Mean and Variance," Production and Operations Management, Production and Operations Management Society, vol. 29(12), pages 2808-2827, December.
    6. Stefano Cipolla & Jacek Gondzio, 2023. "Proximal Stabilized Interior Point Methods and Low-Frequency-Update Preconditioning Techniques," Journal of Optimization Theory and Applications, Springer, vol. 197(3), pages 1061-1103, June.
    7. Bingsheng He & Li-Zhi Liao & Xiang Wang, 2012. "Proximal-like contraction methods for monotone variational inequalities in a unified framework I: Effective quadruplet and primary methods," Computational Optimization and Applications, Springer, vol. 51(2), pages 649-679, March.
    8. Marwan A. Kutbi & Abdul Latif & Xiaolong Qin, 2019. "Convergence of Two Splitting Projection Algorithms in Hilbert Spaces," Mathematics, MDPI, vol. 7(10), pages 1-13, October.
    9. Darinka Dentcheva & Gabriela Martinez & Eli Wolfhagen, 2016. "Augmented Lagrangian Methods for Solving Optimization Problems with Stochastic-Order Constraints," Operations Research, INFORMS, vol. 64(6), pages 1451-1465, December.
    10. Gui-Hua Lin & Zhen-Ping Yang & Hai-An Yin & Jin Zhang, 2023. "A dual-based stochastic inexact algorithm for a class of stochastic nonsmooth convex composite problems," Computational Optimization and Applications, Springer, vol. 86(2), pages 669-710, November.
    11. Xiaoming Yuan, 2011. "An improved proximal alternating direction method for monotone variational inequalities with separable structure," Computational Optimization and Applications, Springer, vol. 49(1), pages 17-29, May.
    12. Zhu, Daoli & Marcotte, Patrice, 1995. "Coupling the auxiliary problem principle with descent methods of pseudoconvex programming," European Journal of Operational Research, Elsevier, vol. 83(3), pages 670-685, June.
    13. Guo, Zhaomiao & Fan, Yueyue, 2017. "A Stochastic Multi-Agent Optimization Model for Energy Infrastructure Planning Under Uncertainty and Competition," Institute of Transportation Studies, Working Paper Series qt89s5s8hn, Institute of Transportation Studies, UC Davis.
    14. Yong-Jin Liu & Jing Yu, 2023. "A semismooth Newton based dual proximal point algorithm for maximum eigenvalue problem," Computational Optimization and Applications, Springer, vol. 85(2), pages 547-582, June.
    15. Julian Rasch & Antonin Chambolle, 2020. "Inexact first-order primal–dual algorithms," Computational Optimization and Applications, Springer, vol. 76(2), pages 381-430, June.
    16. A. Ruszczynski, 1994. "On Augmented Lagrangian Decomposition Methods For Multistage Stochastic Programs," Working Papers wp94005, International Institute for Applied Systems Analysis.
    17. Jason Narsoo, 2017. "Performance Analysis of Portfolio Optimisation Strategies: Evidence from the Exchange Market," International Journal of Economics and Finance, Canadian Center of Science and Education, vol. 9(6), pages 124-132, June.
    18. R. S. Burachik & S. Scheimberg & B. F. Svaiter, 2001. "Robustness of the Hybrid Extragradient Proximal-Point Algorithm," Journal of Optimization Theory and Applications, Springer, vol. 111(1), pages 117-136, October.
    19. Min Tao & Xiaoming Yuan, 2018. "The generalized proximal point algorithm with step size 2 is not necessarily convergent," Computational Optimization and Applications, Springer, vol. 70(3), pages 827-839, July.
    20. A. F. Izmailov & M. V. Solodov, 2022. "Perturbed Augmented Lagrangian Method Framework with Applications to Proximal and Smoothed Variants," Journal of Optimization Theory and Applications, Springer, vol. 193(1), pages 491-522, June.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:inm:orijoc:v:34:y:2022:i:6:p:2989-3006. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Chris Asher (email available below). General contact details of provider: https://edirc.repec.org/data/inforea.html .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.