IDEAS home Printed from https://ideas.repec.org/a/spr/coopap/v67y2017i2d10.1007_s10589-017-9895-8.html
   My bibliography  Save this article

Solving nearly-separable quadratic optimization problems as nonsmooth equations

Author

Listed:
  • Frank E. Curtis

    (Lehigh University)

  • Arvind U. Raghunathan

    (Mitsubishi Electric Research Laboratories)

Abstract

An algorithm for solving nearly-separable quadratic optimization problems (QPs) is presented. The approach is based on applying a semismooth Newton method to solve the implicit complementarity problem arising as the first-order stationarity conditions of such a QP. An important feature of the approach is that, as in dual decomposition methods, separability of the dual function of the QP can be exploited in the search direction computation. Global convergence of the method is promoted by enforcing decrease in component(s) of a Fischer–Burmeister formulation of the complementarity conditions, either via a merit function or through a filter mechanism. The results of numerical experiments when solving convex and nonconvex instances are provided to illustrate the efficacy of the method.

Suggested Citation

  • Frank E. Curtis & Arvind U. Raghunathan, 2017. "Solving nearly-separable quadratic optimization problems as nonsmooth equations," Computational Optimization and Applications, Springer, vol. 67(2), pages 317-360, June.
  • Handle: RePEc:spr:coopap:v:67:y:2017:i:2:d:10.1007_s10589-017-9895-8
    DOI: 10.1007/s10589-017-9895-8
    as

    Download full text from publisher

    File URL: http://link.springer.com/10.1007/s10589-017-9895-8
    File Function: Abstract
    Download Restriction: Access to the full text of the articles in this series is restricted.

    File URL: https://libkey.io/10.1007/s10589-017-9895-8?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    As the access to this document is restricted, you may want to search for a different version of it.

    References listed on IDEAS

    as
    1. Jacek Gondzio & Andreas Grothey, 2007. "Parallel interior-point solver for structured quadratic programs: Application to financial planning problems," Annals of Operations Research, Springer, vol. 152(1), pages 319-339, July.
    2. Quoc Tran Dinh & Ion Necoara & Moritz Diehl, 2014. "Path-following gradient-based decomposition algorithms for separable convex optimization," Journal of Global Optimization, Springer, vol. 59(1), pages 59-80, May.
    3. Quoc Tran Dinh & Carlo Savorgnan & Moritz Diehl, 2013. "Combining Lagrangian decomposition and excessive gap smoothing technique for solving large-scale separable convex optimization problems," Computational Optimization and Applications, Springer, vol. 55(1), pages 75-111, May.
    4. NESTEROV, Yu., 2005. "Excessive gap technique in nonsmooth convex minimization," LIDAM Reprints CORE 1818, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    5. Todd S. Munson & Francisco Facchinei & Michael C. Ferris & Andreas Fischer & Christian Kanzow, 2001. "The Semismooth Algorithm for Large Scale Complementarity Problems," INFORMS Journal on Computing, INFORMS, vol. 13(4), pages 294-311, November.
    6. NESTEROV, Yu., 2005. "Smooth minimization of non-smooth functions," LIDAM Reprints CORE 1819, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    Full references (including those not matched with items on IDEAS)

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Stuart M. Harwood, 2021. "Analysis of the Alternating Direction Method of Multipliers for Nonconvex Problems," SN Operations Research Forum, Springer, vol. 2(1), pages 1-29, March.

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Quoc Tran-Dinh, 2017. "Adaptive smoothing algorithms for nonsmooth composite convex minimization," Computational Optimization and Applications, Springer, vol. 66(3), pages 425-451, April.
    2. Jueyou Li & Zhiyou Wu & Changzhi Wu & Qiang Long & Xiangyu Wang, 2016. "An Inexact Dual Fast Gradient-Projection Method for Separable Convex Optimization with Linear Coupled Constraints," Journal of Optimization Theory and Applications, Springer, vol. 168(1), pages 153-171, January.
    3. Guoyin Li & Alfred Ma & Ting Pong, 2014. "Robust least square semidefinite programming with applications," Computational Optimization and Applications, Springer, vol. 58(2), pages 347-379, June.
    4. Masaru Ito, 2016. "New results on subgradient methods for strongly convex optimization problems with a unified analysis," Computational Optimization and Applications, Springer, vol. 65(1), pages 127-172, September.
    5. Masoud Ahookhosh, 2019. "Accelerated first-order methods for large-scale convex optimization: nearly optimal complexity under strong convexity," Mathematical Methods of Operations Research, Springer;Gesellschaft für Operations Research (GOR);Nederlands Genootschap voor Besliskunde (NGB), vol. 89(3), pages 319-353, June.
    6. Daskalakis, Constantinos & Deckelbaum, Alan & Kim, Anthony, 2015. "Near-optimal no-regret algorithms for zero-sum games," Games and Economic Behavior, Elsevier, vol. 92(C), pages 327-348.
    7. Nima Rabiei & Jose Muñoz, 2015. "AAR-based decomposition algorithm for non-linear convex optimisation," Computational Optimization and Applications, Springer, vol. 62(3), pages 761-786, December.
    8. Yurii Nesterov, 2009. "Unconstrained Convex Minimization in Relative Scale," Mathematics of Operations Research, INFORMS, vol. 34(1), pages 180-193, February.
    9. Samid Hoda & Andrew Gilpin & Javier Peña & Tuomas Sandholm, 2010. "Smoothing Techniques for Computing Nash Equilibria of Sequential Games," Mathematics of Operations Research, INFORMS, vol. 35(2), pages 494-512, May.
    10. Jueyou Li & Guo Chen & Zhaoyang Dong & Zhiyou Wu, 2016. "A fast dual proximal-gradient method for separable convex optimization with linear coupled constraints," Computational Optimization and Applications, Springer, vol. 64(3), pages 671-697, July.
    11. Quoc Tran Dinh & Carlo Savorgnan & Moritz Diehl, 2013. "Combining Lagrangian decomposition and excessive gap smoothing technique for solving large-scale separable convex optimization problems," Computational Optimization and Applications, Springer, vol. 55(1), pages 75-111, May.
    12. Masoud Ahookhosh & Arnold Neumaier, 2018. "Solving structured nonsmooth convex optimization with complexity $$\mathcal {O}(\varepsilon ^{-1/2})$$ O ( ε - 1 / 2 )," TOP: An Official Journal of the Spanish Society of Statistics and Operations Research, Springer;Sociedad de Estadística e Investigación Operativa, vol. 26(1), pages 110-145, April.
    13. Guanghui Lan & Yuyuan Ouyang, 2022. "Accelerated gradient sliding for structured convex optimization," Computational Optimization and Applications, Springer, vol. 82(2), pages 361-394, June.
    14. DEVOLDER, Olivier & GLINEUR, François & NESTEROV, Yurii, 2011. "First-order methods of smooth convex optimization with inexact oracle," LIDAM Discussion Papers CORE 2011002, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    15. Jens Hübner & Martin Schmidt & Marc C. Steinbach, 2020. "Optimization techniques for tree-structured nonlinear problems," Computational Management Science, Springer, vol. 17(3), pages 409-436, October.
    16. Dirk Lorenz & Marc Pfetsch & Andreas Tillmann, 2014. "An infeasible-point subgradient method using adaptive approximate projections," Computational Optimization and Applications, Springer, vol. 57(2), pages 271-306, March.
    17. TAYLOR, Adrien B. & HENDRICKX, Julien M. & François GLINEUR, 2016. "Exact worst-case performance of first-order methods for composite convex optimization," LIDAM Discussion Papers CORE 2016052, Université catholique de Louvain, Center for Operations Research and Econometrics (CORE).
    18. Dimitris Bertsimas & Nishanth Mundru, 2021. "Sparse Convex Regression," INFORMS Journal on Computing, INFORMS, vol. 33(1), pages 262-279, January.
    19. Amir Beck & Shoham Sabach, 2015. "Weiszfeld’s Method: Old and New Results," Journal of Optimization Theory and Applications, Springer, vol. 164(1), pages 1-40, January.
    20. Donghwan Kim & Jeffrey A. Fessler, 2021. "Optimizing the Efficiency of First-Order Methods for Decreasing the Gradient of Smooth Convex Functions," Journal of Optimization Theory and Applications, Springer, vol. 188(1), pages 192-219, January.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:spr:coopap:v:67:y:2017:i:2:d:10.1007_s10589-017-9895-8. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Sonal Shukla or Springer Nature Abstracting and Indexing (email available below). General contact details of provider: http://www.springer.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.