IDEAS home Printed from https://ideas.repec.org/a/inm/ormoor/v47y2022i3p2082-2111.html
   My bibliography  Save this article

Distributed Stochastic Optimization with Large Delays

Author

Listed:
  • Zhengyuan Zhou

    (Stern School of Business, New York University, New York, New York 10012)

  • Panayotis Mertikopoulos

    (Univ. Grenoble Alpes, CNRS, Inria, LIG, 38000 Grenoble, France)

  • Nicholas Bambos

    (Department of Management Science and Engineering, Stanford University, Stanford, California 94305)

  • Peter Glynn

    (Department of Management Science and Engineering, Stanford University, Stanford, California 94305)

  • Yinyu Ye

    (Department of Management Science and Engineering, Stanford University, Stanford, California 94305)

Abstract

The recent surge of breakthroughs in machine learning and artificial intelligence has sparked renewed interest in large-scale stochastic optimization problems that are universally considered hard. One of the most widely used methods for solving such problems is distributed asynchronous stochastic gradient descent (DASGD), a family of algorithms that result from parallelizing stochastic gradient descent on distributed computing architectures (possibly) asychronously. However, a key obstacle in the efficient implementation of DASGD is the issue of delays : when a computing node contributes a gradient update, the global model parameter may have already been updated by other nodes several times over, thereby rendering this gradient information stale. These delays can quickly add up if the computational throughput of a node is saturated, so the convergence of DASGD may be compromised in the presence of large delays. Our first contribution is that, by carefully tuning the algorithm’s step size, convergence to the critical set is still achieved in mean square, even if the delays grow unbounded at a polynomial rate. We also establish finer results in a broad class of structured optimization problems (called variationally coherent), where we show that DASGD converges to a global optimum with a probability of one under the same delay assumptions. Together, these results contribute to the broad landscape of large-scale nonconvex stochastic optimization by offering state-of-the-art theoretical guarantees and providing insights for algorithm design.

Suggested Citation

  • Zhengyuan Zhou & Panayotis Mertikopoulos & Nicholas Bambos & Peter Glynn & Yinyu Ye, 2022. "Distributed Stochastic Optimization with Large Delays," Mathematics of Operations Research, INFORMS, vol. 47(3), pages 2082-2111, August.
  • Handle: RePEc:inm:ormoor:v:47:y:2022:i:3:p:2082-2111
    DOI: 10.1287/moor.2021.1200
    as

    Download full text from publisher

    File URL: http://dx.doi.org/10.1287/moor.2021.1200
    Download Restriction: no

    File URL: https://libkey.io/10.1287/moor.2021.1200?utm_source=ideas
    LibKey link: if access is restricted and if your library uses this service, LibKey will redirect you to where you can use your library subscription to access this item
    ---><---

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:inm:ormoor:v:47:y:2022:i:3:p:2082-2111. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: Chris Asher (email available below). General contact details of provider: https://edirc.repec.org/data/inforea.html .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.