IDEAS home Printed from https://ideas.repec.org/a/gam/jmathe/v13y2025i18p3050-d1754849.html
   My bibliography  Save this article

A Deep Reinforcement Learning Model to Solve the Stochastic Capacitated Vehicle Routing Problem with Service Times and Deadlines

Author

Listed:
  • Sergio Flavio Marroquín-Cano

    (Tecnológico Nacional de México, Instituto Tecnológico de Tuxtla Gutierrez, Carr. Panamericana Km 1080, Tuxtla Gutierrez 29050, Mexico)

  • Elías Neftalí Escobar-Gómez

    (Tecnológico Nacional de México, Instituto Tecnológico de Tuxtla Gutierrez, Carr. Panamericana Km 1080, Tuxtla Gutierrez 29050, Mexico)

  • Eduardo F. Morales

    (Instituto Nacional de Astrofísica, Óptica y Electrónica (INAOE), Luis Enrique Erro No. 1, San Andrés Cholula 72840, Mexico)

  • Elizeth Ramírez-Álvarez

    (Tecnológico Nacional de México, Instituto Tecnológico de Querétaro, Avenida Tecnológico S/N, Santiago de Queretaro 76000, Mexico)

  • Pedro Gasga-García

    (Tecnológico Nacional de México, Instituto Tecnológico de Tuxtla Gutierrez, Carr. Panamericana Km 1080, Tuxtla Gutierrez 29050, Mexico)

  • Eduardo Chandomí-Castellanos

    (Tecnológico Nacional de México, Instituto Tecnológico de Tuxtla Gutierrez, Carr. Panamericana Km 1080, Tuxtla Gutierrez 29050, Mexico)

  • J. Renán Velázquez-González

    (Tecnológico Nacional de México, Instituto Tecnológico de Tuxtla Gutierrez, Carr. Panamericana Km 1080, Tuxtla Gutierrez 29050, Mexico)

  • Julio Alberto Guzmán-Rabasa

    (Departamento de Mecatrónica, Universidad Politécnica de Chiapas, Carr. Tuxtla Gutierrez Km 21+500, Suchiapa 29082, Mexico)

  • José Roberto Bermúdez

    (Tecnológico Nacional de México, Instituto Tecnológico de Tuxtla Gutierrez, Carr. Panamericana Km 1080, Tuxtla Gutierrez 29050, Mexico)

  • Francisco Rodríguez-Sánchez

    (Tecnológico Nacional de México, Instituto Tecnológico de Tuxtla Gutierrez, Carr. Panamericana Km 1080, Tuxtla Gutierrez 29050, Mexico)

Abstract

Vehicle Routing Problems are central to logistics and operational research, arising in diverse contexts such as transportation planning, manufacturing systems, and military operations. While Deep Reinforcement Learning has been successfully applied to both deterministic and stochastic variants of Vehicle Routing Problems, existing approaches often neglect critical time-sensitive conditions. This work addresses the Stochastic Capacitated Vehicle Routing Problem with Service Times and Deadlines, a challenging formulation that is suited to model time routing conditions. The proposal, POMO-DC, integrates a novel dynamic context mechanism. At each decision step, this mechanism incorporates the vehicle’s cumulative travel time and delays—features absent in prior models—enabling the policy to adapt to changing conditions and avoid time violations. The model is evaluated on stochastic instances with 20, 30, and 50 customers and benchmarked against Google OR-Tools using multiple metaheuristics. Results show that POMO-DC reduces average delays by up to 88% (from 169.63 to 20.35 min for instances of 30 customers) and 75% (from 4352.43 to 1098.97 min for instances of 50 customers), while maintaining competitive travel times. These outcomes highlight the potential of Deep Reinforcement Learning-based frameworks to learn patterns from stochastic data and effectively manage time uncertainty in Vehicle Routing Problems.

Suggested Citation

  • Sergio Flavio Marroquín-Cano & Elías Neftalí Escobar-Gómez & Eduardo F. Morales & Elizeth Ramírez-Álvarez & Pedro Gasga-García & Eduardo Chandomí-Castellanos & J. Renán Velázquez-González & Julio Albe, 2025. "A Deep Reinforcement Learning Model to Solve the Stochastic Capacitated Vehicle Routing Problem with Service Times and Deadlines," Mathematics, MDPI, vol. 13(18), pages 1-25, September.
  • Handle: RePEc:gam:jmathe:v:13:y:2025:i:18:p:3050-:d:1754849
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2227-7390/13/18/3050/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2227-7390/13/18/3050/
    Download Restriction: no
    ---><---

    More about this item

    Keywords

    ;
    ;
    ;
    ;
    ;

    Statistics

    Access and download statistics

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:13:y:2025:i:18:p:3050-:d:1754849. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.