IDEAS home Printed from https://ideas.repec.org/a/gam/jsusta/v15y2023i14p11051-d1194446.html
   My bibliography  Save this article

Deep Reinforcement Q-Learning for Intelligent Traffic Control in Mass Transit

Author

Listed:
  • Shurok Khozam

    (Cosys-Grettia, University Gustave Eiffel, F-77447 Marne-la-Vallée, France)

  • Nadir Farhi

    (Cosys-Grettia, University Gustave Eiffel, F-77447 Marne-la-Vallée, France)

Abstract

Traffic control in mass transit consists of the regulation of both vehicle dynamics and passenger flows. While most of the existing approaches focus on the optimization of vehicle dwell time, vehicle time headway, and passenger stocks, we propose in this article an approach which also includes the optimization of the passenger inflows to the platforms. We developed in this work a deep reinforcement Q-learning model for the traffic control in a mass transit line. We first propose a new mathematical traffic model for the train and passengers dynamics. The model combines a discrete-event description of the vehicle dynamics, with a macroscopic model for the passenger flows. We use this new model as the environment of the traffic in mass transit for the reinforcement learning optimization. For this aim, we defined, under the new traffic model, the state variables as well as the control ones, including in particular the number of running vehicles, the vehicle dwell times at stations, and the passenger inflow to platforms. Second, we present our new deep Q-network (DQN) model for the reinforcement learning (RL) with the state representation, action space, and reward function definitions. We also provide the neural network architecture as well as the main hyper-parameters. Finally, we give an evaluation of the model under multiple scenarios. We show in particular the efficiency of the control of the passenger inflows into the platforms.

Suggested Citation

  • Shurok Khozam & Nadir Farhi, 2023. "Deep Reinforcement Q-Learning for Intelligent Traffic Control in Mass Transit," Sustainability, MDPI, vol. 15(14), pages 1-22, July.
  • Handle: RePEc:gam:jsusta:v:15:y:2023:i:14:p:11051-:d:1194446
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2071-1050/15/14/11051/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2071-1050/15/14/11051/
    Download Restriction: no
    ---><---

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jsusta:v:15:y:2023:i:14:p:11051-:d:1194446. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.