Author
Listed:
- Yujing Zhou
(Department of Electrical Engineering and Computer Science, Embry-Riddle Aeronautical University, Daytona Beach Campus, Daytona Beach, FL 32114, USA)
- Yupeng Yang
(College of Engineering, University of North Carolina at Charlotte, Charlotte, NC 28223, USA)
- Bill Deng Pan
(School of Graduate Studies in College of Aviation, Embry-Riddle Aeronautical University, Daytona Beach Campus, Daytona Beach, FL 32114, USA)
- Yongxin Liu
(Department of Mathematics in College of Arts and Sciences, Embry-Riddle Aeronautical University, Daytona Beach Campus, Daytona Beach, FL 32114, USA)
- Sirish Namilae
(Department of Aerospace Engineering, Embry-Riddle Aeronautical University, Daytona Beach Campus, Daytona Beach, FL 32114, USA)
- Houbing Herbert Song
(College of Engineering and Information Technology, University of Maryland, Baltimore County, Baltimore, MD 21250, USA)
- Dahai Liu
(School of Graduate Studies in College of Aviation, Embry-Riddle Aeronautical University, Daytona Beach Campus, Daytona Beach, FL 32114, USA)
Abstract
Emergencies can occur unexpectedly and require immediate action, especially in aviation, where time pressure and uncertainty are high. This study focused on improving emergency evacuation in airport and aircraft scenarios using real-time decision-making support. A system based on the Asynchronous Advantage Actor–Critic (A3C) algorithm, an advanced deep reinforcement learning method, was developed to generate faster and more efficient evacuation routes compared to traditional models. The A3C model was tested in various scenarios, including different environmental conditions and numbers of agents, and its performance was compared with the Deep Q-Network (DQN) algorithm. The results showed that A3C achieved evacuations 43.86% faster on average and converged in fewer episodes (100 vs. 250 for DQN). In dynamic environments with moving threats, A3C also outperformed DQN in maintaining agent safety and adapting routes in real time. As the number of agents increased, A3C maintained high levels of efficiency and robustness. These findings demonstrate A3C’s strong potential to enhance evacuation planning through improved speed, adaptability, and scalability. The study concludes by highlighting the practical benefits of applying such models in real-world emergency response systems, including significantly faster evacuation times, real-time adaptability to evolving threats, and enhanced scalability for managing large crowds in high-density environments including airport terminals. The A3C-based model offers a cost-effective alternative to full-scale evacuation drills by enabling virtual scenario testing, supports proactive safety planning through predictive modeling, and contributes to the development of intelligent decision-support tools that improve coordination and reduce response time during emergencies.
Suggested Citation
Yujing Zhou & Yupeng Yang & Bill Deng Pan & Yongxin Liu & Sirish Namilae & Houbing Herbert Song & Dahai Liu, 2025.
"Deep Reinforcement Learning for Real-Time Airport Emergency Evacuation Using Asynchronous Advantage Actor–Critic (A3C) Algorithm,"
Mathematics, MDPI, vol. 13(14), pages 1-22, July.
Handle:
RePEc:gam:jmathe:v:13:y:2025:i:14:p:2269-:d:1701636
Download full text from publisher
Corrections
All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:13:y:2025:i:14:p:2269-:d:1701636. See general information about how to correct material in RePEc.
If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.
We have no bibliographic references for this item. You can help adding them by using this form .
If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.
For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .
Please note that corrections may take a couple of weeks to filter through
the various RePEc services.