IDEAS home Printed from https://ideas.repec.org/a/gam/jsusta/v14y2022i14p8825-d866301.html
   My bibliography  Save this article

An Adaptive Multi-Level Quantization-Based Reinforcement Learning Model for Enhancing UAV Landing on Moving Targets

Author

Listed:
  • Najmaddin Abo Mosali

    (Research Center for Unmanned Vehicles, Faculty of Mechanical and Manufacturing Engineering, Universiti Tun Hussein Onn Malaysia, Parit Raja, Batu Pahat 86400, Johor, Malaysia)

  • Syariful Syafiq Shamsudin

    (Research Center for Unmanned Vehicles, Faculty of Mechanical and Manufacturing Engineering, Universiti Tun Hussein Onn Malaysia, Parit Raja, Batu Pahat 86400, Johor, Malaysia)

  • Salama A. Mostafa

    (Faculty of Computer Science and Information Technology, Universiti Tun Hussin Onn Malaysia, Parit Raja, Batu Pahat 84600, Johor, Malaysia)

  • Omar Alfandi

    (College of Technological Innovation, Zayed University, Abu Dhabi 4783, United Arab Emirates)

  • Rosli Omar

    (Faculty of Electrical and Electronic Engineering, Universiti Tun Hussein Onn Malaysia, Parit Raja, Batu Pahat 86400, Johor, Malaysia)

  • Najib Al-Fadhali

    (Faculty of Electrical and Electronic Engineering, Universiti Tun Hussein Onn Malaysia, Parit Raja, Batu Pahat 86400, Johor, Malaysia)

  • Mazin Abed Mohammed

    (College of Computer Science and Information Technology, University of Anbar, Ramadi 31001, Iraq)

  • R. Q. Malik

    (Department of Medical Instrumentation Techniques Engineering, Al-Mustaqbal University College, Hillah 51001, Iraq)

  • Mustafa Musa Jaber

    (Department of Medical Instruments Engineering Techniques, Dijlah University College, Baghdad 10021, Iraq
    Department of Medical Instruments Engineering Techniques, Al-Farahidi University, Baghdad 10021, Iraq)

  • Abdu Saif

    (Department of Electrical Engineering, Faculty of Engineering, Universiti Malaya, Kuala Lumpur 50603, Selangor, Malaysia)

Abstract

The autonomous landing of an unmanned aerial vehicle (UAV) on a moving platform is an essential functionality in various UAV-based applications. It can be added to a teleoperation UAV system or part of an autonomous UAV control system. Various robust and predictive control systems based on the traditional control theory are used for operating a UAV. Recently, some attempts were made to land a UAV on a moving target using reinforcement learning (RL). Vision is used as a typical way of sensing and detecting the moving target. Mainly, the related works have deployed a deep-neural network (DNN) for RL, which takes the image as input and provides the optimal navigation action as output. However, the delay of the multi-layer topology of the deep neural network affects the real-time aspect of such control. This paper proposes an adaptive multi-level quantization-based reinforcement learning (AMLQ) model. The AMLQ model quantizes the continuous actions and states to directly incorporate simple Q-learning to resolve the delay issue. This solution makes the training faster and enables simple knowledge representation without needing the DNN. For evaluation, the AMLQ model was compared with state-of-art approaches and was found to be superior in terms of root mean square error (RMSE), which was 8.7052 compared with the proportional–integral–derivative (PID) controller, which achieved an RMSE of 10.0592.

Suggested Citation

  • Najmaddin Abo Mosali & Syariful Syafiq Shamsudin & Salama A. Mostafa & Omar Alfandi & Rosli Omar & Najib Al-Fadhali & Mazin Abed Mohammed & R. Q. Malik & Mustafa Musa Jaber & Abdu Saif, 2022. "An Adaptive Multi-Level Quantization-Based Reinforcement Learning Model for Enhancing UAV Landing on Moving Targets," Sustainability, MDPI, vol. 14(14), pages 1-17, July.
  • Handle: RePEc:gam:jsusta:v:14:y:2022:i:14:p:8825-:d:866301
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2071-1050/14/14/8825/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2071-1050/14/14/8825/
    Download Restriction: no
    ---><---

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jsusta:v:14:y:2022:i:14:p:8825-:d:866301. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.