IDEAS home Printed from https://ideas.repec.org/a/gam/jftint/v14y2022i12p348-d981029.html
   My bibliography  Save this article

rl4dtn: Q-Learning for Opportunistic Networks

Author

Listed:
  • Jorge Visca

    (Faculty of Engineering, Universidad de la República, Montevideo 11600, Uruguay)

  • Javier Baliosian

    (Faculty of Engineering, Universidad de la República, Montevideo 11600, Uruguay)

Abstract

Opportunistic networks are highly stochastic networks supported by sporadic encounters between mobile devices. To route data efficiently, opportunistic-routing algorithms must capitalize on devices’ movement and data transmission patterns. This work proposes a routing method based on reinforcement learning, specifically Q-learning. As usual in routing algorithms, the objective is to select the best candidate devices to put forward once an encounter occurs. However, there is also the possibility of not forwarding if we know that a better candidate might be encountered in the future. This decision is not usually considered in learning schemes because there is no obvious way to represent the temporal evolution of the network. We propose a novel, distributed, and online method that allows learning both the network’s connectivity and its temporal evolution with the help of a temporal graph. This algorithm allows learning to skip forwarding opportunities to capitalize on future encounters. We show that explicitly representing the action for deferring forwarding increases the algorithm’s performance. The algorithm’s scalability is discussed and shown to perform well in a network of considerable size.

Suggested Citation

  • Jorge Visca & Javier Baliosian, 2022. "rl4dtn: Q-Learning for Opportunistic Networks," Future Internet, MDPI, vol. 14(12), pages 1-17, November.
  • Handle: RePEc:gam:jftint:v:14:y:2022:i:12:p:348-:d:981029
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/1999-5903/14/12/348/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/1999-5903/14/12/348/
    Download Restriction: no
    ---><---

    References listed on IDEAS

    as
    1. Huizhen Yu & Dimitri Bertsekas, 2013. "Q-learning and policy iteration algorithms for stochastic shortest path problems," Annals of Operations Research, Springer, vol. 208(1), pages 95-132, September.
    2. Vishnupriya Kuppusamy & Udaya Miriya Thanthrige & Asanga Udugama & Anna Förster, 2019. "Evaluating Forwarding Protocols in Opportunistic Networks: Trends, Advances, Challenges and Best Practices," Future Internet, MDPI, vol. 11(5), pages 1-26, May.
    Full references (including those not matched with items on IDEAS)

    Most related items

    These are the items that most often cite the same works as this one and are cited by the same works as this one.
    1. Mishra, Neermegha & Kastner, Marvin & Jahn, Carlos, 2022. "Marine communication for shipping: Using ad-hoc networks at sea," Chapters from the Proceedings of the Hamburg International Conference of Logistics (HICL), in: Kersten, Wolfgang & Jahn, Carlos & Blecker, Thorsten & Ringle, Christian M. (ed.), Changing Tides: The New Role of Resilience and Sustainability in Logistics and Supply Chain Management – Innovative Approaches for the Shift to a New , volume 33, pages 523-557, Hamburg University of Technology (TUHH), Institute of Business Logistics and General Management.
    2. Dimitri P. Bertsekas, 2019. "Robust shortest path planning and semicontractive dynamic programming," Naval Research Logistics (NRL), John Wiley & Sons, vol. 66(1), pages 15-37, February.
    3. Dimitri P. Bertsekas, 2018. "Proximal algorithms and temporal difference methods for solving fixed point problems," Computational Optimization and Applications, Springer, vol. 70(3), pages 709-736, July.
    4. Huizhen Yu & Dimitri P. Bertsekas, 2015. "A Mixed Value and Policy Iteration Method for Stochastic Control with Universally Measurable Policies," Mathematics of Operations Research, INFORMS, vol. 40(4), pages 926-968, October.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jftint:v:14:y:2022:i:12:p:348-:d:981029. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    If CitEc recognized a bibliographic reference but did not link an item in RePEc to it, you can help with this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.