IDEAS home Printed from https://ideas.repec.org/a/gam/jmathe/v10y2022i18p3293-d912158.html
   My bibliography  Save this article

Deep Reinforcement Learning-Based RMSA Policy Distillation for Elastic Optical Networks

Author

Listed:
  • Bixia Tang

    (School of Physics and Telecommunication Engineering, South China Normal University, Guangzhou 510006, China)

  • Yue-Cai Huang

    (School of Electronics and Information Engineering, South China Normal University, Foshan 528200, China)

  • Yun Xue

    (School of Physics and Telecommunication Engineering, South China Normal University, Guangzhou 510006, China
    School of Electronics and Information Engineering, South China Normal University, Foshan 528200, China)

  • Weixing Zhou

    (School of Physics and Telecommunication Engineering, South China Normal University, Guangzhou 510006, China
    School of Electronics and Information Engineering, South China Normal University, Foshan 528200, China)

Abstract

The reinforcement learning-based routing, modulation, and spectrum assignment has been regarded as an emerging paradigm for resource allocation in the elastic optical networks. One limitation is that the learning process is highly dependent on the training environment, such as the traffic pattern or the optical network topology. Therefore, re-training is required in case of network topology or traffic pattern variations, which consumes a great amount of computation power and time. To ease the requirement of re-training, we propose a policy distillation scheme, which distills knowledge from a well-trained teacher model and then transfers the knowledge to the to-be-trained student model, so that the training of the latter can be accelerated. Specifically, the teacher model is trained for one training environment (e.g., the topology and traffic pattern) and the student model is for another training environment. The simulation results indicate that our proposed method can effectively speed up the training process of the student model, and it even leads to a lower blocking probability, compared with the case that the student model is trained without knowledge distillation.

Suggested Citation

  • Bixia Tang & Yue-Cai Huang & Yun Xue & Weixing Zhou, 2022. "Deep Reinforcement Learning-Based RMSA Policy Distillation for Elastic Optical Networks," Mathematics, MDPI, vol. 10(18), pages 1-19, September.
  • Handle: RePEc:gam:jmathe:v:10:y:2022:i:18:p:3293-:d:912158
    as

    Download full text from publisher

    File URL: https://www.mdpi.com/2227-7390/10/18/3293/pdf
    Download Restriction: no

    File URL: https://www.mdpi.com/2227-7390/10/18/3293/
    Download Restriction: no
    ---><---

    Citations

    Citations are extracted by the CitEc Project, subscribe to its RSS feed for this item.
    as


    Cited by:

    1. Weihua Ou & Jianping Gou & Shaoning Zeng & Lan Du, 2023. "Preface to the Special Issue “Advancement of Mathematical Methods in Feature Representation Learning for Artificial Intelligence, Data Mining and Robotics”—Special Issue Book," Mathematics, MDPI, vol. 11(4), pages 1-4, February.

    Corrections

    All material on this site has been provided by the respective publishers and authors. You can help correct errors and omissions. When requesting a correction, please mention this item's handle: RePEc:gam:jmathe:v:10:y:2022:i:18:p:3293-:d:912158. See general information about how to correct material in RePEc.

    If you have authored this item and are not yet registered with RePEc, we encourage you to do it here. This allows to link your profile to this item. It also allows you to accept potential citations to this item that we are uncertain about.

    We have no bibliographic references for this item. You can help adding them by using this form .

    If you know of missing items citing this one, you can help us creating those links by adding the relevant references in the same way as above, for each refering item. If you are a registered author of this item, you may also want to check the "citations" tab in your RePEc Author Service profile, as there may be some citations waiting for confirmation.

    For technical questions regarding this item, or to correct its authors, title, abstract, bibliographic or download information, contact: MDPI Indexing Manager (email available below). General contact details of provider: https://www.mdpi.com .

    Please note that corrections may take a couple of weeks to filter through the various RePEc services.

    IDEAS is a RePEc service. RePEc uses bibliographic data supplied by the respective publishers.