A Lightweight Reinforcement Learning Based Packet Routing Method Using Online Sequential Learning

被引:1
|
作者
Nemoto, Kenji [1 ]
Matsutani, Hiroki [1 ]
机构
[1] Keio Univ, Grad Sch Sci & Technol, Yokohama 2238522, Japan
关键词
reinforcement learning; packet routing; neural networks; OS-ELM;
D O I
10.1587/transinf.2022EDP7231
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Existing simple routing protocols (e.g., OSPF, RIP) have some disadvantages of being inflexible and prone to congestion due to the concentration of packets on particular routers. To address these issues, packet routing methods using machine learning have been proposed recently. Compared to these algorithms, machine learning based methods can choose a routing path intelligently by learning efficient routes. However, machine learning based methods have a disadvantage of training time overhead. We thus focus on a lightweight machine learning algorithm, OS-ELM (Online Sequential Extreme Learning Machine), to reduce the training time. Although previous work on reinforcement learning using OS-ELM exists, it has a problem of low learning accuracy. In this paper, we propose OS-ELM QN (Q-Network) with a prioritized experience replay buffer to improve the learning performance. It is compared to a deep reinforcement learning based packet routing method using a network simulator. Experimental results show that introducing the experience replay buffer improves the learning performance. OS-ELM QN achieves a 2.33 times speedup than a DQN (Deep Q-Network) in terms of learning speed. Regarding the packet transfer latency, OS-ELM QN is comparable or slightly inferior to the DQN while they are better than OSPF in most cases since they can distribute congestions.
引用
收藏
页码:1796 / 1807
页数:12
相关论文
共 50 条
  • [21] Edge Collaborative Online Task Offloading Method Based on Reinforcement Learning
    Sun, Ming
    Bao, Tie
    Xie, Dan
    Lv, Hengyi
    Si, Guoliang
    ELECTRONICS, 2023, 12 (18)
  • [22] An Efficient Method for Solving Routing Problems with Energy Constraints Using Reinforcement Learning
    Do, Haggi
    Son, Hakmo
    Kim, Jinwhan
    2024 21ST INTERNATIONAL CONFERENCE ON UBIQUITOUS ROBOTS, UR 2024, 2024, : 293 - 298
  • [23] An Online Rapid Mesh Segmentation Method Based on an Online Sequential Extreme Learning Machine
    Zhao, Feiyu
    Sheng, Buyun
    Yin, Xiyan
    Wang, Hui
    Lu, Xincheng
    Zhao, Yuncheng
    IEEE ACCESS, 2019, 7 : 109094 - 109110
  • [24] Reinforcement Learning Using a Stochastic Gradient Method with Memory-Based Learning
    Yamada, Takafumi
    Yamaguchi, Satoshi
    ELECTRICAL ENGINEERING IN JAPAN, 2010, 173 (01) : 32 - 40
  • [25] Autonomous Lightweight Scheduling in LoRa-based Networks Using Reinforcement Learning
    Baimukhanov, Batyrkhan
    Gilazh, Bibarys
    Zorbas, Dimitrios
    2024 IEEE INTERNATIONAL BLACK SEA CONFERENCE ON COMMUNICATIONS AND NETWORKING, BLACKSEACOM 2024, 2024, : 268 - 271
  • [26] Tank War Using Online Reinforcement Learning
    Andersen, Kresten Toftgaard
    Zeng, Yifeng
    Christensen, Dennis Dahl
    Tran, Dung
    2009 IEEE/WIC/ACM INTERNATIONAL JOINT CONFERENCES ON WEB INTELLIGENCE (WI) AND INTELLIGENT AGENT TECHNOLOGIES (IAT), VOL 2, 2009, : 497 - 500
  • [27] Online Vehicle Routing With Neural Combinatorial Optimization and Deep Reinforcement Learning
    Yu, James J. Q.
    Yu, Wen
    Gu, Jiatao
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2019, 20 (10) : 3806 - 3817
  • [28] Reinforcement learning strategies for sequential action learning
    Fermin, Alan
    Takehiko, Yoshida
    Tanaka, Saori
    Ito, Makoto
    Yoshimoto, Junichiro
    Doya, Kenji
    NEUROSCIENCE RESEARCH, 2009, 65 : S236 - S236
  • [29] Reinforcement learning for adaptive routing
    Peshkin, L
    Savova, V
    PROCEEDING OF THE 2002 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, VOLS 1-3, 2002, : 1825 - 1830
  • [30] Using Deep Reinforcement Learning for Routing in IP Networks
    Singh, Abhiram
    Sharma, Sidharth
    Gumaste, Ashwin
    30TH INTERNATIONAL CONFERENCE ON COMPUTER COMMUNICATIONS AND NETWORKS (ICCCN 2021), 2021,