Federated Discrete Reinforcement Learning for Automatic Guided Vehicle Control

被引:2
|
作者
Sierra-Garcia, J. Enrique [1 ]
Santos, Matilde [2 ]
机构
[1] Univ Burgos, Electromech Engn Dept, Burgos 09006, Spain
[2] Univ Complutense Madrid, Inst Knowledge Technol, Madrid 28040, Spain
关键词
Automated guided vehicle (AGV); Federated learning; Industry; 4.0; Intelligent control; Path following; Reinforcement learning; AGV;
D O I
10.1016/j.future.2023.08.021
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Under the federated learning paradigm, the agents learn in parallel and combine their knowledge to build a global knowledge model. This new machine learning strategy increases privacy and reduces communication costs, some benefits that can be very useful for industry applications deployed in the edge. Automatic Guided Vehicles (AGVs) can take advantage of this approach since they can be considered intelligent agents, operate in fleets, and are normally managed by a central system that can run in the edge and handles the knowledge of each of them to obtain a global emerging behavioral model. Furthermore, this idea can be combined with the concept of reinforcement learning (RL). This way, the AGVs can interact with the system to learn according to the policy implemented by the RL algorithm in order to follow specified routes, and send their findings to the main system. The centralized system collects this information in a group policy to turn it over to the AGVs. In this work, a novel Federated Discrete Reinforcement Learning (FDRL) approach is implemented to control the trajectories of a fleet of AGVs. Each industrial AGV runs the modules that correspond to an RL system: a state estimator, a rewards calculator, an action selector, and a policy update algorithm. AGVs share their policy variation with the federated server, which combines them into a group policy with a learning aggregation function. To validate the proposal, simulation results of the FDRL control for five hybrid tricycle-differential AGVs and four different trajectories (ellipse, lemniscate, octagon, and a closed 16-polyline) have been obtained and compared with a Proportional Integral Derivative (PID) controller optimized with genetic algorithms. The intelligent control approach shows an average improvement of 78% in mean absolute error, 75% in root mean square error, and 73% in terms of standard deviation. It has been shown that this approach also accelerates the learning up to a 50 % depending on the trajectory, with an average of 36% speed up while allowing precise tracking. The suggested federated-learning based technique outperforms an optimized fuzzy logic controller (FLC) for all of the measured trajectories as well. In addition, different learning aggregation functions have been proposed and evaluated. The influence of the number of vehicles (from 2 to 10) on the path following performance and on network transmission has been analyzed too.& COPY; 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND license (http://creativecommons.org/licenses/by-nc-nd/4.0/).
引用
收藏
页码:78 / 89
页数:12
相关论文
共 50 条
  • [41] Unmanned Aerial Vehicle Pitch Control Using Deep Reinforcement Learning with Discrete Actions in Wind Tunnel Test
    Wada, Daichi
    Araujo-Estrada, Sergio A.
    Windsor, Shane
    AEROSPACE, 2021, 8 (01) : 1 - 16
  • [42] Federated Reinforcement Learning For Fast Personalization
    Nadiger, Chetan
    Kumar, Anil
    Abdelhak, Sherine
    2019 IEEE SECOND INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND KNOWLEDGE ENGINEERING (AIKE), 2019, : 123 - 127
  • [43] Federated Reinforcement Learning with Environment Heterogeneity
    Jin, Hao
    Peng, Yang
    Yang, Wenhao
    Wang, Shusen
    Zhang, Zhihua
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151 : 18 - 37
  • [44] Federated Reinforcement Learning for the Building Facilities
    Fujita, Koki
    Fujimura, Shugo
    Sun, Yuwei
    Esaki, Hiroshi
    Ochiai, Hideya
    2022 IEEE INTERNATIONAL CONFERENCE ON OMNI-LAYER INTELLIGENT SYSTEMS (IEEE COINS 2022), 2022, : 331 - 336
  • [45] FERED: Federated Reinforcement Learning in the DBMS
    Tzamaras, Sotirios
    Ciucanu, Radu
    Soare, Marta
    Amer-Yahia, Sihem
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 4989 - 4993
  • [46] Distributed Deep Reinforcement Learning-Based Gradient Quantization for Federated Learning Enabled Vehicle Edge Computing
    Zhang, Cui
    Zhang, Wenjun
    Wu, Qiong
    Fan, Pingyi
    Fan, Qiang
    Wang, Jiangzhou
    Letaief, Khaled B.
    IEEE INTERNET OF THINGS JOURNAL, 2025, 12 (05): : 4899 - 4913
  • [47] Traveling course interpolation method for automatic guided vehicle systems with discrete reference markers
    Kimura, M
    Oguchi, K
    ISIE '97 - PROCEEDINGS OF THE IEEE INTERNATIONAL SYMPOSIUM ON INDUSTRIAL ELECTRONICS, VOLS 1-3, 1997, : 12 - 17
  • [48] Federated Reinforcement Learning Based on Multi-head Attention Mechanism for Vehicle Edge Caching
    Li, XinRan
    Wei, ZhenChun
    Lyu, ZengWei
    Yuan, XiaoHui
    Xu, Juan
    Zhang, ZeYu
    WIRELESS ALGORITHMS, SYSTEMS, AND APPLICATIONS, PT III, 2022, 13473 : 648 - 656
  • [49] Lyapunov-guided Deep Reinforcement Learning for Vehicle task Stable offloading
    Huang, Ziyang
    Chen, Yanming
    Zhang, Yiwen
    PROCEEDINGS OF THE 2024 27 TH INTERNATIONAL CONFERENCE ON COMPUTER SUPPORTED COOPERATIVE WORK IN DESIGN, CSCWD 2024, 2024, : 833 - 838
  • [50] Study on the Guidance Control System of Inductively Powered Automatic Guided Vehicle
    Liu Wei
    Sun Yne
    Su Yugang
    Dai Xin
    Qin Yunyou
    Xia Chenyang
    25TH WORLD BATTERY, HYBRID AND FUEL CELL ELECTRIC VEHICLE SYMPOSIUM AND EXHIBITION PROCEEDINGS, VOLS 1 & 2, 2010, : 1349 - 1353