A Decentralized Partially Observable Markov Decision Process for complete coverage onboard multiple shape changing reconfigurable robots

被引:0
|
作者
Pey, J. J. J. [1 ]
Bhagya, S. M. [1 ]
Samarakoon, P. [1 ]
Muthugala, M. A. Viraj J. [1 ]
Elara, Mohan Rajesh [1 ]
机构
[1] Singapore Univ Technol & Design, Engn Prod Dev Pillar, 8 Somapah Rd, Singapore 487372, Singapore
关键词
Deep reinforcement learning; Multi-agent; Reconfigurable robots; Decentralized Partially Observable Markov; Decision Process; Complete Coverage Planning; REINFORCEMENT; CHALLENGES; SYSTEMS;
D O I
10.1016/j.eswa.2025.126565
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Achieving complete area coverage by multiple collaborative robots is an essential aspect of productive deployment for applications such as cleaning, maintenance, and patrol. However, in real-world scenarios, physical constraints such as structural obstacles in the environment that generate narrow spaces, hinder the area coverage achieved by the fixed morphology robots, resulting in ineffective deployments. This paper proposes a Multi-Agent Reinforcement Learning (MARL) framework leveraging the Decentralized Partially Observable Markov Decision Process (Dec-POMDP) for Complete Coverage Planning (CPP) of multiple shapereconfigurable robots. This decentralized framework runs onboard each robot to mitigate the physical limitations while achieving complete area coverage. The proposed approach was trained on diverse map environments with different tight spaces to generate policies that enabled the robots to learn and coordinate their joint actions. The framework was further demonstrated on multiple unseen test environments to evaluate the generalization capabilities and area coverage performance. Additionally, baseline comparisons with the different combinations of multiple fixed morphology robots were conducted to validate the area coverage effectiveness of the proposed approach. In all scenarios, the proposed framework achieved 100% area coverage while the baselines only achieved partial area coverage ranging from 46.21% to 96.55%. The improved coverage performance compared to the baselines demonstrates the potential and usefulness of such an approach in the deployment of multiple reconfigurable robots.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Decentralized Control of Partially Observable Markov Decision Processes
    Amato, Christopher
    Chowdhary, Girish
    Geramifard, Alborz
    Uere, N. Kemal
    Kochenderfer, Mykel J.
    2013 IEEE 52ND ANNUAL CONFERENCE ON DECISION AND CONTROL (CDC), 2013, : 2398 - 2405
  • [2] A partially-observable markov decision process for dealing with dynamically changing environments
    Chatzis, Sotirios P.
    Kosmopoulos, Dimitrios
    IFIP Advances in Information and Communication Technology, 2014, 436 : 111 - 120
  • [3] Robust partially observable Markov decision process
    Osogami, Takayuki
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 37, 2015, 37 : 106 - 115
  • [4] Autonomous Thermalling as a Partially Observable Markov Decision Process
    Guilliard, Iain
    Rogahn, Richard J.
    Piavis, Jim
    Kolobov, Andrey
    ROBOTICS: SCIENCE AND SYSTEMS XIV, 2018,
  • [5] Fuzzy Reinforcement Learning Control for Decentralized Partially Observable Markov Decision Processes
    Sharma, Rajneesh
    Spaan, Matthijs T. J.
    IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS (FUZZ 2011), 2011, : 1422 - 1429
  • [6] Actual Causality and Responsibility Attribution in Decentralized Partially Observable Markov Decision Processes
    Triantafyllou, Stelios
    Singla, Adish
    Radanovic, Goran
    PROCEEDINGS OF THE 2022 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, AIES 2022, 2022, : 739 - 752
  • [7] Motion Planning of Mobile Robots in Indoor Topological Environments using Partially Observable Markov Decision Process
    Monteiro, Neemias Silva
    Goncalves, Vinicius Mariano
    Maia, Carlos Andrey
    IEEE LATIN AMERICA TRANSACTIONS, 2021, 19 (08) : 1315 - 1324
  • [8] Partially Observable Markov Decision Process Approximations for Adaptive Sensing
    Edwin K. P. Chong
    Christopher M. Kreucher
    Alfred O. Hero
    Discrete Event Dynamic Systems, 2009, 19 : 377 - 422
  • [9] A PARTIALLY OBSERVABLE MARKOV DECISION-PROCESS WITH LAGGED INFORMATION
    KIM, SH
    JEONG, BH
    JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 1987, 38 (05) : 439 - 446
  • [10] Session search modeling by partially observable Markov decision process
    Grace Hui Yang
    Xuchu Dong
    Jiyun Luo
    Sicong Zhang
    Information Retrieval Journal, 2018, 21 : 56 - 80