A Decentralized Partially Observable Markov Decision Process for complete coverage onboard multiple shape changing reconfigurable robots

被引:0
|
作者
Pey, J. J. J. [1 ]
Bhagya, S. M. [1 ]
Samarakoon, P. [1 ]
Muthugala, M. A. Viraj J. [1 ]
Elara, Mohan Rajesh [1 ]
机构
[1] Singapore Univ Technol & Design, Engn Prod Dev Pillar, 8 Somapah Rd, Singapore 487372, Singapore
关键词
Deep reinforcement learning; Multi-agent; Reconfigurable robots; Decentralized Partially Observable Markov; Decision Process; Complete Coverage Planning; REINFORCEMENT; CHALLENGES; SYSTEMS;
D O I
10.1016/j.eswa.2025.126565
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Achieving complete area coverage by multiple collaborative robots is an essential aspect of productive deployment for applications such as cleaning, maintenance, and patrol. However, in real-world scenarios, physical constraints such as structural obstacles in the environment that generate narrow spaces, hinder the area coverage achieved by the fixed morphology robots, resulting in ineffective deployments. This paper proposes a Multi-Agent Reinforcement Learning (MARL) framework leveraging the Decentralized Partially Observable Markov Decision Process (Dec-POMDP) for Complete Coverage Planning (CPP) of multiple shapereconfigurable robots. This decentralized framework runs onboard each robot to mitigate the physical limitations while achieving complete area coverage. The proposed approach was trained on diverse map environments with different tight spaces to generate policies that enabled the robots to learn and coordinate their joint actions. The framework was further demonstrated on multiple unseen test environments to evaluate the generalization capabilities and area coverage performance. Additionally, baseline comparisons with the different combinations of multiple fixed morphology robots were conducted to validate the area coverage effectiveness of the proposed approach. In all scenarios, the proposed framework achieved 100% area coverage while the baselines only achieved partial area coverage ranging from 46.21% to 96.55%. The improved coverage performance compared to the baselines demonstrates the potential and usefulness of such an approach in the deployment of multiple reconfigurable robots.
引用
收藏
页数:13
相关论文
共 50 条
  • [21] A Partially Observable Markov-Decision-Process-Based Blackboard Architecture for Cognitive Agents in Partially Observable Environments
    Itoh, Hideaki
    Nakano, Hidehiko
    Tokushima, Ryota
    Fukumoto, Hisao
    Wakuya, Hiroshi
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2022, 14 (01) : 189 - 204
  • [22] Hidden Markov Model for Human Decision Process in a Partially Observable Environment
    Adomi, Masahiro
    Shikauchi, Yumi
    Ishii, Shin
    ARTIFICIAL NEURAL NETWORKS-ICANN 2010, PT II, 2010, 6353 : 94 - 103
  • [23] Voice Activity Detection Using Partially Observable Markov Decision Process
    Park, Chiyoun
    Kim, Namhoon
    Cho, Jeongmi
    INTERSPEECH 2009: 10TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2009, VOLS 1-5, 2009, : 2195 - 2198
  • [24] Tax Evasion as an Optimal Solution to a Partially Observable Markov Decision Process
    Papadopoulou, Paraskevi
    Hristu-Varsakelis, Dimitrios
    APPROXIMATION AND OPTIMIZATION: ALGORITHMS, COMPLEXITY AND APPLICATIONS, 2019, 145 : 219 - 237
  • [25] Parameter decision in adaptive partially observable Markov decision process with finite planning horizon
    Li, J.H.
    Han, Z.Z.
    Shanghai Jiaotong Daxue Xuebao/Journal of Shanghai Jiaotong University, 2000, 34 (12): : 1653 - 1657
  • [26] A generalised partially observable Markov decision process updated by decision trees for maintenance optimisation
    Faddoul, Rafic
    Raphael, Wassim
    Chateauneuf, Alaa
    STRUCTURE AND INFRASTRUCTURE ENGINEERING, 2011, 7 (10) : 783 - 796
  • [27] Optimizing Spatial and Temporal Reuse in Wireless Networks by Decentralized Partially Observable Markov Decision Processes
    Pajarinen, Joni
    Hottinen, Ari
    Peltonen, Jaakko
    IEEE TRANSACTIONS ON MOBILE COMPUTING, 2014, 13 (04) : 866 - 879
  • [28] Learning hierarchical partially observable Markov decision process models for robot navigation
    Theocharous, G
    Rohanimanesh, K
    Mahadevan, S
    2001 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, VOLS I-IV, PROCEEDINGS, 2001, : 511 - 516
  • [29] ON THE ADAPTIVE-CONTROL OF A PARTIALLY OBSERVABLE BINARY MARKOV DECISION-PROCESS
    FERNANDEZGAUCHERAND, E
    ARAPOSTATHIS, A
    MARCUS, SI
    LECTURE NOTES IN CONTROL AND INFORMATION SCIENCES, 1989, 130 : 217 - 229
  • [30] Underwater chemical plume tracing based on partially observable Markov decision process
    Jiu Hai-Feng
    Chen Yu
    Deng Wei
    Pang Shuo
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2019, 16 (02):