A Decentralized Partially Observable Markov Decision Process for complete coverage onboard multiple shape changing reconfigurable robots

被引:0
|
作者
Pey, J. J. J. [1 ]
Bhagya, S. M. [1 ]
Samarakoon, P. [1 ]
Muthugala, M. A. Viraj J. [1 ]
Elara, Mohan Rajesh [1 ]
机构
[1] Singapore Univ Technol & Design, Engn Prod Dev Pillar, 8 Somapah Rd, Singapore 487372, Singapore
关键词
Deep reinforcement learning; Multi-agent; Reconfigurable robots; Decentralized Partially Observable Markov; Decision Process; Complete Coverage Planning; REINFORCEMENT; CHALLENGES; SYSTEMS;
D O I
10.1016/j.eswa.2025.126565
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Achieving complete area coverage by multiple collaborative robots is an essential aspect of productive deployment for applications such as cleaning, maintenance, and patrol. However, in real-world scenarios, physical constraints such as structural obstacles in the environment that generate narrow spaces, hinder the area coverage achieved by the fixed morphology robots, resulting in ineffective deployments. This paper proposes a Multi-Agent Reinforcement Learning (MARL) framework leveraging the Decentralized Partially Observable Markov Decision Process (Dec-POMDP) for Complete Coverage Planning (CPP) of multiple shapereconfigurable robots. This decentralized framework runs onboard each robot to mitigate the physical limitations while achieving complete area coverage. The proposed approach was trained on diverse map environments with different tight spaces to generate policies that enabled the robots to learn and coordinate their joint actions. The framework was further demonstrated on multiple unseen test environments to evaluate the generalization capabilities and area coverage performance. Additionally, baseline comparisons with the different combinations of multiple fixed morphology robots were conducted to validate the area coverage effectiveness of the proposed approach. In all scenarios, the proposed framework achieved 100% area coverage while the baselines only achieved partial area coverage ranging from 46.21% to 96.55%. The improved coverage performance compared to the baselines demonstrates the potential and usefulness of such an approach in the deployment of multiple reconfigurable robots.
引用
收藏
页数:13
相关论文
共 50 条
  • [11] Mixed reinforcement learning for partially observable Markov decision process
    Dung, Le Tien
    Komeda, Takashi
    Takagi, Motoki
    2007 INTERNATIONAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE IN ROBOTICS AND AUTOMATION, 2007, : 436 - +
  • [12] Session search modeling by partially observable Markov decision process
    Yang, Grace Hui
    Dong, Xuchu
    Luo, Jiyun
    Zhang, Sicong
    INFORMATION RETRIEVAL JOURNAL, 2018, 21 (01): : 56 - 80
  • [13] Partially Observable Markov Decision Process Approximations for Adaptive Sensing
    Chong, Edwin K. P.
    Kreucher, Christopher M.
    Hero, Alfred O., III
    DISCRETE EVENT DYNAMIC SYSTEMS-THEORY AND APPLICATIONS, 2009, 19 (03): : 377 - 422
  • [14] A partially observable Markov decision process model for bridge inspection and maintenance
    Ellis, H
    Corotis, RB
    Jiang, MX
    APPLICATIONS OF STATISTICS AND PROBABILITY: CIVIL ENGINEERING RELIABILITY AND RISK ANALYSIS, VOL 3 - FINAL REPORT, 1996, : 1523 - 1528
  • [15] A PARTIALLY OBSERVABLE MARKOV DECISION-PROCESS WITH LAGGED INFORMATION - REPLY
    KIM, SH
    JOURNAL OF THE OPERATIONAL RESEARCH SOCIETY, 1988, 39 (02) : 217 - 218
  • [16] An efficient heuristic for a partially observable Markov decision process of machine replacement
    SinuanyStern, Z
    David, I
    Biran, S
    COMPUTERS & OPERATIONS RESEARCH, 1997, 24 (02) : 117 - 126
  • [17] Partially Observable Markov Decision Process for Managing Robot Collaboration with Human
    Karami, Abir-Beatrice
    Jeanpierre, Laurent
    Mouaddib, Abdel-Illah
    ICTAI: 2009 21ST INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, 2009, : 518 - 521
  • [18] HISTORY-DEPENDENT EVALUATIONS IN PARTIALLY OBSERVABLE MARKOV DECISION PROCESS
    Venel, Xavier
    Ziliotto, Bruno
    SIAM JOURNAL ON CONTROL AND OPTIMIZATION, 2021, 59 (02) : 1730 - 1755
  • [19] Partially Observable Markov Decision Process for Monitoring Multilayer Wafer Fabrication
    Khakifirooz, Marzieh
    Fathi, Mahdi
    Chien, Chen-Fu
    IEEE TRANSACTIONS ON AUTOMATION SCIENCE AND ENGINEERING, 2021, 18 (04) : 1742 - 1753
  • [20] Endangered Seabird Habitat Management as a Partially Observable Markov Decision Process
    Tomberlin, David
    MARINE RESOURCE ECONOMICS, 2010, 25 (01) : 93 - 104