Data-Efficient Offline Reinforcement Learning with Approximate Symmetries

被引:0
|
作者
Angelotti, Giorgio [1 ,2 ]
Drougard, Nicolas [1 ,2 ]
Chanel, Caroline P. C. [1 ,2 ]
机构
[1] Univ Toulouse, ANITI, Toulouse, France
[2] Univ Toulouse, ISAE Supaero, Toulouse, France
关键词
Offline reinforcement learning; Approximate symmetries; Data augmentation;
D O I
10.1007/978-3-031-55326-4_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The performance of Offline Reinforcement Learning (ORL) models in Markov Decision Processes (MDPs) is heavily contingent upon the quality and diversity of the training data. This research furthers the exploration of expert-guided symmetry detection and data augmentation techniques by considering approximate symmetries in discrete MDPs, providing a fresh perspective on data efficiency in the domain of ORL. We scrutinize the adaptability and resilience of these established methodologies in varied stochastic environments, featuring alterations in transition probabilities with respect to the already tested stochastic environments. Key findings from these investigations elucidate the potential of approximate symmetries for the data augmentation process and confirm the robustness of the existing methods under altered stochastic conditions. Our analysis reinforces the applicability of the established symmetry detection techniques in diverse scenarios while opening new horizons for enhancing the efficiency of ORL models.
引用
收藏
页码:164 / 186
页数:23
相关论文
共 50 条
  • [31] Data-Efficient Off-Policy Policy Evaluation for Reinforcement Learning
    Thomas, Philip S.
    Brunskill, Emma
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [32] Data-Efficient Graph Learning
    Ding, Kaize
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 20, 2024, : 22663 - 22663
  • [33] On Efficient Sampling in Offline Reinforcement Learning
    Jia, Qing-Shan
    2024 14TH ASIAN CONTROL CONFERENCE, ASCC 2024, 2024, : 1 - 6
  • [34] Data-efficient Co-Adaptation of Morphology and Behaviour with Deep Reinforcement Learning
    Luck, Kevin Sebastian
    Ben Amor, Heni
    Calandra, Roberto
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [35] Mix-up Consistent Cross Representations for Data-Efficient Reinforcement Learning
    Liu, Shiyu
    Cao, Guitao
    Liu, Yong
    Li, Yan
    Wu, Chunwei
    Xi, Xidong
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [36] Data-efficient deep reinforcement learning with expert demonstration for active flow control
    Zheng, Changdong
    Xie, Fangfang
    Ji, Tingwei
    Zhang, Xinshuai
    Lu, Yufeng
    Zhou, Hongjie
    Zheng, Yao
    PHYSICS OF FLUIDS, 2022, 34 (11)
  • [37] Data-Efficient Reinforcement Learning for Energy Optimization of Power-Assisted Wheelchairs
    Feng, Guoxi
    Busoniu, Lucian
    Guerra, Thierry-Marie
    Mohammad, Sami
    IEEE TRANSACTIONS ON INDUSTRIAL ELECTRONICS, 2019, 66 (12) : 9734 - 9744
  • [38] Robust On-Policy Sampling for Data-Efficient Policy Evaluation in Reinforcement Learning
    Zhong, Rujie
    Zhang, Duohan
    Schafer, Lukas
    Albrecht, Stefano V.
    Hanna, Josiah P.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [39] Load Balancing for Communication Networks via Data-Efficient Deep Reinforcement Learning
    Wu, Di
    Kang, Jikun
    Xu, Yi Tian
    Li, Hang
    Li, Jimmy
    Chen, Xi
    Rivkin, Dmitriy
    Jenkin, Michael
    Lee, Taeseop
    Park, Intaik
    Liu, Xue
    Dudek, Gregory
    2021 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2021,
  • [40] SDRL: Interpretable and Data-Efficient Deep Reinforcement Learning Leveraging Symbolic Planning
    Lyu, Daoming
    Yang, Fangkai
    Liu, Bo
    Gustafson, Steven
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 2970 - 2977