Consolidated actor-critic model for partially-observable Markov decision processes

被引:0
|
作者
Elhanany, I. [1 ]
Niedzwiedz, C. [1 ]
Liu, Z.
Livingston, S. [1 ]
机构
[1] Univ Tennessee, Dept Elect Engn & Comp Sci, Knoxville, TN 37996 USA
关键词
D O I
10.1049/el:20081346
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
A method for consolidating the traditionally separate actor and critic neural networks in temporal difference learning for addressing partially-observable Markov decision processes (POMDPs) is presented. Simulation results for solving a five-state POMDP problem support the claim that the consolidated model achieves higher performance while reducing computational and storage requirements to approximately half those of the traditional approach.
引用
收藏
页码:1317 / U41
页数:2
相关论文
共 50 条
  • [41] Equivalence Relations in Fully and Partially Observable Markov Decision Processes
    Castro, Pablo Samuel
    Panangaden, Prakash
    Precup, Doina
    21ST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE (IJCAI-09), PROCEEDINGS, 2009, : 1653 - 1658
  • [42] Recursively-Constrained Partially Observable Markov Decision Processes
    Ho, Qi Heng
    Becker, Tyler
    Kraske, Benjamin
    Laouar, Zakariya
    Feather, Martin S.
    Rossi, Federico
    Lahijanian, Morteza
    Sunberg, Zachary
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2024, 244 : 1658 - 1680
  • [43] A Fast Approximation Method for Partially Observable Markov Decision Processes
    LIU Bingbing
    KANG Yu
    JIANG Xiaofeng
    QIN Jiahu
    JournalofSystemsScience&Complexity, 2018, 31 (06) : 1423 - 1436
  • [44] Active Chemical Sensing With Partially Observable Markov Decision Processes
    Gosangi, Rakesh
    Gutierrez-Osuna, Ricardo
    OLFACTION AND ELECTRONIC NOSE, PROCEEDINGS, 2009, 1137 : 562 - 565
  • [45] Stochastic optimization of controlled partially observable Markov decision processes
    Bartlett, PL
    Baxter, J
    PROCEEDINGS OF THE 39TH IEEE CONFERENCE ON DECISION AND CONTROL, VOLS 1-5, 2000, : 124 - 129
  • [46] Actor-Critic Model Predictive Control
    Romero, Angel
    Song, Yunlong
    Scaramuzza, Davide
    2024 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2024), 2024, : 14777 - 14784
  • [47] Reinforcement learning algorithm for partially observable Markov decision processes
    Wang, Xue-Ning
    He, Han-Gen
    Xu, Xin
    Kongzhi yu Juece/Control and Decision, 2004, 19 (11): : 1263 - 1266
  • [48] Partially Observable Markov Decision Processes and Performance Sensitivity Analysis
    Li, Yanjie
    Yin, Baoqun
    Xi, Hongsheng
    IEEE TRANSACTIONS ON SYSTEMS MAN AND CYBERNETICS PART B-CYBERNETICS, 2008, 38 (06): : 1645 - 1651
  • [49] Learning factored representations for partially observable Markov decision processes
    Sallans, B
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 12, 2000, 12 : 1050 - 1056
  • [50] Partially Observable Markov Decision Processes: A Geometric Technique and Analysis
    Zhang, Hao
    OPERATIONS RESEARCH, 2010, 58 (01) : 214 - 228