TD based reinforcement learning using neural networks in control problems with continuous action space

被引:0
|
作者
Lee, JH [1 ]
Oh, SY [1 ]
Choi, DH [1 ]
机构
[1] Pohang Univ Sci & Technol, Dept Elect Engn, Pohang 790784, South Korea
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
While most of the research on reinforcement learning assumed a discrete control space, many of the real world control problems need to have continuous output This can be achieved by using continuous mapping functions for the value and action functions of the reinforcement learning architecture. Two questions arise here however One is what sort of function representation to use and the other is how to determine the amount of noise for search in action space. The ubiquitous back-propagation neural network is used here to learn the value and action functions. Next the reinforcement predictor that is intended to predict the next reinforcement is introduced that also determines the amount of noise to add to the controller output This proposed Reinforcement Learning architecture is found to have a sound on-line learning control performance through a computer simulation of the ball and beam system as an example plant.
引用
收藏
页码:2028 / 2033
页数:6
相关论文
共 50 条
  • [1] Convergent Reinforcement Learning Control with Neural Networks and Continuous Action Search
    Lee, Minwoo
    Anderson, Charles W.
    2014 IEEE SYMPOSIUM ON ADAPTIVE DYNAMIC PROGRAMMING AND REINFORCEMENT LEARNING (ADPRL), 2014, : 33 - 40
  • [2] Fuzzy neural control of satellite attitude by TD based reinforcement learning
    Cui, Xiao-ting
    Liu, Xiang-dong
    WCICA 2006: SIXTH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-12, CONFERENCE PROCEEDINGS, 2006, : 3983 - +
  • [3] Swarm Reinforcement Learning Methods for Problems with Continuous State-Action Space
    Iima, Hitoshi
    Kuroe, Yasuaki
    Emoto, Kazuo
    2011 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2011, : 2173 - 2180
  • [4] Switching reinforcement learning for continuous action space
    Nagayoshi, Masato
    Murao, Hajime
    Tamaki, Hisashi
    ELECTRONICS AND COMMUNICATIONS IN JAPAN, 2012, 95 (03) : 37 - 44
  • [5] Quantum reinforcement learning in continuous action space
    Wu, Shaojun
    Jin, Shan
    Wen, Dingding
    Han, Donghong
    Wang, Xiaoting
    QUANTUM, 2025, 9 : 1 - 18
  • [6] Algorithmic trading using continuous action space deep reinforcement learning
    Majidi, Naseh
    Shamsi, Mahdi
    Marvasti, Farokh
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 235
  • [7] Vision-based Navigation of UAV with Continuous Action Space Using Deep Reinforcement Learning
    Zhou, Benchun
    Wang, Weihong
    Liu, Zhenghua
    Wang, Jia
    PROCEEDINGS OF THE 2019 31ST CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2019), 2019, : 5030 - 5035
  • [8] Autonomous blimp control using model-free reinforcement learning in a continuous state and action space
    Rottmann, Axel
    Plagemann, Christian
    Hilgers, Peter
    Burgard, Wolfram
    2007 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-9, 2007, : 1901 - +
  • [9] Personalized vital signs control based on continuous action-space reinforcement learning with supervised experience
    Sun, Chenxi
    Hong, Shenda
    Song, Moxian
    Shang, Junyuan
    Li, Hongyan
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2021, 69
  • [10] Control of associative chaotic neural networks using a reinforcement learning
    Sato, N
    Adachi, M
    Kotani, M
    ADVANCES IN NEURAL NETWORKS - ISNN 2004, PT 1, 2004, 3173 : 395 - 400