Analysis of Q-learning on ANNs for Robot Control using Live Video Feed

被引:0
|
作者
Murali, Nihal [1 ]
Gupta, Kunal [1 ]
Bhanot, Surekha [1 ]
机构
[1] BITS Pilani, Dept Elect & Elect Engn, Pilani Campus, Pilani 333031, Rajasthan, India
关键词
Artificial neural networks; Hardware implementation; Q-learning; Raw image inputs; Reinforcement learning; Robot learning;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Training of artificial neural networks (ANNs) using reinforcement learning (RL) techniques is being widely discussed in the robot learning literature. The high model complexity of ANNs along with the model-free nature of RL algorithms provides a desirable combination for many robotics applications. There is a huge need for algorithms that generalize using raw sensory inputs, such as vision, without any hand-engineered features or domain heuristics. In this paper, the standard control problem of line following robot was used as a test-bed, and an ANN controller for the robot was trained on images from a live video feed using Q-learning. A virtual agent was first trained in simulation environment and then deployed onto a robot's hardware. The robot successfully learns to traverse a wide range of curves and displays excellent generalization ability. Qualitative analysis of the evolution of policies, performance and weights of the network provide insights into the nature and convergence of the learning algorithm.
引用
收藏
页码:524 / 529
页数:6
相关论文
共 50 条
  • [21] Learning to control an inverted pendulum using Q-learning and neural networks
    Jiang, Guofei
    Wu, Cangpu
    Zidonghua Xuebao/Acta Automatica Sinica, 1998, 24 (05): : 662 - 666
  • [22] Enhanced Robot Learning using Fuzzy Q-Learning & Context-Aware Middleware
    Phiri, Charles C.
    Ju, Zhaojie
    Kubota, Naoyuki
    Liu, Honghai
    2016 INTERNATIONAL SYMPOSIUM ON MICRO-NANOMECHATRONICS AND HUMAN SCIENCE (MHS), 2016,
  • [23] Enhancing Nash Q-learning and Team Q-learning mechanisms by using bottlenecks
    Ghazanfari, Behzad
    Mozayani, Nasser
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2014, 26 (06) : 2771 - 2783
  • [24] LEARNING HOSE TRANSPORT CONTROL WITH Q-LEARNING
    Fernandez-Gauna, Borja
    Manuel Lopez-Guede, Jose
    Zulueta, Ekaitz
    Grana, Manuel
    NEURAL NETWORK WORLD, 2010, 20 (07) : 913 - 923
  • [25] Autonomous Decentralized Traffic Control Using Q-Learning in LPWAN
    Kaburaki, Aoto
    Adachi, Koichi
    Takyu, Osamu
    Ohta, Mai
    Fujii, Takeo
    IEEE ACCESS, 2021, 9 : 93651 - 93661
  • [26] Faster Deep Q-learning using Neural Episodic Control
    Nishio, Daichi
    Yamane, Satoshi
    2018 IEEE 42ND ANNUAL COMPUTER SOFTWARE AND APPLICATIONS CONFERENCE (COMPSAC), VOL 1, 2018, : 486 - 491
  • [27] Optimal control using adaptive resonance theory and Q-learning
    Kiumarsi, Bahare
    AlQaudi, Bakur
    Modares, Hamidreza
    Lewis, Frank L.
    Levine, Daniel S.
    NEUROCOMPUTING, 2019, 361 : 119 - 125
  • [28] MULTI-ROBOT COOPERATIVE TRANSPORTATION OF OBJECTS USING MODIFIED Q-LEARNING
    Siriwardana, Pallege Gamini Dilupa
    de Silva, Clarence
    PROCEEDINGS OF THE ASME INTERNATIONAL MECHANICAL ENGINEERING CONGRESS AND EXPOSITION - 2010, VOL 8, PTS A AND B, 2012, : 745 - 753
  • [29] Solving the optimal path planning of a mobile robot using improved Q-learning
    Low, Ee Soong
    Ong, Pauline
    Cheah, Kah Chun
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2019, 115 : 143 - 161
  • [30] Automating Robot Motion Planning for Magnetic Resonance Navigation Using Q-Learning
    Wang, Xiaoyan
    An, Zhenzhou
    Zhou, Yihang
    Wang, Haifeng
    Chang, Yuchou
    2018 IEEE INTERNATIONAL CONFERENCE ON CYBORG AND BIONIC SYSTEMS (CBS), 2018, : 304 - 307