A comparison of learning performance in two-dimensional Q-learning by the difference of Q-values alignment

被引:0
|
作者
Kathy Thi Aung
Takayasu Fuchida
机构
[1] Kagoshima University,Department of Information and Computer Science, Graduate School of Science and Engineering
关键词
Q-learning; Q-value; Voronoi Q-value element; Single agent; State space;
D O I
10.1007/s10015-011-0961-5
中图分类号
学科分类号
摘要
In this article, we examine the learning performance of various strategies under different conditions using the Voronoi Q-value element (VQE) based on reward in a single-agent environment, and decide how to act in a certain state. In order to test our hypotheses, we performed computational experiments using several situations such as various angles of rotation of VQEs which are arranged into a lattice structure, various angles of an agent’s action rotation that has 4 actions, and a random arrangement of VQEs to correctly evaluate the optimal Q-values for state and action pairs in order to deal with continuous-valued inputs. As a result, the learning performance changes when the angle of VQEs and the angle of action are changed by a specific relative position.
引用
收藏
页码:473 / 477
页数:4
相关论文
共 50 条
  • [1] A comparison of learning performance in two-dimensional Q-learning by the difference of Q-values alignment
    Aung, Kathy Thi
    Fuchida, Takayasu
    PROCEEDINGS OF THE SIXTEENTH INTERNATIONAL SYMPOSIUM ON ARTIFICIAL LIFE AND ROBOTICS (AROB 16TH '11), 2011, : 597 - 600
  • [2] A comparison of learning performance in two-dimensional Q-learning by the difference of Q-values alignment
    Aung, Kathy Thi
    Fuchida, Takayasu
    ARTIFICIAL LIFE AND ROBOTICS, 2012, 16 (04) : 473 - 477
  • [3] Improving the Performance of Q-learning Using Simultanouse Q-values Updating
    Pouyan, Maryam
    Mousavi, Amin
    Golzari, Shahram
    Hatam, Ahmad
    2014 INTERNATIONAL CONGRESS ON TECHNOLOGY, COMMUNICATION AND KNOWLEDGE (ICTCK), 2014,
  • [4] A proposition of addition and integration of q-values in Q-Learning
    Aung, Kathy Thi
    Fuchida, Takayasu
    PROCEEDINGS OF THE EIGHTEENTH INTERNATIONAL SYMPOSIUM ON ARTIFICIAL LIFE AND ROBOTICS (AROB 18TH '13), 2013, : 304 - 307
  • [5] Initialization of Q-values by fuzzy rules for accelerating Q-learning
    Oh, CH
    Nakashima, T
    Ishibuchi, H
    IEEE WORLD CONGRESS ON COMPUTATIONAL INTELLIGENCE, 1998, : 2051 - 2056
  • [6] Q-learning: computation of optimal Q-values for evaluating the learning level in robotic tasks
    D'Orazio, T
    Cicirelli, G
    JOURNAL OF EXPERIMENTAL & THEORETICAL ARTIFICIAL INTELLIGENCE, 2001, 13 (03) : 241 - 270
  • [7] Emotion behavior learning system based on meta-parameter control of Q-learning with plural Q-values
    Akiguchi, S
    Maeda, Y
    PROCEEDINGS OF THE 3RD INTERNATIONAL SYMPOSIUM ON AUTONOMOUS MINIROBOTS FOR RESEARCH AND EDUTAINMENT (AMIRE 2005), 2006, : 344 - +
  • [8] Two mode Q-learning
    Park, KH
    Kim, JH
    CEC: 2003 CONGRESS ON EVOLUTIONARY COMPUTATION, VOLS 1-4, PROCEEDINGS, 2003, : 2449 - 2454
  • [9] Comparison of Deep Q-Learning, Q-Learning and SARSA Reinforced Learning for Robot Local Navigation
    Anas, Hafiq
    Ong, Wee Hong
    Malik, Owais Ahmed
    ROBOT INTELLIGENCE TECHNOLOGY AND APPLICATIONS 6, 2022, 429 : 443 - 454
  • [10] Learning to Play Pac-Xon with Q-Learning and Two Double Q-Learning Variants
    Schilperoort, Jits
    Mak, Ivar
    Drugan, Madalina M.
    Wiering, Marco A.
    2018 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2018, : 1151 - 1158