Fuzzy Rule Interpolation-based Q-learning

被引:0
|
作者
Vincze, David [1 ]
Kovacs, Szilveszter [1 ]
机构
[1] Univ Miskolc, Dept Informat Technol, Miskolc, Hungary
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is a well known topic in computational intelligence. It can be used to solve control problems in unknown environments without defining an exact method on how to solve problems in various situations. Instead the goal is defined and all the actions done in the different states are given feedback, called reward or punishment (positive or negative reward). Based on these rewards the system can learn which action is considered the best in a given state. A method called Q-learning can be used for building up the state-action-value function. This method uses discrete states. With the application of fuzzy reasoning the method can be extended to be used in continuous environment, called Fuzzy Q-learning (FQ-Learning). Traditional Fuzzy Q-learning uses 0-order Takagi-Sugeno fuzzy inference. The main goal of this paper is to introduce Fuzzy Rule Interpolation (FRI), namely the FIVE (Fuzzy rule Interpolation based on Vague Environment) to be the model applied with Q-learning (FRIQ-learning). The paper also includes an application example: the well known cart pole (reversed pendulum) problem is used for demonstrating the applicability of the FIVE model in Q-learning.
引用
收藏
页码:45 / 49
页数:5
相关论文
共 50 条
  • [1] Incremental Rule Base Creation with Fuzzy Rule Interpolation-Based Q-Learning
    Vincze, David
    Kovacs, Szilveszter
    COMPUTATIONAL INTELLIGENCE IN ENGINEERING, 2010, 313 : 191 - 203
  • [2] Fuzzy interpolation-based Q-learning with profit sharing plan scheme
    Horiuchi, T
    Fujino, A
    Katai, O
    Sawaragi, T
    PROCEEDINGS OF THE SIXTH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS I - III, 1997, : 1707 - 1712
  • [3] Demonstration of expert knowledge injection in Fuzzy Rule Interpolation based Q-learning
    Tompa, Tama
    Kovacs, Szilveszter
    Vincze, David
    Niitsuma, Mihoko
    2021 IEEE/SICE INTERNATIONAL SYMPOSIUM ON SYSTEM INTEGRATION (SII), 2021, : 843 - 844
  • [4] Antecedent Redundancy Exploitation in Fuzzy Rule Interpolation-based Reinforcement Learning
    Vincze, David
    Toth, Alex
    Niitsuma, Mihoko
    2020 IEEE/ASME INTERNATIONAL CONFERENCE ON ADVANCED INTELLIGENT MECHATRONICS (AIM), 2020, : 1316 - 1321
  • [5] Football Simulation Modeling with Fuzzy Rule Interpolation-based Fuzzy Automaton
    Vincze, David
    Toth, Alex
    Niitsuma, Mihoko
    2020 17TH INTERNATIONAL CONFERENCE ON UBIQUITOUS ROBOTS (UR), 2020, : 87 - 92
  • [6] Fuzzy Q-learning
    Glorennec, PY
    Jouffe, L
    PROCEEDINGS OF THE SIXTH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS I - III, 1997, : 659 - 662
  • [7] Fuzzy Rule Interpolation-based Behavior Modeling for Human-Robot Interaction
    Kovacs, Szilveszter
    2014 IEEE 12TH INTERNATIONAL SYMPOSIUM ON APPLIED MACHINE INTELLIGENCE AND INFORMATICS (SAMI), 2014, : 115 - 115
  • [8] Accuracy based fuzzy Q-learning for robot behaviours
    Gu, DB
    Hu, HS
    2004 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-3, PROCEEDINGS, 2004, : 1455 - 1460
  • [9] Dynamic scheduling with fuzzy clustering based Q-learning
    Wang, Guo-Lei
    Lin, Lin
    Zhong, Shi-Sheng
    Jisuanji Jicheng Zhizao Xitong/Computer Integrated Manufacturing Systems, CIMS, 2009, 15 (04): : 751 - 757
  • [10] Fuzzy Q-Learning for generalization of reinforcement learning
    Berenji, HR
    FUZZ-IEEE '96 - PROCEEDINGS OF THE FIFTH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-3, 1996, : 2208 - 2214