Fuzzy Rule Interpolation-based Q-learning

被引:0
|
作者
Vincze, David [1 ]
Kovacs, Szilveszter [1 ]
机构
[1] Univ Miskolc, Dept Informat Technol, Miskolc, Hungary
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is a well known topic in computational intelligence. It can be used to solve control problems in unknown environments without defining an exact method on how to solve problems in various situations. Instead the goal is defined and all the actions done in the different states are given feedback, called reward or punishment (positive or negative reward). Based on these rewards the system can learn which action is considered the best in a given state. A method called Q-learning can be used for building up the state-action-value function. This method uses discrete states. With the application of fuzzy reasoning the method can be extended to be used in continuous environment, called Fuzzy Q-learning (FQ-Learning). Traditional Fuzzy Q-learning uses 0-order Takagi-Sugeno fuzzy inference. The main goal of this paper is to introduce Fuzzy Rule Interpolation (FRI), namely the FIVE (Fuzzy rule Interpolation based on Vague Environment) to be the model applied with Q-learning (FRIQ-learning). The paper also includes an application example: the well known cart pole (reversed pendulum) problem is used for demonstrating the applicability of the FIVE model in Q-learning.
引用
收藏
页码:45 / 49
页数:5
相关论文
共 50 条
  • [31] Improving interpolation-based oversampling for imbalanced data learning
    Zhu, Tuanfei
    Lin, Yaping
    Liu, Yonghe
    KNOWLEDGE-BASED SYSTEMS, 2020, 187
  • [32] Multi-Agent Coordination Method Based on Fuzzy Q-Learning
    Peng, Jun
    Liu, Miao
    Wu, Min
    Zhang, Xiaoyong
    Lin, Kuo-Chi
    2008 7TH WORLD CONGRESS ON INTELLIGENT CONTROL AND AUTOMATION, VOLS 1-23, 2008, : 5411 - +
  • [33] Genetic Based Fuzzy Q-Learning Energy Management for Smart Grid
    Li Xin
    Zang Chuanzhi
    Zeng Peng
    Yu Haibin
    PROCEEDINGS OF THE 31ST CHINESE CONTROL CONFERENCE, 2012, : 6924 - 6927
  • [34] Fuzzy interporation-based Q-learning with continuous states and actions
    Horiuchi, T
    Fujino, A
    Katai, O
    Sawaragi, T
    FUZZ-IEEE '96 - PROCEEDINGS OF THE FIFTH IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS, VOLS 1-3, 1996, : 594 - 600
  • [35] Footstep planning for biped robot based on fuzzy Q-learning approach
    Sabourin, Christophe
    Madani, Kurosh
    Yu, Weiwei
    Yan, Jie
    ICINCO 2008: PROCEEDINGS OF THE FIFTH INTERNATIONAL CONFERENCE ON INFORMATICS IN CONTROL, AUTOMATION AND ROBOTICS, VOL RA-1: ROBOTICS AND AUTOMATION, VOL 1, 2008, : 183 - +
  • [36] Fuzzy Q-learning with the modified fuzzy ART neural network
    Ueda, H
    Hanada, N
    Kimoto, H
    Naraki, T
    Takahashi, K
    Miyahara, T
    2005 IEEE/WIC/ACM INTERNATIONAL CONFERENCE ON INTELLIGENT AGENT TECHNOLOGY, PROCEEDINGS, 2005, : 308 - 315
  • [37] Fuzzy Q-Learning with the modified fuzzy ART neural network
    Ueda, Hiroaki
    Naraki, Takeshi
    Hanada, Naoki
    Kimoto, Hideaki
    Takahashi, Kenichi
    Miyahara, Tetsuhiro
    Web Intelligence and Agent Systems, 2007, 5 (03): : 331 - 341
  • [38] An Online Learning Control Strategy for Hybrid Electric Vehicle Based on Fuzzy Q-Learning
    Hu, Yue
    Li, Weimin
    Xu, Hui
    Xu, Guoqing
    ENERGIES, 2015, 8 (10): : 11167 - 11186
  • [39] Fuzzy adaptive Q-learning method with dynamic learning parameters
    Maeda, Y
    JOINT 9TH IFSA WORLD CONGRESS AND 20TH NAFIPS INTERNATIONAL CONFERENCE, PROCEEDINGS, VOLS. 1-5, 2001, : 2778 - 2780
  • [40] Modeling and fuzzy Q-learning control of biped walking
    Meng Joo Er
    Yi Zhou
    Proceedings of the 24th Chinese Control Conference, Vols 1 and 2, 2005, : 641 - 646