Fuzzy Rule Interpolation-based Q-learning

被引:0
|
作者
Vincze, David [1 ]
Kovacs, Szilveszter [1 ]
机构
[1] Univ Miskolc, Dept Informat Technol, Miskolc, Hungary
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Reinforcement learning is a well known topic in computational intelligence. It can be used to solve control problems in unknown environments without defining an exact method on how to solve problems in various situations. Instead the goal is defined and all the actions done in the different states are given feedback, called reward or punishment (positive or negative reward). Based on these rewards the system can learn which action is considered the best in a given state. A method called Q-learning can be used for building up the state-action-value function. This method uses discrete states. With the application of fuzzy reasoning the method can be extended to be used in continuous environment, called Fuzzy Q-learning (FQ-Learning). Traditional Fuzzy Q-learning uses 0-order Takagi-Sugeno fuzzy inference. The main goal of this paper is to introduce Fuzzy Rule Interpolation (FRI), namely the FIVE (Fuzzy rule Interpolation based on Vague Environment) to be the model applied with Q-learning (FRIQ-learning). The paper also includes an application example: the well known cart pole (reversed pendulum) problem is used for demonstrating the applicability of the FIVE model in Q-learning.
引用
收藏
页码:45 / 49
页数:5
相关论文
共 50 条
  • [41] A Hybrid Fuzzy Q-Learning algorithm for robot navigation
    Gordon, Sean W.
    Reyes, Napoleon H.
    Barczak, Andre
    2011 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2011, : 2625 - 2631
  • [42] Intelligent Fuzzy Q-Learning control of humanoid robots
    Er, MJ
    Zhou, Y
    ADVANCES IN NEURAL NETWORKS - ISNN 2005, PT 3, PROCEEDINGS, 2005, 3498 : 216 - 221
  • [43] Anomaly Detection using Fuzzy Q-learning Algorithm
    Shamshirband, Shahaboddin
    Anuar, Nor Badrul
    Kiah, Miss Laiha Mat
    Misra, Sanjay
    ACTA POLYTECHNICA HUNGARICA, 2014, 11 (08) : 5 - 28
  • [44] A Q-learning based continuous tuning of fuzzy wall tracking without exploration
    Ghaderi, R. (r_ghaderi@nit.ac.ir), 1600, Materials and Energy Research Center (25):
  • [45] Automatic generation of fuzzy inference systems by dynamic fuzzy Q-Learning
    Deng, C
    Er, MJ
    2003 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN AND CYBERNETICS, VOLS 1-5, CONFERENCE PROCEEDINGS, 2003, : 3206 - 3211
  • [46] Routing in VANETs: A Fuzzy Constraint Q-Learning Approach
    Wu, Celimuge
    Ohzahata, Satoshi
    Kato, Toshihiko
    2012 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM), 2012, : 195 - 200
  • [47] A Q-learning Based Continuous Tuning of Fuzzy Wall Tracking without Exploration
    Valiollahi, S.
    Ghaderi, R.
    Ebrahimzadeh, A.
    INTERNATIONAL JOURNAL OF ENGINEERING, 2012, 25 (04): : 355 - 366
  • [48] Fuzzy Q-learning in continuous state and action space
    Xu M.-L.
    Xu W.-B.
    Journal of China Universities of Posts and Telecommunications, 2010, 17 (04): : 100 - 109
  • [50] Design of a fuzzy logic controller with Evolutionary Q-Learning
    Kim, Min-Soeng
    Lee, Ju-Jang
    INTELLIGENT AUTOMATION AND SOFT COMPUTING, 2006, 12 (04): : 369 - 381