Modeling Biological Agents Beyond the Reinforcement-Learning Paradigm

被引:5
|
作者
Georgeon, Olivier L. [1 ]
Casado, Remi C. [1 ]
Matignon, Laetitia A. [1 ]
机构
[1] Univ Lyon 1, LIRIS, UMR5205, F-69622 Villeurbanne, France
关键词
D O I
10.1016/j.procs.2015.12.179
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
It is widely acknowledged that biological beings (animals) are not Markov: modelers generally do not model them as agents receiving a complete representation of their environment's state in input (except perhaps in simple controlled tasks). In this paper, we claim that biological beings generally cannot recognize rewarding Markov states of their environment either. Therefore, we model them as agents trying to perform rewarding interactions with their environment (interaction-driven tasks), but not as agents trying to reach rewarding states (state-driven tasks). We review two interaction-driven tasks: the AB and AABB task, and implement a non-Markov Reinforcement-Learning (RL) algorithm based upon historical sequences and Q-learning. Results show that this RL algorithm takes significantly longer than a constructivist algorithm implemented previously by Georgeon, Ritter, & Haynes (2009). This is because the constructivist algorithm directly learns and repeats hierarchical sequences of interactions, whereas the RL algorithm spends time learning Q-values. Along with theoretical arguments, these results support the constructivist paradigm for modeling biological agents.
引用
收藏
页码:17 / 22
页数:6
相关论文
共 50 条
  • [41] A Reinforcement-Learning Algorithm for Sampling Design in Markov Random Fields
    Bonneau, Mathieu
    Peyrard, Nathalie
    Sabbadin, Regis
    20TH EUROPEAN CONFERENCE ON ARTIFICIAL INTELLIGENCE (ECAI 2012), 2012, 242 : 181 - 186
  • [42] Identifying optimal cycles in quantum thermal machines with reinforcement-learning
    Erdman, Paolo A.
    Noe, Frank
    NPJ QUANTUM INFORMATION, 2022, 8 (01)
  • [43] Evolving internal reinforcers for an intrinsically motivated reinforcement-learning robot
    Schembri, Massimiliano
    Mirolli, Marco
    Baldassarre, Gianluca
    2007 IEEE 6TH INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING, 2007, : 175 - 180
  • [44] Identifying optimal cycles in quantum thermal machines with reinforcement-learning
    Paolo A. Erdman
    Frank Noé
    npj Quantum Information, 8
  • [45] A Reinforcement-Learning Based Cognitive Scheme for Opportunistic Spectrum Access
    Angeliki V. Kordali
    Panayotis G. Cottis
    Wireless Personal Communications, 2016, 86 : 751 - 769
  • [46] Reinforcement-learning generation of four-qubit entangled states
    Giordano, Sara
    Martin-Delgado, Miguel A.
    PHYSICAL REVIEW RESEARCH, 2022, 4 (04):
  • [47] Computational modeling of behavioral tasks: An illustration on a classic reinforcement learning paradigm
    Suthaharan, Praveen
    Corlett, R. Philip
    Ang, Yuen-Siang
    QUANTITATIVE METHODS FOR PSYCHOLOGY, 2021, 17 (02): : 105 - 140
  • [48] GBDT Modeling of Deep Reinforcement Learning Agents Using Distillation
    Hatano, Toshiki
    Tsuneda, Toi
    Suzuki, Yuta
    Imade, Kuniyasu
    Shesimo, Kazuki
    Yamane, Satoshi
    2021 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS (ICM), 2021,
  • [49] High-intensity interval exercise impairs neuroelectric indices of reinforcement-learning
    Walsh, Jeremy J.
    Colino, Francisco L.
    Krigolson, Olave E.
    Luehr, Stephen
    Gurd, Brendon J.
    Tschakovsky, Michael E.
    PHYSIOLOGY & BEHAVIOR, 2019, 198 : 18 - 26
  • [50] A Reinforcement-Learning Application to Biodiversity Conservation in Costa-Rican Forest
    Sabbadin, R.
    Spring, D.
    Bergonnier, E.
    MODSIM 2007: INTERNATIONAL CONGRESS ON MODELLING AND SIMULATION: LAND, WATER AND ENVIRONMENTAL MANAGEMENT: INTEGRATED SYSTEMS FOR SUSTAINABILITY, 2007, : 2189 - 2195