Dynamic Inverse Reinforcement Learning for Characterizing Animal Behavior

被引:0
|
作者
Ashwood, Zoe C. [1 ,2 ]
Jha, Aditi [1 ,3 ]
Pillow, JonathanW. [1 ]
机构
[1] Princeton Univ, Princeton Neurosci Inst, Princeton, NJ 08544 USA
[2] Princeton Univ, Dept Comp Sci, Princeton, NJ 08544 USA
[3] Princeton Univ, Dept Elect & Comp Engn, Princeton, NJ USA
关键词
MODELS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Understanding decision-making is a core objective in both neuroscience and psychology, and computational models have often been helpful in the pursuit of this goal. While many models have been developed for characterizing behavior in binary decision-making and bandit tasks, comparatively little work has focused on animal decision-making in more complex tasks, such as navigation through a maze. Inverse reinforcement learning (IRL) is a promising approach for understanding such behavior, as it aims to infer the unknown reward function of an agent from its observed trajectories through state space. However, IRL has yet to be widely applied in neuroscience. One potential reason for this is that existing IRL frameworks assume that an agent's reward function is fixed over time. To address this shortcoming, we introduce dynamic inverse reinforcement learning (DIRL), a novel IRL framework that allows for time-varying intrinsic rewards. Our method parametrizes the unknown reward function as a time-varying linear combination of spatial reward maps (which we refer to as "goal maps"). We develop an efficient inference method for recovering this dynamic reward function from behavioral data. We demonstrate DIRL in simulated experiments and then apply it to a dataset of mice exploring a labyrinth. Our method returns interpretable reward functions for two separate cohorts of mice, and provides a novel characterization of exploratory behavior. We expect DIRL to have broad applicability in neuroscience, and to facilitate the design of biologically-inspired reward functions for training artificial agents.
引用
收藏
页数:14
相关论文
共 50 条
  • [11] Reinforcement Learning for Characterizing Hysteresis Behavior of Shape Memory Alloys
    Kirkpatrick, Kenton
    Valasek, John
    JOURNAL OF AEROSPACE COMPUTING INFORMATION AND COMMUNICATION, 2009, 6 (03): : 227 - 238
  • [12] Learning and Adapting Behavior of Autonomous Vehicles through Inverse Reinforcement Learning
    Trauth, Rainer
    Kaufeld, Marc
    Geisslinger, Maximilian
    Betz, Johannes
    2023 IEEE INTELLIGENT VEHICLES SYMPOSIUM, IV, 2023,
  • [13] Robust Bayesian Inverse Reinforcement Learning with Sparse Behavior Noise
    Zheng, Jiangchuan
    Liu, Siyuan
    Ni, Lionel M.
    PROCEEDINGS OF THE TWENTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2014, : 2198 - 2205
  • [14] Modeling Driver Behavior using Adversarial Inverse Reinforcement Learning
    Sackmann, Moritz
    Bey, Henrik
    Hofmann, Ulrich
    Thielecke, Joern
    2022 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2022, : 1683 - 1690
  • [15] Can AI predict animal movements? Filling gaps in animal trajectories using inverse reinforcement learning
    Hirakawa, Tsubasa
    Yamashita, Takayoshi
    Tamaki, Toru
    Fujiyoshi, Hironobu
    Umezu, Yuta
    Takeuchi, Ichiro
    Matsumoto, Sakiko
    Yoda, Ken
    ECOSPHERE, 2018, 9 (10):
  • [16] Projection based inverse reinforcement learning for the analysis of dynamic treatment regimes
    Syed Ihtesham Hussain Shah
    Giuseppe De Pietro
    Giovanni Paragliola
    Antonio Coronato
    Applied Intelligence, 2023, 53 : 14072 - 14084
  • [17] Inverse Reinforcement Learning for Identification in Linear-Quadratic Dynamic Games
    Koepf, Florian
    Inga, Jairo
    Rothfufss, Simon
    Flad, Michael
    Hohmann, Soeren
    IFAC PAPERSONLINE, 2017, 50 (01): : 14902 - 14908
  • [18] Projection based inverse reinforcement learning for the analysis of dynamic treatment regimes
    Shah, Syed Ihtesham Hussain
    De Pietro, Giuseppe
    Paragliola, Giovanni
    Coronato, Antonio
    APPLIED INTELLIGENCE, 2023, 53 (11) : 14072 - 14084
  • [19] Rationally inattentive inverse reinforcement learning explains youtube commenting behavior
    Hoiles, William
    Krishnamurthy, Vikram
    Pattanayak, Kunal
    Journal of Machine Learning Research, 2020, 21
  • [20] Inverse Reinforcement Learning via Neural Network in Driver Behavior Modeling
    Zou, QiJie
    Li, Haoyu
    Zhang, Rubo
    2018 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2018, : 1245 - 1250