Dynamic Inverse Reinforcement Learning for Characterizing Animal Behavior

被引:0
|
作者
Ashwood, Zoe C. [1 ,2 ]
Jha, Aditi [1 ,3 ]
Pillow, JonathanW. [1 ]
机构
[1] Princeton Univ, Princeton Neurosci Inst, Princeton, NJ 08544 USA
[2] Princeton Univ, Dept Comp Sci, Princeton, NJ 08544 USA
[3] Princeton Univ, Dept Elect & Comp Engn, Princeton, NJ USA
关键词
MODELS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Understanding decision-making is a core objective in both neuroscience and psychology, and computational models have often been helpful in the pursuit of this goal. While many models have been developed for characterizing behavior in binary decision-making and bandit tasks, comparatively little work has focused on animal decision-making in more complex tasks, such as navigation through a maze. Inverse reinforcement learning (IRL) is a promising approach for understanding such behavior, as it aims to infer the unknown reward function of an agent from its observed trajectories through state space. However, IRL has yet to be widely applied in neuroscience. One potential reason for this is that existing IRL frameworks assume that an agent's reward function is fixed over time. To address this shortcoming, we introduce dynamic inverse reinforcement learning (DIRL), a novel IRL framework that allows for time-varying intrinsic rewards. Our method parametrizes the unknown reward function as a time-varying linear combination of spatial reward maps (which we refer to as "goal maps"). We develop an efficient inference method for recovering this dynamic reward function from behavioral data. We demonstrate DIRL in simulated experiments and then apply it to a dataset of mice exploring a labyrinth. Our method returns interpretable reward functions for two separate cohorts of mice, and provides a novel characterization of exploratory behavior. We expect DIRL to have broad applicability in neuroscience, and to facilitate the design of biologically-inspired reward functions for training artificial agents.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Characterizing Crowd Preferences on Stadium Facilities through Dynamic Inverse Reinforcement Learning
    Dong, Yiwen
    Huang, Peide
    Noh, Hae Young
    PROCEEDINGS OF THE 10TH ACM INTERNATIONAL CONFERENCE ON SYSTEMS FOR ENERGY-EFFICIENT BUILDINGS, CITIES, AND TRANSPORTATION, BUILDSYS 2023, 2023, : 305 - 306
  • [2] Learning Behavior Styles with Inverse Reinforcement Learning
    Lee, Seong Jae
    popovic, Zoran
    ACM TRANSACTIONS ON GRAPHICS, 2010, 29 (04):
  • [3] Identification of animal behavioral strategies by inverse reinforcement learning
    Yamaguchi, Shoichiro
    Naoki, Honda
    Ikeda, Muneki
    Tsukada, Yuki
    Nakano, Shunji
    Mori, Ikue
    Ishii, Shin
    PLOS COMPUTATIONAL BIOLOGY, 2018, 14 (05):
  • [4] BAYESIAN INVERSE REINFORCEMENT LEARNING FOR COLLECTIVE ANIMAL MOVEMENT
    Schafer, Toryn L. J.
    Wikle, Christopher K.
    Hooten, Mevin B.
    ANNALS OF APPLIED STATISTICS, 2022, 16 (02): : 999 - 1013
  • [5] Modular inverse reinforcement learning for visuomotor behavior
    Rothkopf, Constantin A.
    Ballard, Dana H.
    BIOLOGICAL CYBERNETICS, 2013, 107 (04) : 477 - 490
  • [6] Modular inverse reinforcement learning for visuomotor behavior
    Constantin A. Rothkopf
    Dana H. Ballard
    Biological Cybernetics, 2013, 107 : 477 - 490
  • [7] Inverse Reinforcement Learning Based Stochastic Driver Behavior Learning
    Ozkan, Mehmet F.
    Rocque, Abishek J.
    Ma, Yao
    IFAC PAPERSONLINE, 2021, 54 (20): : 882 - 888
  • [8] Inverse Reinforcement Learning Using Dynamic Policy Programming
    Uchibe, Eiji
    Doya, Kenji
    FOUTH JOINT IEEE INTERNATIONAL CONFERENCES ON DEVELOPMENT AND LEARNING AND EPIGENETIC ROBOTICS (IEEE ICDL-EPIROB 2014), 2014, : 222 - 228
  • [9] Dynamic multiobjective optimization driven by inverse reinforcement learning
    Zou, Fei
    Yen, Gary G.
    Zhao, Chen
    INFORMATION SCIENCES, 2021, 575 : 468 - 484
  • [10] A behavior fusion method based on inverse reinforcement learning
    Shi, Haobin
    Li, Jingchen
    Chen, Shicong
    Hwang, Kao-Shing
    INFORMATION SCIENCES, 2022, 609 : 429 - 444