Dynamic Inverse Reinforcement Learning for Characterizing Animal Behavior

被引:0
|
作者
Ashwood, Zoe C. [1 ,2 ]
Jha, Aditi [1 ,3 ]
Pillow, JonathanW. [1 ]
机构
[1] Princeton Univ, Princeton Neurosci Inst, Princeton, NJ 08544 USA
[2] Princeton Univ, Dept Comp Sci, Princeton, NJ 08544 USA
[3] Princeton Univ, Dept Elect & Comp Engn, Princeton, NJ USA
关键词
MODELS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Understanding decision-making is a core objective in both neuroscience and psychology, and computational models have often been helpful in the pursuit of this goal. While many models have been developed for characterizing behavior in binary decision-making and bandit tasks, comparatively little work has focused on animal decision-making in more complex tasks, such as navigation through a maze. Inverse reinforcement learning (IRL) is a promising approach for understanding such behavior, as it aims to infer the unknown reward function of an agent from its observed trajectories through state space. However, IRL has yet to be widely applied in neuroscience. One potential reason for this is that existing IRL frameworks assume that an agent's reward function is fixed over time. To address this shortcoming, we introduce dynamic inverse reinforcement learning (DIRL), a novel IRL framework that allows for time-varying intrinsic rewards. Our method parametrizes the unknown reward function as a time-varying linear combination of spatial reward maps (which we refer to as "goal maps"). We develop an efficient inference method for recovering this dynamic reward function from behavioral data. We demonstrate DIRL in simulated experiments and then apply it to a dataset of mice exploring a labyrinth. Our method returns interpretable reward functions for two separate cohorts of mice, and provides a novel characterization of exploratory behavior. We expect DIRL to have broad applicability in neuroscience, and to facilitate the design of biologically-inspired reward functions for training artificial agents.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Individual Human Behavior Identification Using an Inverse Reinforcement Learning Method
    Inga, Jairo
    Koepf, Florian
    Flad, Michael
    Hohmann, Soeren
    2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, : 99 - 104
  • [22] Contextual Action with Multiple Policies Inverse Reinforcement Learning for Behavior Simulation
    Alvarez, Nahum
    Noda, Itsuki
    PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON AGENTS AND ARTIFICIAL INTELLIGENCE (ICAART), VOL 2, 2019, : 887 - 894
  • [23] Rationally Inattentive Inverse Reinforcement Learning Explains YouTube Commenting Behavior
    Hoiles, William
    Krishnamurthy, Vikram
    Pattanayak, Kunal
    JOURNAL OF MACHINE LEARNING RESEARCH, 2020, 21
  • [24] Modeling Driver Lane Change Behavior Using Inverse Reinforcement Learning
    Zhou, Zhaodong
    Chen, Jun
    2024 IEEE 3RD INTERNATIONAL CONFERENCE ON COMPUTING AND MACHINE INTELLIGENCE, ICMI 2024, 2024,
  • [25] Connections Between Relational Event Model and Inverse Reinforcement Learning for Characterizing Group Interaction Sequences
    Wu, Congyu
    IEEE TRANSACTIONS ON COMPUTATIONAL SOCIAL SYSTEMS, 2022, 9 (04) : 1029 - 1037
  • [26] Repeated Inverse Reinforcement Learning
    Amin, Kareem
    Jiang, Nan
    Singh, Satinder
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [27] Cooperative Inverse Reinforcement Learning
    Hadfield-Menell, Dylan
    Dragan, Anca
    Abbeel, Pieter
    Russell, Stuart
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 29 (NIPS 2016), 2016, 29
  • [28] Misspecification in Inverse Reinforcement Learning
    Skalse, Joar
    Abate, Alessandro
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 12, 2023, : 15136 - 15143
  • [29] Lifelong Inverse Reinforcement Learning
    Mendez, Jorge A.
    Shivkumar, Shashank
    Eaton, Eric
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [30] Bayesian Inverse Reinforcement Learning
    Ramachandran, Deepak
    Amir, Eyal
    20TH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2007, : 2586 - 2591