Dynamic Inverse Reinforcement Learning for Characterizing Animal Behavior

被引:0
|
作者
Ashwood, Zoe C. [1 ,2 ]
Jha, Aditi [1 ,3 ]
Pillow, JonathanW. [1 ]
机构
[1] Princeton Univ, Princeton Neurosci Inst, Princeton, NJ 08544 USA
[2] Princeton Univ, Dept Comp Sci, Princeton, NJ 08544 USA
[3] Princeton Univ, Dept Elect & Comp Engn, Princeton, NJ USA
关键词
MODELS;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Understanding decision-making is a core objective in both neuroscience and psychology, and computational models have often been helpful in the pursuit of this goal. While many models have been developed for characterizing behavior in binary decision-making and bandit tasks, comparatively little work has focused on animal decision-making in more complex tasks, such as navigation through a maze. Inverse reinforcement learning (IRL) is a promising approach for understanding such behavior, as it aims to infer the unknown reward function of an agent from its observed trajectories through state space. However, IRL has yet to be widely applied in neuroscience. One potential reason for this is that existing IRL frameworks assume that an agent's reward function is fixed over time. To address this shortcoming, we introduce dynamic inverse reinforcement learning (DIRL), a novel IRL framework that allows for time-varying intrinsic rewards. Our method parametrizes the unknown reward function as a time-varying linear combination of spatial reward maps (which we refer to as "goal maps"). We develop an efficient inference method for recovering this dynamic reward function from behavioral data. We demonstrate DIRL in simulated experiments and then apply it to a dataset of mice exploring a labyrinth. Our method returns interpretable reward functions for two separate cohorts of mice, and provides a novel characterization of exploratory behavior. We expect DIRL to have broad applicability in neuroscience, and to facilitate the design of biologically-inspired reward functions for training artificial agents.
引用
收藏
页数:14
相关论文
共 50 条
  • [41] Inverse Reinforcement Learning with Failed Demonstrations towards Stable Driving Behavior Modeling
    Zhao, Minglu
    Shimosaka, Masamichi
    2024 35TH IEEE INTELLIGENT VEHICLES SYMPOSIUM, IEEE IV 2024, 2024, : 2537 - 2544
  • [42] Reinforcement Learning and Inverse Reinforcement Learning with System 1 and System 2
    Peysakhovich, Alexander
    AIES '19: PROCEEDINGS OF THE 2019 AAAI/ACM CONFERENCE ON AI, ETHICS, AND SOCIETY, 2019, : 409 - 415
  • [43] Recognition and interfere deceptive behavior based on inverse reinforcement learning and game theory
    ZENG Yunxiu
    XU Kai
    JournalofSystemsEngineeringandElectronics, 2023, 34 (02) : 270 - 288
  • [44] The Effects of Feedback on Human Behavior in Social Media: An Inverse Reinforcement Learning Model
    Das, Sanmay
    Lavoie, Allen
    AAMAS'14: PROCEEDINGS OF THE 2014 INTERNATIONAL CONFERENCE ON AUTONOMOUS AGENTS & MULTIAGENT SYSTEMS, 2014, : 653 - 660
  • [45] Recognition and interfere deceptive behavior based on inverse reinforcement learning and game theory
    Zeng, Yunxiu
    Xu, Kai
    JOURNAL OF SYSTEMS ENGINEERING AND ELECTRONICS, 2023, 34 (02) : 270 - 288
  • [46] Probabilistic Prediction of Interactive Driving Behavior via Hierarchical Inverse Reinforcement Learning
    Sun, Liting
    Zhan, Wei
    Tomizuka, Masayoshi
    2018 21ST INTERNATIONAL CONFERENCE ON INTELLIGENT TRANSPORTATION SYSTEMS (ITSC), 2018, : 2111 - 2117
  • [47] INVERSE REINFORCEMENT LEARNING BASED DRIVER BEHAVIOR ANALYSIS AND FUEL ECONOMY ASSESSMENT
    Ozkan, Mehmet Fatih
    Ma, Yao
    PROCEEDINGS OF THE ASME DYNAMIC SYSTEMS AND CONTROL CONFERENCE, DSCC2020, VOL 1, 2020,
  • [48] Fast Inverse Reinforcement Learning with Interval Consistent Graph for Driving Behavior Prediction
    Shimosaka, Masamichi
    Sato, Junichi
    Takenaka, Kazuhito
    Hitomi, Kentarou
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 1532 - 1538
  • [49] User Behavior Analysis in Online Health Community Based on Inverse Reinforcement Learning
    Zhang, Yaqi
    Wang, Xi
    Zuo, Zhiya
    Fan, Dan
    E-BUSINESS: NEW CHALLENGES AND OPPORTUNITIES FOR DIGITAL-ENABLED INTELLIGENT FUTURE, PT III, WHICEB 2024, 2024, 517 : 250 - 259
  • [50] Inverse Reinforcement Learning for Text Summarization
    Fu, Yu
    Xiong, Deyi
    Dong, Yue
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 6559 - 6570