A reinforcement learning diffusion decision model for value-based decisions

被引:94
|
作者
Fontanesi, Laura [1 ]
Gluth, Sebastian [1 ]
Spektor, Mikhail S. [1 ]
Rieskamp, Joerg [1 ]
机构
[1] Univ Basel, Fac Psychol, Missionsstr 62a, CH-4055 Basel, Switzerland
基金
瑞士国家科学基金会;
关键词
Decision-making; Computational modeling; Bayesian inference and parameter estimation; Response time models; CHOICE; EXPLAIN; BRAIN; FMRI;
D O I
10.3758/s13423-018-1554-2
中图分类号
B841 [心理学研究方法];
学科分类号
040201 ;
摘要
Psychological models of value-based decision-making describe how subjective values are formed and mapped to single choices. Recently, additional efforts have been made to describe the temporal dynamics of these processes by adopting sequential sampling models from the perceptual decision-making tradition, such as the diffusion decision model (DDM). These models, when applied to value-based decision-making, allow mapping of subjective values not only to choices but also to response times. However, very few attempts have been made to adapt these models to situations in which decisions are followed by rewards, thereby producing learning effects. In this study, we propose a new combined reinforcement learning diffusion decision model (RLDDM) and test it on a learning task in which pairs of options differ with respect to both value difference and overall value. We found that participants became more accurate and faster with learning, responded faster and more accurately when options had more dissimilar values, and decided faster when confronted with more attractive (i.e., overall more valuable) pairs of options. We demonstrate that the suggested RLDDM can accommodate these effects and does so better than previously proposed models. To gain a better understanding of the model dynamics, we also compare it to standard DDMs and reinforcement learning models. Our work is a step forward towards bridging the gap between two traditions of decision-making research.
引用
收藏
页码:1099 / 1121
页数:23
相关论文
共 50 条
  • [31] Value-based deep reinforcement learning for adaptive isolated intersection signal control
    Wan, Chia-Hao
    Hwang, Ming-Chorng
    IET INTELLIGENT TRANSPORT SYSTEMS, 2018, 12 (09) : 1005 - 1010
  • [32] Sample Complexity Bounds for Two Timescale Value-based Reinforcement Learning Algorithms
    Xu, Tengyu
    Liang, Yingbin
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130
  • [33] Value-Based Model: A new perspective in Medical Decision-making
    Riva, Silvia
    Pravettoni, Gabriella
    FRONTIERS IN PUBLIC HEALTH, 2016, 4
  • [34] Engaging multiple perspectives: A value-based decision-making model
    Hall, Dianne J.
    Davis, Robert A.
    DECISION SUPPORT SYSTEMS, 2007, 43 (04) : 1588 - 1604
  • [35] Value-based Blended Learning Model for Strengthening Students' Character
    Komalasari, Kokom
    Winarno
    Indrawadi, Junaidi
    INTERNATIONAL JOURNAL OF INSTRUCTION, 2023, 16 (04) : 689 - 706
  • [36] Does value-based management facilitate managerial decision-making? An analysis of divestiture decisions
    Firk, Sebastian
    Richter, Sven
    Wolff, Michael
    MANAGEMENT ACCOUNTING RESEARCH, 2021, 51
  • [37] Mental representations distinguish value-based decisions from perceptual decisions
    Stephanie M. Smith
    Ian Krajbich
    Psychonomic Bulletin & Review, 2021, 28 : 1413 - 1422
  • [38] Mental representations distinguish value-based decisions from perceptual decisions
    Smith, Stephanie M.
    Krajbich, Ian
    PSYCHONOMIC BULLETIN & REVIEW, 2021, 28 (04) : 1413 - 1422
  • [39] Adaptability Analysis of Value-based and Policy-based Deep Reinforcement Learning in Nuclear Field
    Tan, Sichao
    Liu, Zhen
    Liu, Yongchao
    Li, Tong
    Liang, Biao
    Wang, Bo
    Li, Jiangkuan
    Tian, Ruifeng
    Yuanzineng Kexue Jishu/Atomic Energy Science and Technology, 2024, 58 : 382 - 392
  • [40] Correcting biased value estimation in mixing value-based multi-agent reinforcement learning by multiple choice learning
    Liu, Bing
    Xie, Yuxuan
    Feng, Lei
    Fu, Ping
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2022, 116