A reinforcement learning diffusion decision model for value-based decisions

被引:94
|
作者
Fontanesi, Laura [1 ]
Gluth, Sebastian [1 ]
Spektor, Mikhail S. [1 ]
Rieskamp, Joerg [1 ]
机构
[1] Univ Basel, Fac Psychol, Missionsstr 62a, CH-4055 Basel, Switzerland
基金
瑞士国家科学基金会;
关键词
Decision-making; Computational modeling; Bayesian inference and parameter estimation; Response time models; CHOICE; EXPLAIN; BRAIN; FMRI;
D O I
10.3758/s13423-018-1554-2
中图分类号
B841 [心理学研究方法];
学科分类号
040201 ;
摘要
Psychological models of value-based decision-making describe how subjective values are formed and mapped to single choices. Recently, additional efforts have been made to describe the temporal dynamics of these processes by adopting sequential sampling models from the perceptual decision-making tradition, such as the diffusion decision model (DDM). These models, when applied to value-based decision-making, allow mapping of subjective values not only to choices but also to response times. However, very few attempts have been made to adapt these models to situations in which decisions are followed by rewards, thereby producing learning effects. In this study, we propose a new combined reinforcement learning diffusion decision model (RLDDM) and test it on a learning task in which pairs of options differ with respect to both value difference and overall value. We found that participants became more accurate and faster with learning, responded faster and more accurately when options had more dissimilar values, and decided faster when confronted with more attractive (i.e., overall more valuable) pairs of options. We demonstrate that the suggested RLDDM can accommodate these effects and does so better than previously proposed models. To gain a better understanding of the model dynamics, we also compare it to standard DDMs and reinforcement learning models. Our work is a step forward towards bridging the gap between two traditions of decision-making research.
引用
收藏
页码:1099 / 1121
页数:23
相关论文
共 50 条
  • [41] Applying Value-Based Deep Reinforcement Learning on KPI Time Series Anomaly Detection
    Zhang, Yu
    Wang, Tianbo
    2022 IEEE 15TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING (IEEE CLOUD 2022), 2022, : 197 - 202
  • [42] Rapid Visuomotor Responses Reflect Value-Based Decisions
    Carroll, Timothy J.
    McNamee, Daniel
    Ingram, James N.
    Wolpert, Daniel M.
    JOURNAL OF NEUROSCIENCE, 2019, 39 (20): : 3906 - 3920
  • [43] Mechanisms underlying the influence of saliency on value-based decisions
    Chen, Xiaomo
    Mihalas, Stefan
    Niebur, Ernst
    Stuphorn, Veit
    JOURNAL OF VISION, 2013, 13 (12): : 18
  • [44] The hippocampus supports deliberation during value-based decisions
    Bakkour, Akram
    Palombo, Daniela J.
    Zylberberg, Ariel
    Kang, Yul H. R.
    Reid, Allison
    Verfaellie, Mieke
    Shadlen, Michael N.
    Shohamy, Daphna
    ELIFE, 2019, 8
  • [45] Rethinking Exploration and Experience Exploitation in Value-Based Multi-Agent Reinforcement Learning
    Borzilov, Anatolii
    Skrynnik, Alexey
    Panov, Aleksandr
    IEEE ACCESS, 2025, 13 : 13770 - 13781
  • [46] Item memorability has no influence on value-based decisions
    Xinyue Li
    Wilma A. Bainbridge
    Akram Bakkour
    Scientific Reports, 12 (1)
  • [47] Item memorability has no influence on value-based decisions
    Li, Xinyue
    Bainbridge, Wilma A.
    Bakkour, Akram
    SCIENTIFIC REPORTS, 2022, 12 (01):
  • [48] Body dynamics of gait affect value-based decisions
    Griessbach, Eric
    Incagli, Francesca
    Herbort, Oliver
    Canal-Bruland, Rouwen
    SCIENTIFIC REPORTS, 2021, 11 (01)
  • [49] Body dynamics of gait affect value-based decisions
    Eric Grießbach
    Francesca Incagli
    Oliver Herbort
    Rouwen Cañal-Bruland
    Scientific Reports, 11
  • [50] A diffusion model decomposition of value-based decision-making in regular drinkers following experimental manipulation of alcohol demand
    Copeland, A.
    Stafford, T.
    Field, M.
    ALCOHOL-CLINICAL AND EXPERIMENTAL RESEARCH, 2023, 47 : 75 - 75