Conditioning and time representation in long short-term memory networks

被引:9
|
作者
Rivest, Francois [1 ,2 ]
Kalaska, John F. [3 ]
Bengio, Yoshua [4 ]
机构
[1] Royal Mil Coll Canada, Dept Math & Comp Sci, Stn Forces, Kingston, ON K7K 7B4, Canada
[2] Queens Univ, Ctr Neurosci Studies, Kingston, ON, Canada
[3] Univ Montreal, Dept Physiol, Montreal, PQ H3C 3J7, Canada
[4] Univ Montreal, Dept Comp Sci & Operat Res, Montreal, PQ, Canada
关键词
Time representation learning; Temporal-difference learning; Long short-term memory networks; Dopamine; Conditioning; Reinforcement learning; PARAMETRIC WORKING-MEMORY; MONKEY DOPAMINE NEURONS; REWARD-PREDICTION; PREMOTOR CORTEX; MODEL; RESPONSES; HIPPOCAMPUS; INTERVALS; DYNAMICS; STIMULUS;
D O I
10.1007/s00422-013-0575-1
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Dopaminergic models based on the temporal-difference learning algorithm usually do not differentiate trace from delay conditioning. Instead, they use a fixed temporal representation of elapsed time since conditioned stimulus onset. Recently, a new model was proposed in which timing is learned within a long short-term memory (LSTM) artificial neural network representing the cerebral cortex (Rivest et al. in J Comput Neurosci 28(1):107-130, 2010). In this paper, that model's ability to reproduce and explain relevant data, as well as its ability to make interesting new predictions, are evaluated. The model reveals a strikingly different temporal representation between trace and delay conditioning since trace conditioning requires working memory to remember the past conditioned stimulus while delay conditioning does not. On the other hand, the model predicts no important difference in DA responses between those two conditions when trained on one conditioning paradigm and tested on the other. The model predicts that in trace conditioning, animal timing starts with the conditioned stimulus offset as opposed to its onset. In classical conditioning, it predicts that if the conditioned stimulus does not disappear after the reward, the animal may expect a second reward. Finally, the last simulation reveals that the buildup of activity of some units in the networks can adapt to new delays by adjusting their rate of integration. Most importantly, the paper shows that it is possible, with the proposed architecture, to acquire discharge patterns similar to those observed in dopaminergic neurons and in the cerebral cortex on those tasks simply by minimizing a predictive cost function.
引用
收藏
页码:23 / 48
页数:26
相关论文
共 50 条
  • [1] Conditioning and time representation in long short-term memory networks
    Francois Rivest
    John F. Kalaska
    Yoshua Bengio
    Biological Cybernetics, 2014, 108 : 23 - 48
  • [2] Time-distanced gates in long short-term memory networks
    Gao, Riqiang
    Tang, Yucheng
    Xu, Kaiwen
    Huo, Yuankai
    Bao, Shunxing
    Antic, Sanja L.
    Epstein, Emily S.
    Deppen, Steve
    Paulson, Alexis B.
    Sandler, Kim L.
    Massion, Pierre P.
    Landman, Bennett A.
    MEDICAL IMAGE ANALYSIS, 2020, 65 (65)
  • [3] On the Initialization of Long Short-Term Memory Networks
    Ghazi, Mostafa Mehdipour
    Nielsen, Mads
    Pai, Akshay
    Modat, Marc
    Cardoso, M. Jorge
    Ourselin, Sebastien
    Sorensen, Lauge
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT I, 2019, 11953 : 275 - 286
  • [4] Evolving Long Short-Term Memory Networks
    Neto, Vicente Coelho Lobo
    Passos, Leandro Aparecido
    Papa, Joao Paulo
    COMPUTATIONAL SCIENCE - ICCS 2020, PT II, 2020, 12138 : 337 - 350
  • [5] Short-Term Traffic Prediction Using Long Short-Term Memory Neural Networks
    Abbas, Zainab
    Al-Shishtawy, Ahmad
    Girdzijauskas, Sarunas
    Vlassov, Vladimir
    2018 IEEE INTERNATIONAL CONGRESS ON BIG DATA (IEEE BIGDATA CONGRESS), 2018, : 57 - 65
  • [6] Diagnosing Dysarthria with Long Short-Term Memory Networks
    Mayle, Alex
    Mou, Zhiwei
    Bunescu, Razvan
    Mirshekarian, Sadegh
    Xu, Li
    Liu, Chang
    INTERSPEECH 2019, 2019, : 4514 - 4518
  • [7] Short-term traffic travel time forecasting using ensemble approach based on long short-term memory networks
    Jia, Xingli
    Zhou, Wuxiao
    Yang, Hongzhi
    Li, Shuangqing
    Chen, Xingpeng
    IET INTELLIGENT TRANSPORT SYSTEMS, 2023, 17 (06) : 1262 - 1273
  • [8] Molecular Design With Long Short-Term Memory Networks
    Grisoni, Francesca
    Schneider, Gisbert
    JOURNAL OF COMPUTER AIDED CHEMISTRY, 2019, 20 : 35 - 42
  • [9] Long Short Term Memory Networks for Short-Term Electric Load Forecasting
    Narayan, Apurva
    Hipel, Keith W.
    2017 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN, AND CYBERNETICS (SMC), 2017, : 2573 - 2578
  • [10] The time needed to consolidate short-term memory to long-term memory
    Takeyama, E
    Takenoshita, M
    Nishimura, S
    Yoshiya, I
    ANESTHESIOLOGY, 1998, 89 (3A) : U317 - U317