Conditioning and time representation in long short-term memory networks

被引:9
|
作者
Rivest, Francois [1 ,2 ]
Kalaska, John F. [3 ]
Bengio, Yoshua [4 ]
机构
[1] Royal Mil Coll Canada, Dept Math & Comp Sci, Stn Forces, Kingston, ON K7K 7B4, Canada
[2] Queens Univ, Ctr Neurosci Studies, Kingston, ON, Canada
[3] Univ Montreal, Dept Physiol, Montreal, PQ H3C 3J7, Canada
[4] Univ Montreal, Dept Comp Sci & Operat Res, Montreal, PQ, Canada
关键词
Time representation learning; Temporal-difference learning; Long short-term memory networks; Dopamine; Conditioning; Reinforcement learning; PARAMETRIC WORKING-MEMORY; MONKEY DOPAMINE NEURONS; REWARD-PREDICTION; PREMOTOR CORTEX; MODEL; RESPONSES; HIPPOCAMPUS; INTERVALS; DYNAMICS; STIMULUS;
D O I
10.1007/s00422-013-0575-1
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Dopaminergic models based on the temporal-difference learning algorithm usually do not differentiate trace from delay conditioning. Instead, they use a fixed temporal representation of elapsed time since conditioned stimulus onset. Recently, a new model was proposed in which timing is learned within a long short-term memory (LSTM) artificial neural network representing the cerebral cortex (Rivest et al. in J Comput Neurosci 28(1):107-130, 2010). In this paper, that model's ability to reproduce and explain relevant data, as well as its ability to make interesting new predictions, are evaluated. The model reveals a strikingly different temporal representation between trace and delay conditioning since trace conditioning requires working memory to remember the past conditioned stimulus while delay conditioning does not. On the other hand, the model predicts no important difference in DA responses between those two conditions when trained on one conditioning paradigm and tested on the other. The model predicts that in trace conditioning, animal timing starts with the conditioned stimulus offset as opposed to its onset. In classical conditioning, it predicts that if the conditioned stimulus does not disappear after the reward, the animal may expect a second reward. Finally, the last simulation reveals that the buildup of activity of some units in the networks can adapt to new delays by adjusting their rate of integration. Most importantly, the paper shows that it is possible, with the proposed architecture, to acquire discharge patterns similar to those observed in dopaminergic neurons and in the cerebral cortex on those tasks simply by minimizing a predictive cost function.
引用
收藏
页码:23 / 48
页数:26
相关论文
共 50 条
  • [21] Long Short-Term Memory Networks for Vehicle Sensor Fusion
    Gandy, Jonah T.
    Ball, John E.
    AUTONOMOUS SYSTEMS: SENSORS, PROCESSING AND SECURITY FOR GROUND, AIR, SEA AND SPACE VEHICLES AND INFRASTRUCTURE 2022, 2022, 12115
  • [22] Long short-term memory networks in memristor crossbar arrays
    Can Li
    Zhongrui Wang
    Mingyi Rao
    Daniel Belkin
    Wenhao Song
    Hao Jiang
    Peng Yan
    Yunning Li
    Peng Lin
    Miao Hu
    Ning Ge
    John Paul Strachan
    Mark Barnell
    Qing Wu
    R. Stanley Williams
    J. Joshua Yang
    Qiangfei Xia
    Nature Machine Intelligence, 2019, 1 : 49 - 57
  • [23] Classification of HRV using Long Short-Term Memory Networks
    Leite, Argentina
    Silva, Maria Eduarda
    Rocha, Ana Paula
    2020 11TH CONFERENCE OF THE EUROPEAN STUDY GROUP ON CARDIOVASCULAR OSCILLATIONS (ESGCO): COMPUTATION AND MODELLING IN PHYSIOLOGY NEW CHALLENGES AND OPPORTUNITIES, 2020,
  • [24] A New Delay Connection for Long Short-Term Memory Networks
    Wang, Jianyong
    Zhang, Lei
    Chen, Yuanyuan
    Yi, Zhang
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2018, 28 (06)
  • [25] CCG supertagging with bidirectional long short-term memory networks
    Kadari, Rekia
    Zhang, Yu
    Zhang, Weinan
    Liu, Ting
    NATURAL LANGUAGE ENGINEERING, 2018, 24 (01) : 77 - 90
  • [26] Deep Long Short-Term Memory Networks for Speech Recognition
    Chien, Jen-Tzung
    Misbullah, Alim
    2016 10TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2016,
  • [27] Dialogue Intent Classification with Long Short-Term Memory Networks
    Meng, Lian
    Huang, Minlie
    NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING, NLPCC 2017, 2018, 10619 : 42 - 50
  • [28] Optical Music Recognition by Long Short-Term Memory Networks
    Baro, Arnau
    Riba, Pau
    Calvo-Zaragoza, Jorge
    Fornes, Alicia
    GRAPHICS RECOGNITION: CURRENT TRENDS AND EVOLUTIONS, GREC 2017, 2018, 11009 : 81 - 95
  • [29] Facade Layout Completion with Long Short-Term Memory Networks
    Hensel, Simon
    Goebbels, Steffen
    Kada, Martin
    COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS, VISIGRAPP 2021, 2023, 1691 : 21 - 40
  • [30] Matching Biomedical Ontologies with Long Short-Term Memory Networks
    Jiang, Chao
    Xue, Xingsi
    2020 IEEE INTERNATIONAL CONFERENCE ON BIOINFORMATICS AND BIOMEDICINE, 2020, : 2484 - 2489