ON THE PROBLEM OF LOCAL MINIMA IN RECURRENT NEURAL NETWORKS

被引:34
|
作者
BIANCHINI, M
GORI, M
MAGGINI, M
机构
[1] Dipartimento di Sistemi e Informatica, Universita di Firenze, 50139, Firenze
来源
关键词
D O I
10.1109/72.279182
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Many researchers have recently focused their efforts on devising efficient algorithms, mainly based on optimization schemes, for learning the weights of recurrent neural networks. As in the case of feedforward networks, however, these learning algorithms may get stuck in local minima during gradient descent, thus discovering sub-optimal solutions. This paper analyses the problem of optimal learning in recurrent networks by proposing conditions that guarantee local minima free error surfaces. An example is given that also shows the constructive role of the proposed theory in designing networks suitable for solving a given task. Moreover, a formal relationship between recurrent and static feedforward networks is established such that the examples of local minima for feedforward networks already known in the literature can be associated with analogous ones in recurrent networks.
引用
收藏
页码:167 / 172
页数:6
相关论文
共 50 条
  • [41] A neural network for tornado diagnosis: Managing local minima
    Marzban, C
    NEURAL COMPUTING & APPLICATIONS, 2000, 9 (02): : 133 - 141
  • [42] Local structure supports learning of deterministic behavior in recurrent neural networks
    Jonathan Binas
    Giacomo Indiveri
    Michael Pfeiffer
    BMC Neuroscience, 16 (Suppl 1)
  • [43] Local Structure Helps Learning Optimized Automata in Recurrent Neural Networks
    Binas, Jonathan
    Indiveri, Giacomo
    Pfeiffer, Michael
    2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [44] Local Feature based Online Mode Detection with Recurrent Neural Networks
    Otte, Sebastian
    Krechel, Dirk
    Liwicki, Marcus
    Dengel, Andreas
    13TH INTERNATIONAL CONFERENCE ON FRONTIERS IN HANDWRITING RECOGNITION (ICFHR 2012), 2012, : 533 - 537
  • [45] Preprocessing based solution for the vanishing gradient problem in recurrent neural networks
    Squartini, S
    Hussain, A
    Piazzao, F
    PROCEEDINGS OF THE 2003 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOL V: BIO-MEDICAL CIRCUITS & SYSTEMS, VLSI SYSTEMS & APPLICATIONS, NEURAL NETWORKS & SYSTEMS, 2003, : 713 - 716
  • [46] Recurrent neural networks
    Siegelmann, HT
    COMPUTER SCIENCE TODAY, 1995, 1000 : 29 - 45
  • [47] Communities of minima in local optima networks of combinatorial spaces
    Daolio, Fabio
    Tomassini, Marco
    Verel, Sebastien
    Ochoa, Gabriela
    PHYSICA A-STATISTICAL MECHANICS AND ITS APPLICATIONS, 2011, 390 (09) : 1684 - 1694
  • [48] How synchronized human networks escape local minima
    Shniderman, Elad
    Avraham, Yahav
    Shahal, Shir
    Duadi, Hamootal
    Davidson, Nir
    Fridman, Moti
    NATURE COMMUNICATIONS, 2024, 15 (01)
  • [49] BACKPROPAGATION GROWING NETWORKS - TOWARDS LOCAL MINIMA ELIMINATION
    BELLIDO, I
    FERNANDEZ, G
    LECTURE NOTES IN COMPUTER SCIENCE, 1991, 540 : 130 - 135
  • [50] The normalized risk-averting error criterion for avoiding nonglobal local minima in training neural networks
    Lo, James Ting-Ho
    Gui, Yichuan
    Peng, Yun
    NEUROCOMPUTING, 2015, 149 : 3 - 12