Continuous limits of residual neural networks in case of large input data

被引:0
|
作者
Herty, Michael [1 ]
Thuenen, Anna [2 ]
Trimborn, Torsten [3 ]
Visconti, Giuseppe [4 ]
机构
[1] Rhein Westfal TH Aachen, Inst Geometrie & Prakt Math IGPM, Templergraben 55, D-52062 Aachen, Germany
[2] Tech Univ Clausthal, Inst Math, Erzstr 1, D-38678 Clausthal Zellerfeld, Germany
[3] NRW BANK, Kavalleriestr 22, D-40213 Dusseldorf, Germany
[4] Sapienza Univ Rome, Dept Math G Castelnuovo, Ple Aldo Moro 5, I-00185 Rome, Italy
关键词
Neural networks; mean-field limit; well-posedness; optimal control; controllability; LEARNING FRAMEWORK; CONVERGENCE;
D O I
10.2478/caim-2022-0008
中图分类号
O1 [数学];
学科分类号
0701 ; 070101 ;
摘要
Residual deep neural networks (ResNets) are mathematically described as interacting particle systems. In the case of infinitely many layers the ResNet leads to a system of coupled system of ordinary differential equations known as neural differential equations. For large scale input data we derive a mean-field limit and show well-posedness of the resulting description. Further, we analyze the existence of solutions to the training process by using both a controllability and an optimal control point of view. Numerical investigations based on the solution of a formal optimality system illustrate the theoretical findings.
引用
收藏
页码:96 / 120
页数:25
相关论文
共 50 条
  • [1] Deep limits of residual neural networks
    Thorpe, Matthew
    van Gennip, Yves
    RESEARCH IN THE MATHEMATICAL SCIENCES, 2023, 10 (01)
  • [2] An Input Residual Connection for Simplifying Gated Recurrent Neural Networks
    Kuo, Nicholas I. H.
    Harandi, Mehrtash
    Fourrier, Nicolas
    Walder, Christian
    Ferraro, Gabriela
    Suominen, Hanna
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [3] Input data analysis using neural networks
    Yilmaz, A
    Sabuncuoglu, L
    SIMULATION, 2000, 74 (03) : 128 - 137
  • [4] Fast learning artificial neural networks for continuous input applications
    Evans, DJ
    Tay, LP
    KYBERNETES, 1995, 24 (03) : 11 - &
  • [5] Deep limits of residual neural networks (vol 10, 6, 2023)
    Thorpe, Matthew
    van Gennip, Yves
    RESEARCH IN THE MATHEMATICAL SCIENCES, 2024, 11 (02)
  • [6] Training Neural Networks Using Input Data Characteristics
    Cernazanu, Cosmin
    ADVANCES IN ELECTRICAL AND COMPUTER ENGINEERING, 2008, 8 (02) : 65 - 70
  • [7] Modeling of continuous time dynamical systems with input by recurrent neural networks
    Chow, TWS
    Li, XD
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-FUNDAMENTAL THEORY AND APPLICATIONS, 2000, 47 (04): : 575 - 578
  • [8] Training Deep Neural Networks with Low Precision Input Data: A Hurricane Prediction Case Study
    Kahira, Albert
    Bautista Gomez, Leonardo
    Badia, Rosa M.
    HIGH PERFORMANCE COMPUTING, ISC HIGH PERFORMANCE 2018, 2018, 11203 : 562 - 569
  • [9] The Impact of Input Data Density on the Performance of Graphic Neural Networks
    Bondareva, N.A.
    Scientific Visualization, 2024, 16 (05): : 109 - 119
  • [10] Reducing prediction error by transforming input data for neural networks
    Shi, JJS
    JOURNAL OF COMPUTING IN CIVIL ENGINEERING, 2000, 14 (02) : 109 - 116