Neural network with unbounded activation functions is universal approximator

被引:171
|
作者
Sonoda, Sho [1 ]
Murata, Noboru [1 ]
机构
[1] Waseda Univ, Fac Sci & Engn, Shinjuku Ku, 3-4-1 Okubo, Tokyo 1698555, Japan
关键词
Neural network; Integral representation; Rectified linear unit (ReLU); Universal approximation; Ridgelet transform; Admissibility condition; Lizorkin distribution; Radon transform; Backprojection filter; Bounded extension to L-2; TRANSFORM; REPRESENTATION; SUPERPOSITIONS; RATES;
D O I
10.1016/j.acha.2015.12.005
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This paper presents an investigation of the approximation property of neural networks with unbounded activation functions, such as the rectified linear unit (ReLU), which is the new de-facto standard of deep learning. The ReLU network can be analyzed by the ridgelet transform with respect to Lizorkin distributions. By showing three reconstruction formulas by using the Fourier slice theorem, the Radon transform, and Parseval's relation, it is shown that a neural network with unbounded activation functions still satisfies the universal approximation property. As an additional consequence, the ridgelet transform, or the backprojection filter in the Radon domain, is what the network learns after backpropagation. Subject to a constructive admissibility condition, the trained network can be obtained by simply discretizing the ridgelet transform, without backpropagation. Numerical examples not only support the consistency of the admissibility condition but also imply that some non-admissible cases result in low-pass filtering. (C) 2015 Elsevier Inc. All rights reserved.
引用
收藏
页码:233 / 268
页数:36
相关论文
共 50 条
  • [11] Reduction of training computation by network optimization of Integration Neural Network approximator
    Iwata, Yoshiharu
    Wakamatsu, Hidefumi
    2023 IEEE/SICE INTERNATIONAL SYMPOSIUM ON SYSTEM INTEGRATION, SII, 2023,
  • [12] A Compact Neural Network for Fused Lasso Signal Approximator
    Mohammadi, Majid
    IEEE TRANSACTIONS ON CYBERNETICS, 2021, 51 (08) : 4327 - 4336
  • [13] ReLU Network with Bounded Width Is a Universal Approximator in View of an Approximate Identity
    Moon, Sunghwan
    APPLIED SCIENCES-BASEL, 2021, 11 (01): : 1 - 11
  • [14] The learning dynamics of a universal approximator
    West, AHL
    Saad, D
    Nabney, IT
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 9: PROCEEDINGS OF THE 1996 CONFERENCE, 1997, 9 : 288 - 294
  • [15] On a neural approximator to ODEs
    Filici, Cristian
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (03): : 539 - 543
  • [16] Universal approximation capability of EBF neural networks with arbitrary activation functions
    Chen, TP
    Chen, H
    CIRCUITS SYSTEMS AND SIGNAL PROCESSING, 1996, 15 (05) : 671 - 683
  • [17] Universal Approximation Using Probabilistic Neural Networks with Sigmoid Activation Functions
    Murugadoss, R.
    Ramakrishnan, M.
    2014 INTERNATIONAL CONFERENCE ON ADVANCES IN ENGINEERING AND TECHNOLOGY RESEARCH (ICAETR), 2014,
  • [18] Resource efficient activation functions for neural network accelerators
    Wuraola, Adedamola
    Patel, Nitish
    NEUROCOMPUTING, 2022, 482 : 163 - 185
  • [19] Fusion of Chaotic Activation Functions in Training Neural Network
    Kabir, A. N. M. Enamul
    Uddin, A. F. M. Nokib
    Asaduzzaman, M.
    Hasan, M. Fahad
    Hasan, M. Imran
    Shahjahan, Md
    2012 7TH INTERNATIONAL CONFERENCE ON ELECTRICAL AND COMPUTER ENGINEERING (ICECE), 2012,
  • [20] Harnessing Chaotic Activation Functions in Training Neural Network
    Asaduzzaman, Md
    Uddin, A. F. M. Nokib
    Shahjahan, Md
    Murase, Kazuyuki
    NEURAL INFORMATION PROCESSING, ICONIP 2012, PT II, 2012, 7664 : 551 - 558