Neural network with unbounded activation functions is universal approximator

被引:171
|
作者
Sonoda, Sho [1 ]
Murata, Noboru [1 ]
机构
[1] Waseda Univ, Fac Sci & Engn, Shinjuku Ku, 3-4-1 Okubo, Tokyo 1698555, Japan
关键词
Neural network; Integral representation; Rectified linear unit (ReLU); Universal approximation; Ridgelet transform; Admissibility condition; Lizorkin distribution; Radon transform; Backprojection filter; Bounded extension to L-2; TRANSFORM; REPRESENTATION; SUPERPOSITIONS; RATES;
D O I
10.1016/j.acha.2015.12.005
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
This paper presents an investigation of the approximation property of neural networks with unbounded activation functions, such as the rectified linear unit (ReLU), which is the new de-facto standard of deep learning. The ReLU network can be analyzed by the ridgelet transform with respect to Lizorkin distributions. By showing three reconstruction formulas by using the Fourier slice theorem, the Radon transform, and Parseval's relation, it is shown that a neural network with unbounded activation functions still satisfies the universal approximation property. As an additional consequence, the ridgelet transform, or the backprojection filter in the Radon domain, is what the network learns after backpropagation. Subject to a constructive admissibility condition, the trained network can be obtained by simply discretizing the ridgelet transform, without backpropagation. Numerical examples not only support the consistency of the admissibility condition but also imply that some non-admissible cases result in low-pass filtering. (C) 2015 Elsevier Inc. All rights reserved.
引用
收藏
页码:233 / 268
页数:36
相关论文
共 50 条
  • [31] Global Robust Exponential Stability for Interval Delayed Neural Networks with Possibly Unbounded Activation Functions
    Qin, Sitian
    Fan, Dejun
    Yan, Ming
    Liu, Qinghe
    NEURAL PROCESSING LETTERS, 2014, 40 (01) : 35 - 50
  • [32] Smart Security Audit: Reinforcement Learning with a Deep Neural Network Approximator
    Pozdniakov, Konstantin
    Alonso, Eduardo
    Stankovic, Vladimir
    Tam, Kimberly
    Jones, Kevin
    2020 INTERNATIONAL CONFERENCE ON CYBER SITUATIONAL AWARENESS, DATA ANALYTICS AND ASSESSMENT (CYBER SA 2020), 2020,
  • [33] Fuzzy flip-flop based neural network as a function approximator
    Lovassy, Rita
    Koczy, Laszlo T.
    Gal, Laszlo
    2008 IEEE INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE FOR MEASUREMENT SYSTEMS AND APPLICATIONS, 2008, : 44 - 49
  • [34] Dynamical behaviors of Hopfield neural network with multilevel activation functions
    Liu, YG
    You, ZS
    Cao, LP
    CHAOS SOLITONS & FRACTALS, 2005, 25 (05) : 1141 - 1153
  • [35] Recurrent neural network synthesis using interaction activation functions
    Novakovic, BM
    1996 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, PROCEEDINGS, VOLS 1-4, 1996, : 1608 - 1613
  • [36] Efficient Neural Network Robustness Certification with General Activation Functions
    Zhang, Huan
    Weng, Tsui-Wei
    Chen, Pin-Yu
    Hsieh, Cho-Jui
    Daniel, Luca
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [37] High precision FPGA implementation of neural network activation functions
    Ortega-Zamorano, Francisco
    Jerez, Jose M.
    Juarez, Gustavo
    Perez, Jorge O.
    Franco, Leonardo
    2014 IEEE SYMPOSIUM ON INTELLIGENT EMBEDDED SYSTEMS (IES), 2014, : 55 - 60
  • [38] EvoDNN - An Evolutionary Deep Neural Network with Heterogeneous Activation Functions
    Cui, Peiyu
    Shabash, Boris
    Wiese, Kay C.
    2019 IEEE CONGRESS ON EVOLUTIONARY COMPUTATION (CEC), 2019, : 2362 - 2369
  • [39] Study of the Effect of Combining Activation Functions in a Convolutional Neural Network
    Guevara, M.
    Cruz, V
    Vergara, O.
    Nandayapa, M.
    Ochoa, H.
    Sossa, H.
    IEEE LATIN AMERICA TRANSACTIONS, 2021, 19 (05) : 844 - 852
  • [40] Empirical Loss Landscape Analysis of Neural Network Activation Functions
    Bosman, Anna Sergeevna
    Engelbrecht, Andries
    Helbig, Marde
    PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2023 COMPANION, 2023, : 2029 - 2037