Function-Space Optimality of Neural Architectures with Multivariate Nonlinearities

被引:0
|
作者
Parhi, Rahul [1 ,2 ]
Unser, Michael [3 ]
机构
[1] Univ Calif San Diego, Dept Elect & Comp Engn, La Jolla, CA 92093 USA
[2] Ecole Polytech Fed Lausanne, Biomed Imaging Grp, CH-1015 Lausanne, Switzerland
[3] Ecole Polytech Fed Lausanne, Biomed Imaging Grp, CH-1015 Lausanne, Switzerland
来源
基金
瑞士国家科学基金会;
关键词
multi-index models; multivariate nonlinearities; neural networks; regularization; representer theorem; PLANE TRANSFORMS; RIDGE FUNCTIONS; APPROXIMATION; NETWORKS; REGRESSION; REDUCTION; INVERSION; SPLINES;
D O I
10.1137/23M1620971
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
We investigate the function-space optimality (specifically, the Banach-space optimality) of a large class of shallow neural architectures with multivariate nonlinearities/activation functions. To that end, we construct a new family of Banach spaces defined via a regularization operator, the k-plane transform, and a sparsity-promoting norm. We prove a representer theorem that states that the solution sets to learning problems posed over these Banach spaces are completely characterized by neural architectures with multivariate nonlinearities. These optimal architectures have skip connections and are tightly connected to orthogonal weight normalization and multi-index models, both of which have received recent interest in the neural network community. Our framework is compatible with a number of classical nonlinearities including the rectified linear unit activation function, the norm activation function, and the radial basis functions found in the theory of thin-plate/polyharmonic splines. We also show that the underlying spaces are special instances of reproducing kernel Banach spaces and variation spaces. Our results shed light on the regularity of functions learned by neural networks trained on data, particularly with multivariate nonlinearities, and provide new theoretical motivation for several architectural choices found in practice.
引用
收藏
页码:110 / 135
页数:26
相关论文
共 50 条
  • [1] Tractable Function-Space Variational Inference in Bayesian Neural Networks
    Rudner, Tim G. J.
    Chen, Zonghao
    Teh, Yee Whye
    Gal, Yarin
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [2] Function-space compactifications of function spaces
    Escardó, MH
    TOPOLOGY AND ITS APPLICATIONS, 2002, 120 (03) : 441 - 463
  • [3] 2 FUNCTION-SPACE TOPOLOGIES
    EDGAR, GA
    PROCEEDINGS OF THE AMERICAN MATHEMATICAL SOCIETY, 1973, 39 (01) : 219 - 220
  • [4] Function-Space Distributions over Kernels
    Benton, Gregory W.
    Maddox, Wesley J.
    Salkey, Jayson P.
    Albinati, Jujlio
    Wilson, Andrew Gordon
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [5] TRANSITION FUNCTORS TO A FUNCTION-SPACE IN THE UNIFORM TOPOLOGY
    LEMIN, AY
    RUSSIAN MATHEMATICAL SURVEYS, 1985, 40 (06) : 133 - 134
  • [6] Function-space Inference with Sparse Implicit Processes
    Rodriguez Santana, Simon
    Zaldivar, Bryan
    Hernandez-Lobato, Daniel
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [7] ON THE FUNCTION-SPACE CONTROLLABILITY OF LINEAR NEUTRAL SYSTEMS
    OCONNOR, DA
    TARN, TJ
    SIAM JOURNAL ON CONTROL AND OPTIMIZATION, 1983, 21 (02) : 306 - 329
  • [8] TOPOLOGICAL DUALITY ON THE FUNCTION-SPACE H(CN)
    ANSEMIL, JM
    JOURNAL OF MATHEMATICAL ANALYSIS AND APPLICATIONS, 1979, 67 (01) : 188 - 197
  • [9] HARDY TYPE INEQUALITIES ON CONE OF FUNCTION-SPACE
    BEREZHNOI, EI
    DOKLADY AKADEMII NAUK, 1992, 326 (02) : 215 - 218
  • [10] CONSTRUCTING A QUASI-UNIFORM FUNCTION-SPACE
    SUNDERHAUF, P
    TOPOLOGY AND ITS APPLICATIONS, 1995, 67 (01) : 1 - 27