A Representer Theorem for Deep Neural Networks

被引:0
|
作者
Unser, Michael [1 ]
机构
[1] Ecole Polytech Fed Lausanne, Biomed Imaging Grp, CH-1015 Lausanne, Switzerland
基金
瑞士国家科学基金会;
关键词
splines; regularization; sparsity; learning; deep neural networks; activation functions; LINEAR INVERSE PROBLEMS; SPLINES; KERNELS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We propose to optimize the activation functions of a deep neural network by adding a corresponding functional regularization to the cost function. We justify the use of a second-order total-variation criterion. This allows us to derive a general representer theorem for deep neural networks that makes a direct connection with splines and sparsity. Specifically, we show that the optimal network configuration can be achieved with activation functions that are nonuniform linear splines with adaptive knots. The bottom line is that the action of each neuron is encoded by a spline whose parameters (including the number of knots) are optimized during the training procedure. The scheme results in a computational structure that is compatible with existing deep-ReLU, parametric ReLU, APL (adaptive piecewise-linear) and MaxOut architectures. It also suggests novel optimization challenges and makes an explicit link with l(1) minimization and sparsity-promoting techniques.
引用
收藏
页数:30
相关论文
共 50 条
  • [31] Orthogonal Deep Neural Networks
    Li, Shuai
    Jia, Kui
    Wen, Yuxin
    Liu, Tongliang
    Tao, Dacheng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (04) : 1352 - 1368
  • [32] Deep neural networks in psychiatry
    Durstewitz, Daniel
    Koppe, Georgia
    Meyer-Lindenberg, Andreas
    MOLECULAR PSYCHIATRY, 2019, 24 (11) : 1583 - 1598
  • [33] Tracking with Deep Neural Networks
    Jin, Jonghoon
    Dundar, Aysegul
    Bates, Jordan
    Farabet, Clement
    Culurciello, Eugenio
    2013 47TH ANNUAL CONFERENCE ON INFORMATION SCIENCES AND SYSTEMS (CISS), 2013,
  • [34] Tweaking Deep Neural Networks
    Kim, Jinwook
    Yoon, Heeyong
    Kim, Min-Soo
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (09) : 5715 - 5728
  • [35] Deep Convolutional Neural Networks
    Gonzalez, Rafael C.
    IEEE SIGNAL PROCESSING MAGAZINE, 2018, 35 (06) : 79 - 87
  • [36] Ranking with Deep Neural Networks
    Prakash, Chandan
    Sarkar, Amitrajit
    PROCEEDINGS OF 2018 FIFTH INTERNATIONAL CONFERENCE ON EMERGING APPLICATIONS OF INFORMATION TECHNOLOGY (EAIT), 2018,
  • [37] Deep Polynomial Neural Networks
    Chrysos, Grigorios G.
    Moschoglou, Stylianos
    Bouritsas, Giorgos
    Deng, Jiankang
    Panagakis, Yannis
    Zafeiriou, Stefanos
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (08) : 4021 - 4034
  • [38] On the Opacity of Deep Neural Networks
    Sogaard, Anders
    CANADIAN JOURNAL OF PHILOSOPHY, 2024,
  • [39] Deep Morphological Neural Networks
    Shen, Yucong
    Shih, Frank Y.
    Zhong, Xin
    Chang, I-Cheng
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2022, 36 (12)
  • [40] Deep Dive into Deep Neural Networks with Flows
    Hainaut, Adrien
    Giot, Romain
    Bourqui, Romain
    Auber, David
    IVAPP: PROCEEDINGS OF THE 15TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS, VOL 3: IVAPP, 2020, : 231 - 239