A Kernel Perspective for the Decision Boundary of Deep Neural Networks

被引:1
|
作者
Zhang, Yifan [1 ]
Liao, Shizhong [1 ]
机构
[1] Tianjin Univ, Coll Intelligence & Comp, Tianjin 300350, Peoples R China
基金
中国国家自然科学基金;
关键词
deep neural network; kernel method; generalization ability; gradient descent; decision boundary;
D O I
10.1109/ICTAI50040.2020.00105
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning has achieved great success in many fields, but they still lack theoretical understandings. Although some recent theoretical and experimental results have investigated the representation power of deep learning, little effort has been devoted to analyzing the generalization ability of deep learning. In this paper, we analyze deep neural networks from a kernel perspective and use kernel methods to investigate the effect of the implicit regularization introduced by gradient descent on the generalization ability. Firstly, we argue that the multi-layer nonlinear feature transformation in deep neural networks is equivalent to a kernel feature mapping and analyze our point from the perspective of the unique mathematical advantages of kernel methods and the method of constructing multi-layer kernel machines, respectively. Secondly, using the representer theorem, we analyze the decision boundary of deep neural networks and prove that the last hidden layers of deep neural networks converge to nonlinear SVMs. Systematical experiments demonstrate that the decision boundaries of neural networks converge to those of nonlinear SVMs.
引用
收藏
页码:653 / 660
页数:8
相关论文
共 50 条
  • [41] Feature Overcorrelation in Deep Graph Neural Networks: A New Perspective
    Jin, Wei
    Liu, Xiaorui
    Ma, Yao
    Aggarwal, Charu
    Tang, Jiliang
    PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 709 - 719
  • [42] Graph Kernel Neural Networks
    Cosmo, Luca
    Minello, Giorgia
    Bicciato, Alessandro
    Bronstein, Michael M.
    Rodola, Emanuele
    Rossi, Luca
    Torsello, Andrea
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 14
  • [43] ReLU deep neural networks from the hierarchical basis perspective
    He, Juncai
    Li, Lin
    Xu, Jinchao
    COMPUTERS & MATHEMATICS WITH APPLICATIONS, 2022, 120 : 105 - 114
  • [44] Deep Neural Networks Pruning via the Structured Perspective Regularization
    Cacciola, Matteo
    Frangioni, Antonio
    Li, Xinlin
    Lodi, Andrea
    SIAM JOURNAL ON MATHEMATICS OF DATA SCIENCE, 2023, 5 (04): : 1051 - 1077
  • [45] Counting manatee aggregations using deep neural networks and Anisotropic Gaussian Kernel
    Zhiqiang Wang
    Yiran Pang
    Cihan Ulus
    Xingquan Zhu
    Scientific Reports, 13
  • [46] Deep hybrid neural-kernel networks using random Fourier features
    Mehrkanoon, Siamak
    Suykens, Johan A. K.
    NEUROCOMPUTING, 2018, 298 : 46 - 54
  • [47] Counting manatee aggregations using deep neural networks and Anisotropic Gaussian Kernel
    Wang, Zhiqiang
    Pang, Yiran
    Ulus, Cihan
    Zhu, Xingquan
    SCIENTIFIC REPORTS, 2023, 13 (01)
  • [48] Tight Bounds on the Smallest Eigenvalue of the Neural Tangent Kernel for Deep ReLU Networks
    Nguyen, Quynh
    Mondelli, Marco
    Montufar, Guido
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [49] Kernel Product Neural Networks
    Xu, Hao
    Zhou, Shuyue
    Shen, Yang
    Lou, Kenan
    Zhang, Ruihua
    Ye, Zhen
    Li, Xiaobo
    Wang, Shuai
    IEEE ACCESS, 2021, 9 : 167076 - 167083
  • [50] Deep regularization and direct training of the inner layers of Neural Networks with Kernel Flows
    Yoo, Gene Ryan
    Owhadi, Houman
    PHYSICA D-NONLINEAR PHENOMENA, 2021, 426