A Kernel Analysis of Feature Learning in Deep Neural Networks

被引:1
|
作者
Canatar, Abdulkadir [1 ]
Pehlevan, Cengiz [2 ,3 ]
机构
[1] Flatiron Inst, Ctr Computat Neurosci, New York, NY USA
[2] Harvard Univ, John A Paulson Sch Engn & Appl Sci, Cambridge, MA 02138 USA
[3] Harvard Univ, Ctr Brain Sci, Cambridge, MA 02138 USA
关键词
deep learning; kernel methods;
D O I
10.1109/ALLERTON49937.2022.9929375
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks learn useful representations of data, yet the nature of these representations has not been fully understood. Here, we empirically study the kernels induced by the layer representations during training by analyzing their kernel alignment to the network's target function. We show that representations from earlier to deeper layers increasingly align with the target task for both training and test sets, implying better generalization. We analyze these representations across different architectures, optimization methods and batch sizes. Furthermore, we compare the Neural Tangent Kernel (NTK) of deep neural networks and its alignment with the target during training and find that NTK-target alignment also increases during training.
引用
收藏
页数:8
相关论文
共 50 条
  • [31] A Deep Hybrid Graph Kernel Through Deep Learning Networks
    Cui, Lixin
    Bai, Lu
    Rossi, Luca
    Wang, Yue
    Jiao, Yuhang
    Hancock, Edwin R.
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 1030 - 1035
  • [32] Improved feature processing for Deep Neural Networks
    Rath, Shakti P.
    Povey, Daniel
    Vesely, Karel
    Cernocky, Jan
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 109 - 113
  • [33] Discriminative Feature Extraction with Deep Neural Networks
    Stuhlsatz, Andre
    Lippel, Jens
    Zielke, Thomas
    2010 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS IJCNN 2010, 2010,
  • [34] On the Neural Tangent Kernel of Deep Networks with Orthogonal Initialization
    Huang, Wei
    Du, Weitao
    Da Xu, Richard Yi
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2577 - 2583
  • [35] The kernel-balanced equation for deep neural networks
    Nakazato, Kenichi
    PHYSICA SCRIPTA, 2023, 98 (10)
  • [36] Feature Selection using Deep Neural Networks
    Roy, Debaditya
    Murty, K. Sri Rama
    Mohan, C. Krishna
    2015 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2015,
  • [37] Versatile kernel reactivation for deep convolutional neural networks
    Lee, Jeong Jun
    Kim, Hyun
    ELECTRONICS LETTERS, 2022, 58 (19) : 723 - 725
  • [38] KERNEL METHODS MATCH DEEP NEURAL NETWORKS ON TIMIT
    Huang, Po-Sen
    Avron, Haim
    Sainath, Tara N.
    Sindhwani, Vikas
    Ramabhadran, Bhuvana
    2014 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2014,
  • [39] A Kernel Perspective for the Decision Boundary of Deep Neural Networks
    Zhang, Yifan
    Liao, Shizhong
    2020 IEEE 32ND INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2020, : 653 - 660
  • [40] Unsound wheat kernel recognition based on deep convolutional neural network transfer learning and feature fusion
    Zhang, Qinghui
    Tian, Xinxin
    Chen, Weidong
    Yang, Hongwei
    Lv, Pengtao
    Wu, Yong
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2022, 43 (05) : 5833 - 5858