Mechanism for feature learning in neural networks and backpropagation-free machine learning models

被引:12
|
作者
Radhakrishnan, Adityanarayanan [1 ,2 ]
Beaglehole, Daniel [3 ]
Pandit, Parthe [4 ,5 ]
Belkin, Mikhail [3 ,5 ]
机构
[1] Harvard Sch Engn & Appl Sci, Cambridge, MA 02138 USA
[2] Broad Inst MIT & Harvard, Cambridge, MA 02142 USA
[3] Univ Calif San Diego, Comp Sci & Engn, La Jolla, CA 92093 USA
[4] Indian Inst Technol, Ctr Machine Intelligence & Data Sci, Mumbai 400076, India
[5] Univ Calif San Diego, Halicioglu Data Sci Inst, La Jolla, CA 92093 USA
基金
美国国家科学基金会;
关键词
REGRESSION;
D O I
10.1126/science.adi5639
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Understanding how neural networks learn features, or relevant patterns in data, for prediction is necessary for their reliable use in technological and scientific applications. In this work, we presented a unifying mathematical mechanism, known as average gradient outer product (AGOP), that characterized feature learning in neural networks. We provided empirical evidence that AGOP captured features learned by various neural network architectures, including transformer-based language models, convolutional networks, multilayer perceptrons, and recurrent neural networks. Moreover, we demonstrated that AGOP, which is backpropagation-free, enabled feature learning in machine learning models, such as kernel machines, that a priori could not identify task-specific features. Overall, we established a fundamental mechanism that captured feature learning in neural networks and enabled feature learning in general machine learning models.
引用
收藏
页码:1461 / 1467
页数:7
相关论文
共 50 条
  • [41] Local kernel renormalization as a mechanism for feature learning in overparametrized convolutional neural networks
    Aiudi, R.
    Pacelli, R.
    Baglioni, P.
    Vezzani, A.
    Burioni, R.
    Rotondo, P.
    NATURE COMMUNICATIONS, 2025, 16 (01)
  • [42] Learning in the machine: Recirculation is random backpropagation
    Baldi, P.
    Sadowski, P.
    NEURAL NETWORKS, 2018, 108 : 479 - 494
  • [43] BPSpike II: A New Backpropagation Learning Algorithm for Spiking Neural Networks
    Matsuda, Satoshi
    NEURAL INFORMATION PROCESSING, ICONIP 2016, PT II, 2016, 9948 : 56 - 65
  • [44] Experimentally realized in situ backpropagation for deep learning in photonic neural networks
    Pai, Sunil
    Sun, Zhanghao
    Hughes, Tyler W.
    Park, Taewon
    Bartlett, Ben
    Williamson, Ian A. D.
    Minkov, Momchil
    Milanizadeh, Maziyar
    Abebe, Nathnael
    Morichetti, Francesco
    Melloni, Andrea
    Fan, Shanhui
    Solgaard, Olav
    Miller, David A. B.
    SCIENCE, 2023, 380 (6643) : 398 - 403
  • [45] Augmented Efficient BackProp for Backpropagation Learning in Deep Autoassociative Neural Networks
    Embrechts, Mark J.
    Hargis, Blake J.
    Linton, Jonathan D.
    2010 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS IJCNN 2010, 2010,
  • [46] A remark on the error-backpropagation learning algorithm for spiking neural networks
    Yang, Jie
    Yang, Wenyu
    Wu, Wei
    APPLIED MATHEMATICS LETTERS, 2012, 25 (08) : 1118 - 1120
  • [47] Active training of backpropagation neural networks using the learning by experimentation methodology
    Fu-Ren Lin
    Michael J. Shaw
    Annals of Operations Research, 1997, 75 : 105 - 122
  • [48] Complex-Valued Feedforward Neural Networks Learning Without Backpropagation
    Guo, Wei
    Huang, He
    Huang, Tingwen
    NEURAL INFORMATION PROCESSING (ICONIP 2017), PT IV, 2017, 10637 : 100 - 107
  • [49] A new backpropagation learning algorithm for layered neural networks with nondifferentiable units
    Oohori, Takahumi
    Naganuma, Hidenori
    Watanabe, Kazuhisa
    NEURAL COMPUTATION, 2007, 19 (05) : 1422 - 1435
  • [50] Neural network models for survival analysis using backpropagation or Bayesian learning
    Trinh, QA
    Hoang, T
    Asselain, B
    STAIRS 2002, PROCEEDINGS, 2002, 78 : 167 - 176