Learning Student Networks via Feature Embedding

被引:52
|
作者
Chen, Hanting [1 ]
Wang, Yunhe [2 ]
Xu, Chang [3 ]
Xu, Chao [1 ]
Tao, Dacheng [3 ]
机构
[1] Peking Univ, Sch Elect Engn & Comp Sci EECS, Cooperat Medianet Innovat Ctr, Key Lab Machine Percept,Minist Educ, Beijing 100871, Peoples R China
[2] Huawei Technol Co Ltd, Noahs Ark Lab, Beijing 100085, Peoples R China
[3] Univ Sydney, Fac Engn, Sch Comp Sci, Darlington, NSW 2008, Australia
基金
中国国家自然科学基金; 澳大利亚研究理事会;
关键词
Knowledge engineering; Convolution; Training; Learning systems; Computational complexity; Graphics processing units; Mobile handsets; Deep learning; knowledge distillation (KD); teacher-student learning;
D O I
10.1109/TNNLS.2020.2970494
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep convolutional neural networks have been widely used in numerous applications, but their demanding storage and computational resource requirements prevent their applications on mobile devices. Knowledge distillation aims to optimize a portable student network by taking the knowledge from a well-trained heavy teacher network. Traditional teacher-student-based methods used to rely on additional fully connected layers to bridge intermediate layers of teacher and student networks, which brings in a large number of auxiliary parameters. In contrast, this article aims to propagate information from teacher to student without introducing new variables that need to be optimized. We regard the teacher-student paradigm from a new perspective of feature embedding. By introducing the locality preserving loss, the student network is encouraged to generate the low-dimensional features that could inherit intrinsic properties of their corresponding high-dimensional features from the teacher network. The resulting portable network, thus, can naturally maintain the performance as that of the teacher network. Theoretical analysis is provided to justify the lower computation complexity of the proposed method. Experiments on benchmark data sets and well-trained networks suggest that the proposed algorithm is superior to state-of-the-art teacher-student learning methods in terms of computational and storage complexity.
引用
收藏
页码:25 / 35
页数:11
相关论文
共 50 条
  • [1] Unsupervised Feature Selection via Collaborative Embedding Learning
    Li, Junyu
    Qi, Fei
    Sun, Xin
    Zhang, Bin
    Xu, Xiangmin
    Cai, Hongmin
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2024, 8 (03): : 2529 - 2540
  • [2] Improving Performance of Convolutional Neural Networks via Feature Embedding
    Ghoshal, Torumoy
    Zhang, Silu
    Dang, Xin
    Wilkins, Dawn
    Chen, Yixin
    PROCEEDINGS OF THE 2019 ANNUAL ACM SOUTHEAST CONFERENCE (ACMSE 2019), 2019, : 31 - 38
  • [3] Unsupervised Embedding Learning via Invariant and Spreading Instance Feature
    Ye, Mang
    Zhang, Xu
    Yuen, Pong C.
    Chang, Shih-Fu
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 6203 - 6212
  • [4] Deep Metric Learning via Lifted Structured Feature Embedding
    Song, Hyun Oh
    Xiang, Yu
    Jegelka, Stefanie
    Savarese, Silvio
    2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 4004 - 4012
  • [5] Learning and Scaling Directed Networks via Graph Embedding
    Drobyshevskiy, Mikhail
    Korshunov, Anton
    Turdakov, Denis
    MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2017, PT I, 2017, 10534 : 634 - 650
  • [6] Adversarially robust neural networks with feature uncertainty learning and label embedding
    Wang, Ran
    Ke, Haopeng
    Hu, Meng
    Wu, Wenhui
    NEURAL NETWORKS, 2024, 172
  • [7] Tag recommendation model using feature learning via word embedding
    Najafabadi, Maryam Khanian
    Nair, Madhavan A. L. Balan
    Mohamed, Azlinah
    2021 IEEE 19TH WORLD SYMPOSIUM ON APPLIED MACHINE INTELLIGENCE AND INFORMATICS (SAMI 2021), 2021, : 305 - 309
  • [8] Representation Learning for Heterogeneous Information Networks via Embedding Events
    Fu, Guoji
    Yuan, Bo
    Duan, Qiqi
    Yao, Xin
    NEURAL INFORMATION PROCESSING (ICONIP 2019), PT I, 2019, 11953 : 327 - 339
  • [9] TRAJECTORY SIMILARITY ASSESSMENT ON ROAD NETWORKS VIA EMBEDDING LEARNING
    Zhang, Rui
    Rong, Yacheng
    Wu, Zilong
    Zhuo, Yifan
    2020 IEEE SIXTH INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM 2020), 2020, : 1 - 8
  • [10] Latent Space Embedding for Unsupervised Feature Selection via Joint Dictionary Learning
    Fan, Yang
    Dai, Jianhua
    Zhang, Qilai
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,