A note on factor normalization for deep neural network models

被引:0
|
作者
Haobo Qi
Jing Zhou
Hansheng Wang
机构
[1] Peking University,Guanghua School of Management
[2] Renmin University of China,Center for Applied Statistics and School of Statistics
来源
关键词
D O I
暂无
中图分类号
学科分类号
摘要
Deep neural network (DNN) models often involve high-dimensional features. In most cases, these high-dimensional features can be decomposed into two parts: a low-dimensional factor and residual features with much-reduced variability and inter-feature correlation. This decomposition has several interesting theoretical implications for DNN training. Based on these implications, we develop a novel factor normalization method for better performance. The proposed method leads to a new deep learning model with two important characteristics. First, it allows factor-related feature extraction, and second, it allows for adaptive learning rates for factors and residuals. These model features improve the convergence speed on both training and testing datasets. Multiple empirical experiments are presented to demonstrate the model’s superior performance.
引用
收藏
相关论文
共 50 条
  • [41] Scalable planning with deep neural network learned transition models
    Wu, Ga
    Say, Buser
    Sanner, Scott
    Journal of Artificial Intelligence Research, 2020, 68 : 571 - 606
  • [42] Cascaded deep neural network models for dialog state tracking
    Yang, Guohua
    Wang, Xiaojie
    MULTIMEDIA TOOLS AND APPLICATIONS, 2019, 78 (08) : 9625 - 9643
  • [43] Cascaded deep neural network models for dialog state tracking
    Guohua Yang
    Xiaojie Wang
    Multimedia Tools and Applications, 2019, 78 : 9625 - 9643
  • [44] Deep Neural Network Models for the Recognition of Traffic Signs Defects
    Nagy, Amr M.
    Czuni, Laszlo
    PROCEEDINGS OF THE 11TH IEEE INTERNATIONAL CONFERENCE ON INTELLIGENT DATA ACQUISITION AND ADVANCED COMPUTING SYSTEMS: TECHNOLOGY AND APPLICATIONS (IDAACS'2021), VOL 2, 2021, : 725 - 729
  • [45] Privacy preserving layer partitioning for Deep Neural Network models
    Rajasekar, Kishore
    Loh, Randolph
    Fok, Kar Wai
    Thing, Vrizlynn L. L.
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 1129 - 1135
  • [46] Scalable Planning with Deep Neural Network Learned Transition Models
    Wu, Ga
    Say, Buser
    Sanner, Scott
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2020, 68 : 571 - 606
  • [47] Analysis of Deep Neural Network Models for Acoustic Scene Classification
    Basbug, Ahmet Melih
    Sert, Mustafa
    2019 27TH SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2019,
  • [48] Job Prediction: From Deep Neural Network Models to Applications
    Tin Van Huynh
    Kiet Van Nguyen
    Ngan Luu-Thuy Nguyen
    Anh Gia-Tuan Nguyen
    2020 RIVF INTERNATIONAL CONFERENCE ON COMPUTING & COMMUNICATION TECHNOLOGIES (RIVF 2020), 2020, : 53 - 58
  • [49] Explaining Deep Neural Network Models with Adversarial Gradient Integration
    Pan, Deng
    Li, Xin
    Zhu, Dongxiao
    PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 2876 - 2883
  • [50] Application of Deep Neural Network Models in Drug Discovery Programs
    Grebner, Christoph
    Matter, Hans
    Kofink, Daniel
    Wenzel, Jan
    Schmidt, Friedemann
    Hessler, Gerhard
    CHEMMEDCHEM, 2021, 16 (24) : 3772 - 3786