Approximating functions with multi-features by deep convolutional neural networks

被引:35
|
作者
Mao, Tong [1 ]
Shi, Zhongjie [2 ]
Zhou, Ding-Xuan [3 ]
机构
[1] Claremont Grad Univ, Inst Math Sci, 710 N Coll Ave, Claremont, CA 91711 USA
[2] Katholieke Univ Leuven, Dept Elect Engn, ESAT STADIUS, Kasteelpk Arenberg 10, B-3001 Leuven, Belgium
[3] Univ Sydney, Sch Math & Stat, Sydney, NSW 2006, Australia
基金
美国国家科学基金会;
关键词
Deep learning; convolutional neural networks; rates of approximation; curse of dimensionality; feature extraction; ERROR-BOUNDS;
D O I
10.1142/S0219530522400085
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
Deep convolutional neural networks (DCNNs) have achieved great empirical success in many fields such as natural language processing, computer vision, and pattern recognition. But there still lacks theoretical understanding of the flexibility and adaptivity of DCNNs in various learning tasks, and the power of DCNNs at feature extraction. We propose a generic DCNN structure consisting of two groups of convolutional layers associated with two downsampling operators, and a fully connected layer, which is determined only by three structural parameters. Our generic DCNNs are capable of extracting various features including not only polynomial features but also general smooth features. We also show that the curse of dimensionality can be circumvented by our DCNNs for target functions of the compositional form with (symmetric) polynomial features, spatially sparse smooth features, and interaction features. These demonstrate the expressive power of our DCNN structure, while the model selection can be relaxed comparing with other deep neural networks since there are only three hyperparameters controlling the architecture to tune.
引用
收藏
页码:93 / 125
页数:33
相关论文
共 50 条
  • [31] Rethinking the Role of Activation Functions in Deep Convolutional Neural Networks for Image Classification
    Zheng, Qinghe
    Yang, Mingqiang
    Tian, Xinyu
    Wang, Xiaochen
    Wang, Deqiang
    ENGINEERING LETTERS, 2020, 28 (01) : 80 - 92
  • [32] Deep Convolutional Neural Networks for Multi-Instance Multi-Task Learning
    Zeng, Tao
    Ji, Shuiwang
    2015 IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM), 2015, : 579 - 588
  • [33] Deep Anchored Convolutional Neural Networks
    Huang, Jiahui
    Dwivedi, Kshitij
    Roig, Gemma
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2019), 2019, : 639 - 647
  • [34] DEEP CONVOLUTIONAL NEURAL NETWORKS FOR LVCSR
    Sainath, Tara N.
    Mohamed, Abdel-rahman
    Kingsbury, Brian
    Ramabhadran, Bhuvana
    2013 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2013, : 8614 - 8618
  • [35] Deep Unitary Convolutional Neural Networks
    Chang, Hao-Yuan
    Wang, Kang L.
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT II, 2021, 12892 : 170 - 181
  • [36] Approximating Lipschitz continuous functions with GroupSort neural networks
    Tanielian, U.
    Sangnier, M.
    Biau, G.
    24TH INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS (AISTATS), 2021, 130 : 442 - +
  • [37] Universality of deep convolutional neural networks
    Zhou, Ding-Xuan
    APPLIED AND COMPUTATIONAL HARMONIC ANALYSIS, 2020, 48 (02) : 787 - 794
  • [38] A Review on Deep Convolutional Neural Networks
    Aloysius, Neena
    Geetha, M.
    2017 INTERNATIONAL CONFERENCE ON COMMUNICATION AND SIGNAL PROCESSING (ICCSP), 2017, : 588 - 592
  • [39] Spatial deep convolutional neural networks
    Wang, Qi
    Parker, Paul A.
    Lund, Robert
    SPATIAL STATISTICS, 2025, 66
  • [40] Convergence of deep convolutional neural networks
    Xu, Yuesheng
    Zhang, Haizhang
    NEURAL NETWORKS, 2022, 153 : 553 - 563