Better Depth-Width Trade-offs for Neural Networks through the lens of Dynamical Systems

被引:0
|
作者
Chatziafratis, Vaggos [1 ]
Nagarajan, Sai Ganesh [2 ]
Panageas, Ioannis [2 ]
机构
[1] Stanford Univ, Dept Comp Sci, Stanford, CA 94305 USA
[2] Singapore Univ Technol & Design, Singapore, Singapore
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The expressivity of neural networks as a function of their depth, width and type of activation units has been an important question in deep learning theory. Recently, depth separation results for ReLU networks were obtained via a new connection with dynamical systems, using a generalized notion of fixed points of a continuous map f, called periodic points. In this work, we strengthen the connection with dynamical systems and we improve the existing width lower bounds along several aspects. Our first main result is period-specific width lower bounds that hold under the stronger notion of L-1-approximation error, instead of the weaker classification error. Our second contribution is that we provide sharper width lower bounds, still yielding meaningful exponential depth-width separations, in regimes where previous results wouldn't apply. A byproduct of our results is that there exists a universal constant characterizing the depth-width trade-offs, as long as f has odd periods. Technically, our results follow by unveiling a tighter connection between the following three quantities of a given function: its period, its Lipschitz constant and the growth rate of the number of oscillations arising under compositions of the function f with itself.
引用
收藏
页数:10
相关论文
共 42 条
  • [1] Better Depth-Width Trade-offs for Neural Networks through the lens of Dynamical Systems
    Chatziafratis, Vaggos
    Nagarajan, Sai Ganesh
    Panageas, Ioannis
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 119, 2020, 119
  • [2] Trade-Offs Between Energy and Depth of Neural Networks
    Uchizawa, Kei
    Abe, Haruki
    NEURAL COMPUTATION, 2024, 36 (08) : 1541 - 1567
  • [3] TRADE-OFFS BETWEEN DEPTH AND WIDTH IN PARALLEL COMPUTATION
    VISHKIN, U
    WIGDERSON, A
    SIAM JOURNAL ON COMPUTING, 1985, 14 (02) : 303 - 314
  • [4] Depth-Width Tradeoffs in Approximating Natural Functions with Neural Networks
    Safran, Itay
    Shamir, Ohad
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 70, 2017, 70
  • [5] Fast and Accurate Binary Neural Networks based on Depth-Width Reshaping
    Xue, Ping
    Lu, Yang
    Chang, Jingfei
    Wei, Xing
    Wei, Zhen
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 9, 2023, : 10684 - 10692
  • [6] Exploring Trade-Offs in Spiking Neural Networks
    Bacho F.
    Chu D.
    Neural Computation, 2023, 35 (10) : 1627 - 1656
  • [7] Performance improvements seen through the lens of strategic trade-offs
    Sarmiento, Roberto
    Shukla, Vinaya
    Manuel Izar-Landeta, Juan
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2013, 51 (15) : 4682 - 4694
  • [8] Trade-Offs in Marine Policy Decisions Through the Lens of Literature
    de Queiroz, Joyce Dias Gois Rodrigues
    Gutierrez, Debora
    Calado, Helena Maria Gregorio Pina
    OCEANS-SWITZERLAND, 2024, 5 (04): : 982 - 1007
  • [9] Pigmentation and fitness trade-offs through the lens of artificial selection
    Rajpurohit, Subhash
    Richardson, Rani
    Dean, John
    Vazquez, Raul
    Wong, Grace
    Schmidt, Paul S.
    BIOLOGY LETTERS, 2016, 12 (10)
  • [10] Trade-offs in the distribution of neural networks in a wireless sensor network
    Holenderski, M
    Lukkien, J
    Khong, TC
    PROCEEDINGS OF THE 2005 INTELLIGENT SENSORS, SENSOR NETWORKS & INFORMATION PROCESSING CONFERENCE, 2005, : 259 - 264