Scaling for edge inference of deep neural networks

被引:301
|
作者
Xu, Xiaowei [1 ]
Ding, Yukun [1 ]
Hu, Sharon Xiaobo [1 ]
Niemier, Michael [1 ]
Cong, Jason [2 ]
Hu, Yu [3 ]
Shi, Yiyu [1 ]
机构
[1] Univ Notre Dame, Dept Comp Sci, Notre Dame, IN 46556 USA
[2] Univ Calif Los Angeles, Dept Comp Sci, Los Angeles, CA 90024 USA
[3] Huazhong Univ Sci & Technol, Sch Opt & Elect Informat, Wuhan, Hubei, Peoples R China
来源
NATURE ELECTRONICS | 2018年 / 1卷 / 04期
关键词
ENERGY;
D O I
10.1038/s41928-018-0059-3
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep neural networks offer considerable potential across a range of applications, from advanced manufacturing to autonomous cars. A clear trend in deep neural networks is the exponential growth of network size and the associated increases in computational complexity and memory consumption. However, the performance and energy efficiency of edge inference, in which the inference (the application of a trained network to new data) is performed locally on embedded platforms that have limited area and power budget, is bounded by technology scaling. Here we analyse recent data and show that there are increasing gaps between the computational complexity and energy efficiency required by data scientists and the hardware capacity made available by hardware architects. We then discuss various architecture and algorithm innovations that could help to bridge the gaps.
引用
收藏
页码:216 / 222
页数:7
相关论文
共 50 条
  • [21] Parameter inference with deep jointly informed neural networks
    Humbird, Kelli D.
    Peterson, J. Luc
    McClarren, Ryan G.
    STATISTICAL ANALYSIS AND DATA MINING, 2019, 12 (06) : 496 - 504
  • [22] ProteInfer, deep neural networks for protein functional inference
    Sanderson, Theo
    Bileschi, Maxwell L.
    Belanger, David
    Colwell, Lucy J.
    Doetsch, Volker
    ELIFE, 2023, 12
  • [23] Trigonometric Inference Providing Learning in Deep Neural Networks
    Cai, Jingyong
    Takemoto, Masashi
    Qiu, Yuming
    Nakajo, Hironori
    APPLIED SCIENCES-BASEL, 2021, 11 (15):
  • [24] Optimization of Analog Accelerators for Deep Neural Networks Inference
    Fasoli, Andrea
    Ambrogio, Stefano
    Narayanan, Pritish
    Tsai, Hsinyu
    Mackin, Charles
    Spoon, Katherine
    Friz, Alexander
    Chen, An
    Burr, Geoffrey W.
    2020 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2020,
  • [25] Reducing Memory Requirements of Convolutional Neural Networks for Inference at the Edge
    Bravenec, Tomas
    Fryza, Tomas
    2021 31ST INTERNATIONAL CONFERENCE RADIOELEKTRONIKA (RADIOELEKTRONIKA), 2021,
  • [26] COMPRESSING DEEP NEURAL NETWORKS FOR EFFICIENT VISUAL INFERENCE
    Ge, Shiming
    Luo, Zhao
    Zhao, Shengwei
    Jin, Xin
    Zhang, Xiao-Yu
    2017 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), 2017, : 667 - 672
  • [27] Commutative Scaling of Width and Depth in Deep Neural Networks
    Hayou, Soufiane
    JOURNAL OF MACHINE LEARNING RESEARCH, 2024, 25
  • [28] Fast and accurate edge resource scaling for 5G/6G networks with distributed deep neural networks
    Giannakas, Theodoros
    Spyropoulos, Thrasyvoulos
    Smid, Ondrej
    2022 IEEE 23RD INTERNATIONAL SYMPOSIUM ON A WORLD OF WIRELESS, MOBILE AND MULTIMEDIA NETWORKS (WOWMOM 2022), 2022, : 100 - 109
  • [29] Modeling of Deep Neural Network (DNN) Placement and Inference in Edge Computing
    Bensalem, Mounir
    Dizdarevic, Jasenka
    Jukan, Admela
    2020 IEEE INTERNATIONAL CONFERENCE ON COMMUNICATIONS WORKSHOPS (ICC WORKSHOPS), 2020,
  • [30] DeepEdgeBench: Benchmarking Deep Neural Networks on Edge Devices
    Baller, Stephan Patrick
    Jindal, Anshul
    Chadha, Mohak
    Gerndt, Michael
    2021 IEEE INTERNATIONAL CONFERENCE ON CLOUD ENGINEERING, IC2E 2021, 2021, : 20 - 30