Optimizing Memory Efficiency for Deep Convolutional Neural Network Accelerators

被引:1
|
作者
Li, Xiaowei [1 ]
Li, Jiajun
Yan, Guihai [1 ]
机构
[1] Chinese Acad Sci, Inst Comp Technol, State Key Lab Comp Architecture, Beijing, Peoples R China
基金
中国国家自然科学基金;
关键词
Deep Convolutional Neural Networks; Accelerator Architecture; Memory Efficiency;
D O I
10.1166/jolpe.2018.1580
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Convolutional Neural Network (CNN) accelerators have achieved nominal performance and energy efficiency speedup compared to traditional general purpose CPU- and GPU-based solutions. Although optimizations on computation have been intensively studied, the energy efficiency of such accelerators remains limited by off-chip memory accesses since their energy cost is magnitudes higher than other operations. Minimizing off-chip memory access volume, therefore, is the key to further improving energy efficiency. The prior state-of-the-art uses rigid data reuse patterns and is sub-optimal for some, or even all, of the individual convolutional layers. To overcome the problem, this paper proposed an adaptive layer partitioning and scheduling scheme, called SmartShuttle, to minimize off-chip memory accesses for CNN accelerators. Smartshuttle can adaptively switch among different data reuse schemes and the corresponding tiling factor settings to dynamically match different convolutional layers and fully-connected layers. Moreover, SmartShuttle thoroughly investigates the impact of data reusability and sparsity on the memory access volume. The experimental results show that SmartShuttle processes the convolutional layers at 434.8 multiply and accumulations (MACs)/DRAM access for VGG16 (batch size = 3), and 526.3 MACs/DRAM access for AlexNet (batch size = 4), which outperforms the state-of-the-art approach (Eyeriss) by 52.2% and 52.6%, respectively.
引用
收藏
页码:496 / 507
页数:12
相关论文
共 50 条
  • [21] Reduced Memory Region Based Deep Convolutional Neural Network Detection
    Tome, Denis
    Bondi, Luca
    Baroffio, Luca
    Tubaro, Stefano
    Plebani, Emanuele
    Pau, Danilo
    2016 IEEE 6TH INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS - BERLIN (ICCE-BERLIN), 2016,
  • [22] Memory Trojan Attack on Neural Network Accelerators
    Zhao, Yang
    Hu, Xing
    Li, Shuangchen
    Ye, Jing
    Deng, Lei
    Ji, Yu
    Xu, Jianyu
    Wu, Dong
    Xie, Yuan
    2019 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE), 2019, : 1415 - 1420
  • [23] Small Memory Footprint Neural Network Accelerators
    Seto, Kenshu
    Nejatollahi, Hamid
    An, Jiyoung
    Kang, Sujin
    Dutt, Nikil
    PROCEEDINGS OF THE 2019 20TH INTERNATIONAL SYMPOSIUM ON QUALITY ELECTRONIC DESIGN (ISQED), 2019, : 253 - 258
  • [24] Design Tradeoff of Internal Memory Size and Memory Access Energy in Deep Neural Network Hardware Accelerators
    Hsiao, Shen-Fu
    Wu, Pei-Hsuen
    2018 IEEE 7TH GLOBAL CONFERENCE ON CONSUMER ELECTRONICS (GCCE 2018), 2018, : 735 - 736
  • [25] Review of ASIC accelerators for deep neural network
    Machupalli, Raju
    Hossain, Masum
    Mandal, Mrinal
    MICROPROCESSORS AND MICROSYSTEMS, 2022, 89
  • [26] Approximate Adders for Deep Neural Network Accelerators
    Raghuram, S.
    Shashank, N.
    2022 35TH INTERNATIONAL CONFERENCE ON VLSI DESIGN (VLSID 2022) HELD CONCURRENTLY WITH 2022 21ST INTERNATIONAL CONFERENCE ON EMBEDDED SYSTEMS (ES 2022), 2022, : 210 - 215
  • [27] Optimizing Temporal Convolutional Network Inference on FPGA-Based Accelerators
    Carreras, Marco
    Deriu, Gianfranco
    Raffo, Luigi
    Benini, Luca
    Meloni, Paolo
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2020, 10 (03) : 348 - 361
  • [28] POMMEL: Exploring Off-Chip Memory Energy & Power Consumption in Convolutional Neural Network Accelerators
    Montgomerie-Corcoran, Alexander
    Bouganis, Christos-Savvas
    2021 24TH EUROMICRO CONFERENCE ON DIGITAL SYSTEM DESIGN (DSD 2021), 2021, : 442 - 448
  • [29] Latency-Insensitive Controller for Convolutional Neural Network Accelerators
    Seo, Youngho
    Lee, Sanghun
    Kim, Sunwoo
    Wang, Jooho
    Park, Sungkyung
    Park, Chester Sungchung
    2019 INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2019, : 249 - 250
  • [30] Optimizing Off-Chip Memory Access for Deep Neural Network Accelerator
    Zheng, Yong
    Yang, Haigang
    Shu, Yi
    Jia, Yiping
    Huang, Zhihong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2022, 69 (04) : 2316 - 2320