Processing-in-Memory Accelerator for Dynamic Neural Network with Run-Time Tuning of Accuracy, Power and Latency

被引:2
|
作者
Yang, Li [1 ]
He, Zhezhi [1 ]
Angizi, Shaahin [1 ]
Fan, Deliang [1 ]
机构
[1] Arizona State Univ, Sch Elect Comp & Energy Engn, Tempe, AZ 85281 USA
基金
美国国家科学基金会;
关键词
Processing-in-Memory; Dynamic neural network;
D O I
10.1109/SOCC49529.2020.9524770
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
With the widely deployment of powerful deep neural network (DNN) into smart, but resource limited IoT devices, many prior works have been proposed to compress DNN in a hardware-aware manner to reduce the computing complexity, while maintaining accuracy, such as weight quantization, pruning, convolution decomposition, etc. However, in typical DNN compression methods, a smaller, but fixed, network structure is generated from a relative large background model for resource limited hardware accelerator deployment. However, such optimization lacks the ability to tune its structure on-the-fly to best fit for a dynamic computing hardware resource allocation and workloads. In this paper, we mainly review two of our prior works [1], [2] to address this issue, discussing how to construct a dynamic DNN structure through either uniform or non-uniform channel selection based sub-network sampling. The constructed dynamic DNN could tune its computing path to involve different number of channels, thus providing the ability to trade-off between speed, power and accuracy on-the-fly after model deployment. Correspondingly, an emerging Spin-Orbit Torque Magnetic Random-Access-Memory (SOT-MRAM) based Processing-In-Memory (PIM) accelerator will also be discussed for such dynamic neural network structure.
引用
收藏
页码:117 / 122
页数:6
相关论文
共 50 条
  • [1] Dynamic Neural Network to Enable Run-Time Trade-off between Accuracy and Latency
    Yang, Li
    Fan, Deliang
    2021 26TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC), 2021, : 587 - 592
  • [2] NeuroPIM: Felxible Neural Accelerator for Processing-in-Memory Architectures
    Bidgoli, Ali Monavari
    Fattahi, Sepideh
    Rezaei, Seyyed Hossein Seyyedaghaei
    Modarressi, Mehdi
    Daneshtalab, Masoud
    2023 26TH INTERNATIONAL SYMPOSIUM ON DESIGN AND DIAGNOSTICS OF ELECTRONIC CIRCUITS AND SYSTEMS, DDECS, 2023, : 51 - 56
  • [3] DyPIM: Dynamic-Inference-Enabled Processing-In-Memory Accelerator
    Xie, Tongxin
    Zhao, Tianchen
    Zhu, Zhenhua
    Ning, Xuefei
    Li, Bing
    Dai, Guohao
    Yang, Huazhong
    Wang, Yu
    2024 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION, DATE, 2024,
  • [4] Accelerating Neural Network Training with Processing-in-Memory GPU
    Fei, Xiang
    Han, Jianhui
    Huang, Jianqiang
    Zheng, Weimin
    Zhang, Youhui
    2022 22ND IEEE/ACM INTERNATIONAL SYMPOSIUM ON CLUSTER, CLOUD AND INTERNET COMPUTING (CCGRID 2022), 2022, : 414 - 421
  • [5] Automatic dynamic run-time optical network reservations
    Lange, JR
    Sundararaj, AI
    Dinda, PA
    14TH IEEE INTERNATIONAL SYMPOSIUM ON HIGH PERFORMANCE DISTRIBUTED COMPUTING, PROCEEDINGS, 2005, : 255 - 264
  • [6] Functionality-Based Processing-in-Memory Accelerator for Deep Convolutional Neural Networks
    Kim, Min-Jae
    Kim, Jeong-Geun
    Yoon, Su-Kyung
    Kim, Shin-Dug
    IEEE ACCESS, 2021, 9 : 145098 - 145108
  • [7] OPIMA: Optical Processing-in-Memory for Convolutional Neural Network Acceleration
    Sunny, Febin
    Shafiee, Amin
    Balasubramaniam, Abhishek
    Nikdast, Mahdi
    Pasricha, Sudeep
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2024, 43 (11) : 3888 - 3899
  • [8] CMP-PIM: An Energy-Efficient Comparator-based Processing-In-Memory Neural Network Accelerator
    Angizi, Shaahin
    He, Zhezhi
    Rakin, Adnan Siraj
    Fan, Deliang
    2018 55TH ACM/ESDA/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2018,
  • [9] ParaPIM: A Parallel Processing-in-Memory Accelerator for Binary-Weight Deep Neural Networks
    Angizi, Shaahin
    He, Zhezhi
    Fan, Deliang
    24TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE (ASP-DAC 2019), 2019, : 127 - 132
  • [10] Ultra-High-Speed Accelerator Architecture for Convolutional Neural Network Based on Processing-in-Memory Using Resistive Random Access Memory
    Wang, Hongzhe
    Wang, Junjie
    Hu, Hao
    Li, Guo
    Hu, Shaogang
    Yu, Qi
    Liu, Zhen
    Chen, Tupei
    Zhou, Shijie
    Liu, Yang
    SENSORS, 2023, 23 (05)