aCortex: An Energy-Efficient Multipurpose Mixed-Signal Inference Accelerator

被引:10
|
作者
Bavandpour, Mohammad [1 ]
Mahmoodi, Mohammad R. [1 ]
Strukov, Dmitri B. [1 ]
机构
[1] Univ Calif Santa Barbara, Dept Elect & Comp Engn, Santa Barbara, CA 93117 USA
来源
IEEE JOURNAL ON EXPLORATORY SOLID-STATE COMPUTATIONAL DEVICES AND CIRCUITS | 2020年 / 6卷 / 01期
基金
美国国家科学基金会;
关键词
Artificial neural networks; floating-gate memory; machine learning; mixed-signal circuits; neuromorphic inference accelerator; nonvolatile memory (NVM); ANALOG;
D O I
10.1109/JXCDC.2020.2999581
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
We introduce "aCortex," an extremely energy-efficient, fast, compact, and versatile neuromorphic processor architecture suitable for the acceleration of a wide range of neural network inference models. The most important feature of our processor is a configurable mixed-signal computing array of vector-by-matrix multiplier (VMM) blocks utilizing embedded nonvolatile memory arrays for storing weight matrices. Analog peripheral circuitry for data conversion and high-voltage programming are shared among a large array of VMM blocks to facilitate compact and energy-efficient analog-domain VMM operation of different types of neural network layers. Other unique features of aCortex include configurable chain of buffers and data buses, simple and efficient instruction set architecture and its corresponding multiagent controller, programmable quantization range, and a customized refresh-free embedded dynamic random access memory. The energy-optimal aCortex with 4-bit analog computing precision was designed in a 55-nm process with embedded NOR flash memory. Its physical performance was evaluated using experimental data from testing individual circuit elements and physical layout of key components for several common benchmarks, namely, Inception-vl and ResNet-152, two state-of-the-art deep feedforward networks for image classification, and GNTM, Google's deep recurrent network for language translation. The system-level simulation results for these benchmarks show the energy efficiency of 97, 106, and 336 TOp/J, respectively, combined with up to 15 TOp/s computing throughput and 0.27-MB/mm(2) storage efficiency. Such estimated performance results compare favorably with those of previously reported mixed-signal accelerators based on much less mature aggressively scaled resistive switching memories.
引用
收藏
页码:98 / 106
页数:9
相关论文
共 50 条
  • [21] An Energy-Efficient Inference Engine for a Configurable ReRAM-Based Neural Network Accelerator
    Zheng, Yang-Lin
    Yang, Wei-Yi
    Chen, Ya-Shu
    Han, Ding-Hung
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (03) : 740 - 753
  • [22] Efficient simulatlon and validation for mixed-signal SOCs
    Twomey, Jerry
    EDN, 2007, 52 (07) : 65 - +
  • [23] An efficient mixed-signal architecture for minimum output energy blind multiuser detection
    Sirisuk, P
    Worapishet, A
    Tanoi, S
    PROCEEDINGS OF THE 2003 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOL I: ANALOG CIRCUITS AND SIGNAL PROCESSING, 2003, : 765 - 768
  • [24] Nonlinearity Modeling for Mixed-Signal Inference Accelerators in Training Frameworks
    Conrad, Joschua
    Jiang, Biyi
    Kaesser, Paul
    Belagiannis, Vasileios
    Ortmanns, Maurits
    2021 28TH IEEE INTERNATIONAL CONFERENCE ON ELECTRONICS, CIRCUITS, AND SYSTEMS (IEEE ICECS 2021), 2021,
  • [25] A Mixed-signal Time-Domain Generative Adversarial Network Accelerator with Efficient Subthreshold Time Multiplier and Mixed-signal On-chip Training for Low Power Edge Devices
    Chen, Zhengyu
    Fu, Sihua
    Cao, Qiankai
    Gu, Jie
    2020 IEEE SYMPOSIUM ON VLSI CIRCUITS, 2020,
  • [26] A Mixed-Signal Quantized Neural Network Accelerator Using Flash Transistors
    Scott, Kyler R.
    Lee, Cheng-Yen
    Khatri, Sunil P.
    Vrudhula, Sarma
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS I-REGULAR PAPERS, 2024, 71 (03) : 1025 - 1038
  • [27] Harmonica: Hybrid Accelerator to Overcome Imperfections of Mixed-signal DNN Accelerators
    Behnam, Payman
    Kamal, Uday
    Shafiee, Ali
    Tumanov, Alexey
    Mukhopadhyay, Saibal
    PROCEEDINGS 2024 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM, IPDPS 2024, 2024, : 619 - 630
  • [28] SONIC: A Sparse Neural Network Inference Accelerator with Silicon Photonics for Energy-Efficient Deep Learning
    Sunny, Febin
    Nikdast, Mandi
    Pasricha, Sudeep
    27TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE, ASP-DAC 2022, 2022, : 214 - 219
  • [29] MSP-MFCC: Energy-Efficient MFCC Feature Extraction Method With Mixed-Signal Processing Architecture for Wearable Speech Recognition Applications
    Li, Qin
    Yang, Yuze
    Lan, Tianxiang
    Zhu, Huifeng
    Wei, Qi
    Qiao, Fei
    Liu, Xinjun
    Yang, Huazhong
    IEEE ACCESS, 2020, 8 : 48720 - 48730
  • [30] Efficient DFT Strategy for Complex Mixed-Signal SoC
    Hu Jian
    Shen Xubang
    2008 INTERNATIONAL CONFERENCE ON EMBEDDED SOFTWARE AND SYSTEMS SYMPOSIA, PROCEEDINGS, 2008, : 425 - 430