aCortex: An Energy-Efficient Multipurpose Mixed-Signal Inference Accelerator

被引:10
|
作者
Bavandpour, Mohammad [1 ]
Mahmoodi, Mohammad R. [1 ]
Strukov, Dmitri B. [1 ]
机构
[1] Univ Calif Santa Barbara, Dept Elect & Comp Engn, Santa Barbara, CA 93117 USA
来源
IEEE JOURNAL ON EXPLORATORY SOLID-STATE COMPUTATIONAL DEVICES AND CIRCUITS | 2020年 / 6卷 / 01期
基金
美国国家科学基金会;
关键词
Artificial neural networks; floating-gate memory; machine learning; mixed-signal circuits; neuromorphic inference accelerator; nonvolatile memory (NVM); ANALOG;
D O I
10.1109/JXCDC.2020.2999581
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
We introduce "aCortex," an extremely energy-efficient, fast, compact, and versatile neuromorphic processor architecture suitable for the acceleration of a wide range of neural network inference models. The most important feature of our processor is a configurable mixed-signal computing array of vector-by-matrix multiplier (VMM) blocks utilizing embedded nonvolatile memory arrays for storing weight matrices. Analog peripheral circuitry for data conversion and high-voltage programming are shared among a large array of VMM blocks to facilitate compact and energy-efficient analog-domain VMM operation of different types of neural network layers. Other unique features of aCortex include configurable chain of buffers and data buses, simple and efficient instruction set architecture and its corresponding multiagent controller, programmable quantization range, and a customized refresh-free embedded dynamic random access memory. The energy-optimal aCortex with 4-bit analog computing precision was designed in a 55-nm process with embedded NOR flash memory. Its physical performance was evaluated using experimental data from testing individual circuit elements and physical layout of key components for several common benchmarks, namely, Inception-vl and ResNet-152, two state-of-the-art deep feedforward networks for image classification, and GNTM, Google's deep recurrent network for language translation. The system-level simulation results for these benchmarks show the energy efficiency of 97, 106, and 336 TOp/J, respectively, combined with up to 15 TOp/s computing throughput and 0.27-MB/mm(2) storage efficiency. Such estimated performance results compare favorably with those of previously reported mixed-signal accelerators based on much less mature aggressively scaled resistive switching memories.
引用
收藏
页码:98 / 106
页数:9
相关论文
共 50 条
  • [31] Efficient analysis of mixed-signal ASICs for smart sensors
    Kerö, NE
    Sauter, T
    14TH IEEE INTERNATIONAL WORKSHOP ON RAPID SYSTEMS PROTOTYPING, PROCEEDINGS: SHORTENING THE PATH FROM SPECIFICATION TO PROTOTYPE, 2003, : 40 - 46
  • [32] Mixed-Signal Neuromorphic Inference Accelerators: Recent Results and Future Prospects
    Bavandpour, M.
    Mahmoodi, M. R.
    Nili, H.
    Bayat, F. Merrikh
    Prezioso, M.
    Vincent, A.
    Strukov, D. B.
    Likharev, K. K.
    2018 IEEE INTERNATIONAL ELECTRON DEVICES MEETING (IEDM), 2018,
  • [33] Analog/Mixed-Signal Hardware Error Modeling for Deep Learning Inference
    Rekhi, Angad S.
    Zimmer, Brian
    Nedovic, Nikola
    Liu, Ningxi
    Venkatesan, Rangharajan
    Wang, Miaorong
    Khailany, Brucek
    Dally, William J.
    Gray, C. Thomas
    PROCEEDINGS OF THE 2019 56TH ACM/EDAC/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2019,
  • [34] Energy-Efficient Embedded Inference of SVMs on FPGA
    Elgawi, Osman
    Mutawa, A. M.
    Ahmad, Afaq
    2019 IEEE COMPUTER SOCIETY ANNUAL SYMPOSIUM ON VLSI (ISVLSI 2019), 2019, : 165 - 169
  • [35] Energy-Efficient Approximate Edge Inference Systems
    Ghosh, Soumendu Kumar
    Raha, Arnab
    Raghunathan, Vijay
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2023, 22 (04)
  • [36] Efficient Moment Estimation with Extremely Small Sample Size via Bayesian Inference for Analog/Mixed-Signal Validation
    Gu, Chenjie
    Chiprout, Eli
    Li, Xin
    2013 50TH ACM / EDAC / IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2013,
  • [37] OPTICAL AND ENERGY-EFFICIENT OF SIGNAL LIGHTS
    LEWIN, I
    HEINISCH, RV
    LIGHTING DESIGN & APPLICATION, 1985, 15 (04): : 49 - 49
  • [38] A Database Accelerator for Energy-Efficient Query Processing and Optimization
    Haas, Sebastian
    Arnold, Oliver
    Scholze, Stefan
    Hoeppner, Sebastian
    Ellguth, Georg
    Dixius, Andreas
    Ungethuem, Annett
    Mier, Eric
    Noethen, Benedikt
    Matus, Emil
    Schiefer, Stefan
    Cederstroem, Love
    Pilz, Fabian
    Mayr, Christian
    Schueffny, Rene
    Lehner, Wolfgang
    Fettweis, Gerhard P.
    2016 2ND IEEE NORDIC CIRCUITS AND SYSTEMS CONFERENCE (NORCAS), 2016,
  • [39] An Energy-Efficient Deep Neural Network Accelerator Design
    Jung, Jueun
    Lee, Kyuho Jason
    2020 54TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2020, : 272 - 276
  • [40] An Asynchronous Energy-Efficient CNN Accelerator with Reconfigurable Architecture
    Chen, Weijia
    Wu, Hui
    Wei, Shaojun
    He, Anping
    Chen, Hong
    2018 IEEE ASIAN SOLID-STATE CIRCUITS CONFERENCE (A-SSCC): PROCEEDINGS OF TECHNICAL PAPERS, 2018, : 51 - 54