aCortex: An Energy-Efficient Multipurpose Mixed-Signal Inference Accelerator

被引:10
|
作者
Bavandpour, Mohammad [1 ]
Mahmoodi, Mohammad R. [1 ]
Strukov, Dmitri B. [1 ]
机构
[1] Univ Calif Santa Barbara, Dept Elect & Comp Engn, Santa Barbara, CA 93117 USA
来源
IEEE JOURNAL ON EXPLORATORY SOLID-STATE COMPUTATIONAL DEVICES AND CIRCUITS | 2020年 / 6卷 / 01期
基金
美国国家科学基金会;
关键词
Artificial neural networks; floating-gate memory; machine learning; mixed-signal circuits; neuromorphic inference accelerator; nonvolatile memory (NVM); ANALOG;
D O I
10.1109/JXCDC.2020.2999581
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
We introduce "aCortex," an extremely energy-efficient, fast, compact, and versatile neuromorphic processor architecture suitable for the acceleration of a wide range of neural network inference models. The most important feature of our processor is a configurable mixed-signal computing array of vector-by-matrix multiplier (VMM) blocks utilizing embedded nonvolatile memory arrays for storing weight matrices. Analog peripheral circuitry for data conversion and high-voltage programming are shared among a large array of VMM blocks to facilitate compact and energy-efficient analog-domain VMM operation of different types of neural network layers. Other unique features of aCortex include configurable chain of buffers and data buses, simple and efficient instruction set architecture and its corresponding multiagent controller, programmable quantization range, and a customized refresh-free embedded dynamic random access memory. The energy-optimal aCortex with 4-bit analog computing precision was designed in a 55-nm process with embedded NOR flash memory. Its physical performance was evaluated using experimental data from testing individual circuit elements and physical layout of key components for several common benchmarks, namely, Inception-vl and ResNet-152, two state-of-the-art deep feedforward networks for image classification, and GNTM, Google's deep recurrent network for language translation. The system-level simulation results for these benchmarks show the energy efficiency of 97, 106, and 336 TOp/J, respectively, combined with up to 15 TOp/s computing throughput and 0.27-MB/mm(2) storage efficiency. Such estimated performance results compare favorably with those of previously reported mixed-signal accelerators based on much less mature aggressively scaled resistive switching memories.
引用
收藏
页码:98 / 106
页数:9
相关论文
共 50 条
  • [41] Efficient Mixed-Signal Neurocomputing Via Successive Integration and Rescaling
    Bavandpour, Mohammad
    Sahay, Shubham
    Mahmoodi, Mohammad R.
    Strukov, Dmitri
    IEEE TRANSACTIONS ON VERY LARGE SCALE INTEGRATION (VLSI) SYSTEMS, 2020, 28 (03) : 823 - 827
  • [42] An efficient scheme for nonlinear modeling and predistortion in mixed-signal systems
    Koeppl, Heinz
    Singerl, Peter
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2006, 53 (12) : 1368 - 1372
  • [43] BETA: Binarized Energy-Efficient Transformer Accelerator at the Edge
    Ji, Yuhao
    Fang, Chao
    Wang, Zhongfeng
    2024 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS 2024, 2024,
  • [44] An Energy-Efficient Patchable Accelerator and Its Design Methods
    Yoshida, Hiroaki
    Wakizaka, Masayuki
    Yamashita, Shigeru
    Fujita, Masahiro
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2014, E97A (12) : 2507 - 2517
  • [45] ECHO: Energy-Efficient Computation Harnessing Online Arithmetic-An MSDF-Based Accelerator for DNN Inference
    Ibrahim, Muhammad Sohail
    Usman, Muhammad
    Lee, Jeong-A
    ELECTRONICS, 2024, 13 (10)
  • [46] Automated synthesis of mixed-signal ML inference hardware under accuracy constraints
    Kunal, Kishor
    Poojary, Jitesh
    Ramprasath, S.
    Harjani, Ramesh
    Sapatnekar, Sachin S.
    29TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE, ASP-DAC 2024, 2024, : 478 - 483
  • [47] Exploring the Design Space of an Energy-Efficient Accelerator for the SKA1-Low Central Signal Processor
    Leandro Fiorin
    Erik Vermij
    Jan van Lunteren
    Rik Jongerius
    Christoph Hagleitner
    International Journal of Parallel Programming, 2016, 44 : 1003 - 1027
  • [48] Exploring the Design Space of an Energy-Efficient Accelerator for the SKA1-Low Central Signal Processor
    Fiorin, Leandro
    Vermij, Erik
    van Lunteren, Jan
    Jongerius, Rik
    Hagleitner, Christoph
    INTERNATIONAL JOURNAL OF PARALLEL PROGRAMMING, 2016, 44 (05) : 1003 - 1027
  • [49] Mixed-signal energy feature extractor of EEG frequency bands
    Carrasco-Robles, Manuel
    Delgado-Restituto, Manuel
    2014 IEEE 5TH LATIN AMERICAN SYMPOSIUM ON CIRCUITS AND SYSTEMS (LASCAS), 2014,
  • [50] Energy-efficient deep learning inference on edge devices
    Daghero, Francesco
    Pagliari, Daniele Jahier
    Poncino, Massimo
    HARDWARE ACCELERATOR SYSTEMS FOR ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING, 2021, 122 : 247 - 301