Efficient Implementation of Activation Functions for LSTM accelerators

被引:5
|
作者
Chong, Yi Sheng [1 ,2 ]
Goh, Wang Ling [1 ]
Ong, Yew Soon [3 ]
Nambiar, Vishnu P. [4 ]
Anh Tuan Do [4 ]
机构
[1] Nanyang Technol Univ NTU, Sch Elect & Elect Engn, Singapore, Singapore
[2] Nanyang Technol Univ, Interdisciplinary Grad Programme, Energy Res Inst, Singapore, Singapore
[3] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore, Singapore
[4] ASTAR, Inst Microelect, Singapore, Singapore
关键词
DESIGN;
D O I
10.1109/VLSI-SoC53125.2021.9606971
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Activation functions such as hyperbolic tangent (tanh) and logistic sigmoid (sigmoid) are critical computing elements in a long short term memory (LSTM) cell and network. These activation functions are non-linear, leading to challenges in their hardware implementations. Area-efficient and high performance hardware implementation of these activation functions thus becomes crucial to allow high throughput in a LSTM accelerator. In this work, we propose an approximation scheme which is suitable for both tanh and sigmoid functions. The proposed hardware for sigmoid function is 8.3 times smaller than the state-of-the-art, while for tanh function, it is the second smallest design. When applying the approximated tanh and sigmoid of 2% error in a LSTM cell computation, its final hidden state and cell state record errors of 3.1% and 5.8% respectively. When the same approximated functions are applied to a single layer LSTM network of 64 hidden nodes, the accuracy drops by 2.8% only. This proposed small yet accurate activation function hardware is promising to be used in Internet of Things (IoT) applications where accuracy can be traded off for ultra-low power consumption.
引用
收藏
页码:19 / 23
页数:5
相关论文
共 50 条
  • [31] DESIGN AUTOMATION TOOLS FOR EFFICIENT IMPLEMENTATION OF LOGIC FUNCTIONS BY DECOMPOSITION
    VARMA, D
    TRACHTENBERG, EA
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 1989, 8 (08) : 901 - 916
  • [32] AN EFFICIENT IMPLEMENTATION OF BOOLEAN FUNCTIONS AS SELF-TIMED CIRCUITS
    DAVID, I
    GINOSAR, R
    YOELI, M
    IEEE TRANSACTIONS ON COMPUTERS, 1992, 41 (01) : 2 - 11
  • [33] Efficient implementation of generalized Maiorana–McFarland class of cryptographic functions
    Pasalic E.
    Chattopadhyay A.
    Zhang W.G.
    Journal of Cryptographic Engineering, 2017, 7 (4) : 287 - 295
  • [34] Efficient Implementation of Activation Function on FPGA for Accelerating Neural Networks
    Qian, Kai
    Liu, Yinqiu
    Zhang, Zexu
    Wang, Kun
    2023 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, ISCAS, 2023,
  • [35] Crypto Accelerators for Power-Efficient and Real-Time on-Chip Implementation of Secure Algorithms
    Baldanzi, Luca
    Crocetti, Luca
    Di Matteo, Stefano
    Fanucci, Luca
    Saponara, Sergio
    Hameau, Patrice
    2019 26TH IEEE INTERNATIONAL CONFERENCE ON ELECTRONICS, CIRCUITS AND SYSTEMS (ICECS), 2019, : 775 - 778
  • [36] Balancing Computation Loads and Optimizing Input Vector Loading in LSTM Accelerators
    Park, Junki
    Yi, Wooseok
    Ahn, Daehyun
    Kung, Jaeha
    Kim, Jae-Joon
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2020, 39 (09) : 1889 - 1901
  • [37] HALF LIFE PERIODS OF ACTIVATION IN ACCELERATORS
    FREYTAG, E
    HEALTH PHYSICS, 1968, 14 (03): : 267 - +
  • [38] MSBF-LSTM: Most-significant Bit-first LSTM Accelerators with Energy Efficiency Optimisations
    Bian, Sige
    Li, He
    Wang, Chengcheng
    Song, Changjun
    Tang, Yongming
    2023 IEEE 31ST ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES, FCCM, 2023, : 218 - 218
  • [39] EIS - Efficient and Trainable Activation Functions for Better Accuracy and Performance
    Biswas, Koushik
    Kumar, Sandeep
    Banerjee, Shilpak
    Pandey, Ashish Kumar
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT II, 2021, 12892 : 260 - 272
  • [40] Efficient Neural Network Robustness Certification with General Activation Functions
    Zhang, Huan
    Weng, Tsui-Wei
    Chen, Pin-Yu
    Hsieh, Cho-Jui
    Daniel, Luca
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31