Efficient compiler code generation for Deep Learning Snowflake co-processor

被引:0
|
作者
Chang, Andre Xian Ming [1 ]
Zaidy, Aliasger [1 ]
Culurciello, Eugenio [1 ]
机构
[1] FWDNXT, W Lafayette, IN 47906 USA
关键词
Deep learning; neural networks; co-processor; compiler;
D O I
10.1109/EMC2.2018.00013
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Neural Networks (DNNs) are widely used in various applications including image classification, semantic segmentation and natural language processing. Various DNN models were developed to achieve high accuracy on different tasks. Efficiently mapping the workflow of those models onto custom accelerators requires a programmable hardware and a custom compiler. In this work, we use Snowflake, which is a programmable DNN targeted accelerator. We also present a compiler that correctly generated code for Snowflake. Our system were evaluated on various convolution layers present in AlexNet, ResNet and LightCNN. Snowflake with 256 processing units was implemented on Xilinx FPGA, and it achieved 70 frames/s for AlexNet without linear layers.
引用
收藏
页码:24 / 28
页数:5
相关论文
共 50 条
  • [41] Deep Learning Based Code Generation Methods: Literature Review
    Yang Z.-Z.
    Chen S.-R.
    Gao C.-Y.
    Li Z.-H.
    Li G.
    Lyu M.R.-T.
    Ruan Jian Xue Bao/Journal of Software, 2024, 35 (02): : 604 - 628
  • [42] Source-Code Generation Using Deep Learning: A Survey
    Ahmed, Areeg
    Azab, Shahira
    Abdelhamid, Yasser
    Lecture Notes in Computer Science (including subseries Lecture Notes in Artificial Intelligence and Lecture Notes in Bioinformatics), 2023, 14116 LNAI : 467 - 482
  • [43] Source-Code Generation Using Deep Learning: A Survey
    Ahmed, Areeg
    Azab, Shahira
    Abdelhamid, Yasser
    PROGRESS IN ARTIFICIAL INTELLIGENCE, EPIA 2023, PT II, 2023, 14116 : 467 - 482
  • [44] A Deep Learning Model Generation Method for Code Reuse and Automatic Machine Learning
    Lee, Keon Myung
    Hwang, Kyoung Soon
    Kim, Kwang Il
    Lee, Sang Hyun
    Park, Ki Sun
    PROCEEDINGS OF THE 2018 CONFERENCE ON RESEARCH IN ADAPTIVE AND CONVERGENT SYSTEMS (RACS 2018), 2018, : 47 - 52
  • [45] A 128 channel 290 GMACs/W Machine Learning based Co-processor for Intention Decoding in Brain Machine Interfaces
    Chen, Yi
    Yao, Enyi
    Basu, Arindam
    2015 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2015, : 3004 - 3007
  • [46] Efficient and Secure Deep Learning Inference in Trusted Processor Enabled Edge Clouds
    Li, Yuepeng
    Zeng, Deze
    Gu, Lin
    Chen, Quan
    Guo, Song
    Zomaya, Albert
    Guo, Minyi
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2022, 33 (12) : 4311 - 4325
  • [47] Deep Learning for Source Code Modeling and Generation: Models, Applications, and Challenges
    Le, Triet H. M.
    Chen, Hao
    Babar, Muhammad Ali
    ACM COMPUTING SURVEYS, 2020, 53 (03)
  • [48] Efficient and practical quantum compiler towards multi-qubit systems with deep reinforcement learning
    Chen, Qiuhao
    Du, Yuxuan
    Jiao, Yuliang
    Lu, Xiliang
    Wu, Xingyao
    Zhao, Qi
    QUANTUM SCIENCE AND TECHNOLOGY, 2024, 9 (04):
  • [49] Interpretable Deep Learning for Efficient Code Smell Prioritization in Software Development
    Rashid, Maaeda M.
    Osman, Mohd Hafeez
    Sharif, Khaironi Yatim
    Zulzalil, Hazura
    IEEE ACCESS, 2025, 13 : 45290 - 45311
  • [50] A Flexible and Energy-Efficient BLAKE-256/2s Co-Processor for Blockchain-based IoT Applications
    Luan, Pham Hoai
    Tran, Thi Hong
    Le, Vu Trung Duong
    Nakashima, Yasuhiko
    2022 35TH SBC/SBMICRO/IEEE/ACM SYMPOSIUM ON INTEGRATED CIRCUITS AND SYSTEMS DESIGN (SBCCI 2022), 2022,