Deep neural networks compiler for a trace-based accelerator

被引:2
|
作者
Chang, Andre Xian Ming [1 ]
Zaidy, Aliasger [1 ]
Vitez, Marko [1 ]
Burzawa, Lukasz [1 ]
Culurciello, Eugenio [1 ]
机构
[1] FWDNXT Inc, 1281 Win Hentschel Blvd, W Lafayette, IN 47906 USA
关键词
DNN; Compiler; Accelerator;
D O I
10.1016/j.sysarc.2019.101659
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Convolutional Neural Networks (CNNs) are the algorithm of choice for image processing applications. CNNs are a highly parallel workload that leads to the emergence of custom hardware accelerators. Deep Learning (DL) models specialized in different tasks require programmable custom hardware and a compiler/mapper to efficiently translate different CNNs into an efficient dataflow in the accelerator. The goal of this paper is to present a compiler for running CNNs on programmable custom hardware accelerators with a domain-specific ISA that targets CNNs. In this work, the compiler was evaluated and tested on a hardware accelerator that was presented in [18]. The compiler uses model definition files created from popular frameworks to generate custom instructions. The model goes through static compilation and different levels of hardware aware optimizations that improve performance and data reuse of the generated program. The software also exposes an interface to run on various FPGA platforms, providing an end-to-end solution. Various CNN models were benchmarked on different systems while scaling the number of processing units.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Embedded Streaming Deep Neural Networks Accelerator With Applications
    Dundar, Aysegul
    Jin, Jonghoon
    Martini, Berin
    Culurciello, Eugenio
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2017, 28 (07) : 1572 - 1583
  • [32] Optimizing Accelerator on FPGA for Deep Convolutional Neural Networks
    Dong, Yong
    Hu, Wei
    Wang, Yonghao
    Jiao, Qiang
    Chen, Shuang
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT II, 2020, 12453 : 97 - 110
  • [33] BSHIFT: A Low Cost Deep Neural Networks Accelerator
    Yong Yu
    Tian Zhi
    Xuda Zhou
    Shaoli Liu
    Yunji Chen
    Shuyao Cheng
    International Journal of Parallel Programming, 2019, 47 : 360 - 372
  • [34] TraceGra: A trace-based anomaly detection for microservice using graph deep learning
    Chen, Jian
    Liu, Fagui
    Jiang, Jun
    Zhong, Guoxiang
    Xu, Dishi
    Tan, Zhuanglun
    Shi, Shangsong
    COMPUTER COMMUNICATIONS, 2023, 204 : 109 - 117
  • [35] A trace-based model for multiparty contracts
    Hvitved, Tom
    Klaedtke, Felix
    Zalinescu, Eugen
    JOURNAL OF LOGIC AND ALGEBRAIC PROGRAMMING, 2012, 81 (02): : 72 - 98
  • [36] Trace-Based Workload Generation and Execution
    Sfakianakis, Yannis
    Kanellou, Eleni
    Marazakis, Manolis
    Bilas, Angelos
    EURO-PAR 2021: PARALLEL PROCESSING, 2021, 12820 : 37 - 54
  • [37] Trace-based adaptive help system
    Sehaba, Karim
    INTERNATIONAL JOURNAL OF TECHNOLOGIES IN HIGHER EDUCATION, 2012, 9 (03): : 55 - 70
  • [38] NxTF: An API and Compiler for Deep Spiking Neural Networks on Intel Loihi
    Rueckauer, Bodo
    Bybee, Connor
    Goettsche, Ralf
    Singh, Yashwardhan
    Mishra, Joyesh
    Wild, Andreas
    ACM JOURNAL ON EMERGING TECHNOLOGIES IN COMPUTING SYSTEMS, 2022, 18 (03)
  • [39] Latte: A Language, Compiler, and Runtime for Elegant and Efficient Deep Neural Networks
    Truong, Leonard
    Barik, Rajkishore
    Totoni, Ehsan
    Liu, Hai
    Markley, Chick
    Fox, Armando
    Shpeisman, Tatiana
    ACM SIGPLAN NOTICES, 2016, 51 (06) : 209 - 223
  • [40] A Trace-Based View on Operating Guidelines
    Stahl, Christian
    Vogler, Walter
    FOUNDATIONS OF SOFTWARE SCIENCE AND COMPUTATIONAL STRUCTURES, 2011, 6604 : 411 - +