Kernel Mapping Techniques for Deep Learning Neural Network Accelerators

被引:0
|
作者
Ozdemir, Sarp [1 ]
Khasawneh, Mohammad [1 ,2 ]
Rao, Smriti [1 ,3 ]
Madden, Patrick H. [1 ]
机构
[1] SUNY Binghamton CSD, Binghamton, NY 13901 USA
[2] MathWorks, Binghamton, NY USA
[3] Ixigo, Binghamton, NY USA
关键词
deep learning; machine learning; combinatorial optimization; kernel mapping; placement;
D O I
10.1145/3505170.3506730
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning applications are compute intensive and naturally parallel; this has spurred the development of new processor architectures tuned for the work load. In this paper, we consider structural differences between deep learning neural networks and more conventional circuits - highlighting how this impacts strategies for mapping neural network compute kernels onto available hardware. We present an efficient mapping approach based on dynamic programming, and also a method to establish performance bounds. We also propose an architectural approach to extend the practical life time of hardware accelerators, enabling the integration of a variety of heterogenous processors into a high performance system. Experimental results using benchmarks from a recent ISPD contest are also reported.
引用
收藏
页码:21 / 28
页数:8
相关论文
共 50 条
  • [1] Hardware Approximate Techniques for Deep Neural Network Accelerators: A Survey
    Armeniakos, Giorgos
    Zervakis, Georgios
    Soudris, Dimitrios
    Henkel, Joerg
    ACM COMPUTING SURVEYS, 2023, 55 (04)
  • [2] Mapping of Deep Neural Network Accelerators on Wireless Multistage Interconnection NoCs
    Aydi, Yassine
    Mnejja, Sirine
    Mohammed, Faraqid Q.
    Abid, Mohamed
    APPLIED SCIENCES-BASEL, 2024, 14 (01):
  • [3] SEALing Neural Network Models in Encrypted Deep Learning Accelerators
    Zuo, Pengfei
    Hua, Yu
    Liang, Ling
    Xie, Xinfeng
    Hu, Xing
    Xie, Yuan
    2021 58TH ACM/IEEE DESIGN AUTOMATION CONFERENCE (DAC), 2021, : 1255 - 1260
  • [4] Deep Kernel: Learning Kernel Function from Data Using Deep Neural Network
    Le, Linh
    Hao, Jie
    Xie, Ying
    Priestley, Jennifer
    2016 3RD IEEE/ACM INTERNATIONAL CONFERENCE ON BIG DATA COMPUTING, APPLICATIONS AND TECHNOLOGIES (BDCAT), 2016, : 1 - 7
  • [5] Optimizing deep learning inference on mobile devices with neural network accelerators
    曾惜
    Xu Yunlong
    Zhi Tian
    HighTechnologyLetters, 2019, 25 (04) : 417 - 425
  • [6] AutoMap: Automatic Mapping of Neural Networks to Deep Learning Accelerators for Edge Devices
    Wang, Yanhong
    Zhao, Zihao
    Jin, Xu
    Zheng, Haotian
    Nie, Maohua
    Zou, Qiaosha
    Shi, C. -J. Richard
    IEEE TRANSACTIONS ON COMPUTER-AIDED DESIGN OF INTEGRATED CIRCUITS AND SYSTEMS, 2023, 42 (09) : 2994 - 3006
  • [7] Exploring Quantization and Mapping Synergy in Hardware-Aware Deep Neural Network Accelerators
    Klhufek, Jan
    Safar, Miroslav
    Mrazek, Vojtech
    Vasicek, Zdenek
    Sekanina, Lukas
    2024 27TH INTERNATIONAL SYMPOSIUM ON DESIGN & DIAGNOSTICS OF ELECTRONIC CIRCUITS & SYSTEMS, DDECS, 2024, : 1 - 6
  • [8] Understanding Error Propagation in Deep Learning Neural Network (DNN) Accelerators and Applications
    Li, Guanpeng
    Hari, Siva Kumar Sastry
    Sullivan, Michael
    Tsai, Timothy
    Pattabiraman, Karthik
    Emer, Joel
    Keckler, Stephen W.
    SC'17: PROCEEDINGS OF THE INTERNATIONAL CONFERENCE FOR HIGH PERFORMANCE COMPUTING, NETWORKING, STORAGE AND ANALYSIS, 2017,
  • [9] Review of ASIC accelerators for deep neural network
    Machupalli, Raju
    Hossain, Masum
    Mandal, Mrinal
    MICROPROCESSORS AND MICROSYSTEMS, 2022, 89
  • [10] Approximate Adders for Deep Neural Network Accelerators
    Raghuram, S.
    Shashank, N.
    2022 35TH INTERNATIONAL CONFERENCE ON VLSI DESIGN (VLSID 2022) HELD CONCURRENTLY WITH 2022 21ST INTERNATIONAL CONFERENCE ON EMBEDDED SYSTEMS (ES 2022), 2022, : 210 - 215