Kernel Mapping Techniques for Deep Learning Neural Network Accelerators

被引:0
|
作者
Ozdemir, Sarp [1 ]
Khasawneh, Mohammad [1 ,2 ]
Rao, Smriti [1 ,3 ]
Madden, Patrick H. [1 ]
机构
[1] SUNY Binghamton CSD, Binghamton, NY 13901 USA
[2] MathWorks, Binghamton, NY USA
[3] Ixigo, Binghamton, NY USA
关键词
deep learning; machine learning; combinatorial optimization; kernel mapping; placement;
D O I
10.1145/3505170.3506730
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning applications are compute intensive and naturally parallel; this has spurred the development of new processor architectures tuned for the work load. In this paper, we consider structural differences between deep learning neural networks and more conventional circuits - highlighting how this impacts strategies for mapping neural network compute kernels onto available hardware. We present an efficient mapping approach based on dynamic programming, and also a method to establish performance bounds. We also propose an architectural approach to extend the practical life time of hardware accelerators, enabling the integration of a variety of heterogenous processors into a high performance system. Experimental results using benchmarks from a recent ISPD contest are also reported.
引用
收藏
页码:21 / 28
页数:8
相关论文
共 50 条
  • [21] Learning Deep Neural Network Based Kernel Functions for Small Sample Size Classification
    Zheng, Tieran
    Han, Jiqing
    Zheng, Guibin
    NEURAL INFORMATION PROCESSING, ICONIP 2017, PT I, 2017, 10634 : 135 - 143
  • [22] Enhancing Collaborative and Geometric Multi-Kernel Learning Using Deep Neural Network
    Zafar, Bareera
    Naqvi, Syed Abbas Zilqurnain
    Ahsan, Muhammad
    Ditta, Allah
    Baneen, Ummul
    Khan, Muhammad Adnan
    CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 72 (03): : 5099 - 5116
  • [23] An Overview of Efficient Interconnection Networks for Deep Neural Network Accelerators
    Nabavinejad, Seyed Morteza
    Baharloo, Mohammad
    Chen, Kun-Chih
    Palesi, Maurizio
    Kogel, Tim
    Ebrahimi, Masoumeh
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2020, 10 (03) : 268 - 282
  • [24] Coupling RBF neural network with ensemble learning techniques for landslide susceptibility mapping
    Binh Thai Pham
    Trung Nguyen-Thoi
    Qi, Chongchong
    Tran Van Phong
    Dou, Jie
    Ho, Lanh Si
    Hiep Van Le
    Prakash, Indra
    CATENA, 2020, 195
  • [25] BenQ: Benchmarking Automated Quantization on Deep Neural Network Accelerators
    Wei, Zheng
    Zhang, Xingjun
    Li, Jingbo
    Ji, Zeyu
    Wei, Jia
    PROCEEDINGS OF THE 2022 DESIGN, AUTOMATION & TEST IN EUROPE CONFERENCE & EXHIBITION (DATE 2022), 2022, : 1479 - 1484
  • [26] Tango: A Deep Neural Network Benchmark Suite for Various Accelerators
    Karki, Aajna
    Keshava, Chethan Palangotu
    Shivakumar, Spoorthi Mysore
    Skow, Joshua
    Hegde, Goutam Madhukeshwar
    Jeon, Hyeran
    2019 IEEE INTERNATIONAL SYMPOSIUM ON PERFORMANCE ANALYSIS OF SYSTEMS AND SOFTWARE (ISPASS), 2019, : 137 - 138
  • [27] LAMBDA: An Open Framework for Deep Neural Network Accelerators Simulation
    Russo, Enrico
    Palesi, Maurizio
    Monteleone, Salvatore
    Patti, Davide
    Ascia, Giuseppe
    Catania, Vincenzo
    2021 IEEE INTERNATIONAL CONFERENCE ON PERVASIVE COMPUTING AND COMMUNICATIONS WORKSHOPS AND OTHER AFFILIATED EVENTS (PERCOM WORKSHOPS), 2021, : 161 - 166
  • [28] An overview memristor based hardware accelerators for deep neural network
    Gokgoz, Baki
    Gul, Fatih
    Aydin, Tolga
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2024, 36 (09):
  • [29] AINoC: New Interconnect for Future Deep Neural Network Accelerators
    Krichene, Hana
    Prasad, Rohit
    Mouhagir, Ayoub
    DESIGN AND ARCHITECTURE FOR SIGNAL AND IMAGE PROCESSING, DASIP 2023, 2023, 13879 : 55 - 69
  • [30] Optimizing Memory Efficiency for Deep Convolutional Neural Network Accelerators
    Li, Xiaowei
    Li, Jiajun
    Yan, Guihai
    JOURNAL OF LOW POWER ELECTRONICS, 2018, 14 (04) : 496 - 507