Deep learning accelerators: a case study with MAESTRO

被引:5
|
作者
Bolhasani, Hamidreza [1 ]
Jassbi, Somayyeh Jafarali [1 ]
机构
[1] Islamic Azad Univ, Sci & Res Branch, Dept Comp Engn, Tehran, Iran
关键词
Deep learning; Convolutional neural networks; Deep neural networks; Hardware accelerator; Deep learning accelerator;
D O I
10.1186/s40537-020-00377-8
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
In recent years, deep learning has become one of the most important topics in computer sciences. Deep learning is a growing trend in the edge of technology and its applications are now seen in many aspects of our life such as object detection, speech recognition, natural language processing, etc. Currently, almost all major sciences and technologies are benefiting from the advantages of deep learning such as high accuracy, speed and flexibility. Therefore, any efforts in improving performance of related techniques is valuable. Deep learning accelerators are considered as hardware architecture, which are designed and optimized for increasing speed, efficiency and accuracy of computers that are running deep learning algorithms. In this paper, after reviewing some backgrounds on deep learning, a well-known accelerator architecture named MAERI (Multiply-Accumulate Engine with Reconfigurable interconnects) is investigated. Performance of a deep learning task is measured and compared in two different data flow strategies: NLR (No Local Reuse) and NVDLA (NVIDIA Deep Learning Accelerator), using an open source tool called MAESTRO (Modeling Accelerator Efficiency via Spatio-Temporal Resource Occupancy). Measured performance indicators of novel optimized architecture, NVDLA shows higher L1 and L2 computation reuse, and lower total runtime (cycles) in comparison to the other one.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Deep learning accelerators: a case study with MAESTRO
    Hamidreza Bolhasani
    Somayyeh Jafarali Jassbi
    Journal of Big Data, 7
  • [2] Exploring Software Models for the Resilience Analysis of Deep Learning Accelerators: the NVDLA Case Study
    Veronesi, A.
    Dall'Occo, F.
    Bertozzi, D.
    Favalli, M.
    Krstic, M.
    2022 25TH INTERNATIONAL SYMPOSIUM ON DESIGN AND DIAGNOSTICS OF ELECTRONIC CIRCUITS AND SYSTEMS (DDECS), 2022, : 142 - 147
  • [3] Parallelism in Deep Learning Accelerators
    Song, Linghao
    Chen, Fan
    Chen, Yiran
    Li, Hai Helen
    2020 25TH ASIA AND SOUTH PACIFIC DESIGN AUTOMATION CONFERENCE, ASP-DAC 2020, 2020, : 645 - 650
  • [4] Deep Learning Accelerators' Configuration Space Exploration Effect on Performance and Resource Utilization: A Gemmini Case Study
    Gookyi, Dennis Agyemanh Nana
    Lee, Eunchong
    Kim, Kyungho
    Jang, Sung-Joon
    Lee, Sang-Seol
    SENSORS, 2023, 23 (05)
  • [5] Ergodic Approximate Deep Learning Accelerators
    van Lijssel, Tim
    Balatsoukas-Stimming, Alexios
    FIFTY-SEVENTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, IEEECONF, 2023, : 734 - 738
  • [6] Hardware Accelerators for Deep Reinforcement Learning
    Mishra, Vinod K.
    Basu, Kanad
    Arunachalam, Ayush
    ARTIFICIAL INTELLIGENCE AND MACHINE LEARNING FOR MULTI-DOMAIN OPERATIONS APPLICATIONS V, 2023, 12538
  • [7] AdequateDL: Approximating Deep Learning Accelerators
    Sentieys, Olivier
    Filip, Silviu
    Briand, David
    Novo, David
    Dupuis, Etienne
    O'Connor, Ian
    Bosio, Alberto
    2021 24TH INTERNATIONAL SYMPOSIUM ON DESIGN AND DIAGNOSTICS OF ELECTRONIC CIRCUITS & SYSTEMS (DDECS), 2021, : 37 - 40
  • [8] Exploiting deep learning accelerators for neuromorphic workloads
    Sun, Pao-Sheng Vincent
    Titterton, Alexander
    Gopiani, Anjlee
    Santos, Tim
    Basu, Arindam
    Lu, Wei D.
    Eshraghian, Jason K.
    NEUROMORPHIC COMPUTING AND ENGINEERING, 2024, 4 (01):
  • [9] Assembly language and assembler for deep learning accelerators
    Lan H.
    Wu L.
    Han D.
    Du Z.
    High Technology Letters, 2019, 25 (04): : 386 - 394
  • [10] Data Orchestration in Deep Learning Accelerators Krishna
    Krishna T.
    Kwon H.
    Parashar A.
    Pellauer M.
    Samajdar A.
    Synthesis Lectures on Computer Architecture, 2020, 15 (03): : 1 - 164