High performance accelerators for deep neural networks: A review

被引:6
|
作者
Akhoon, Mohd Saqib [1 ]
Suandi, Shahrel A. [1 ]
Alshahrani, Abdullah [2 ]
Saad, Abdul-Malik H. Y. [3 ]
Albogamy, Fahad R. [4 ]
Bin Abdullah, Mohd Zaid [1 ]
Loan, Sajad A. [5 ]
机构
[1] Univ Sains, Sch Elect & Elect Engn, Intelligent Biometr Grp, George Town, Malaysia
[2] Univ Jeddah, Dept Comp Sci & Artificial Intelligence, Coll Comp Sci & Engn, Jeddah, Saudi Arabia
[3] Univ Teknol Malaysia, Div Elect & Comp Engn, Fac Engn, Sch Elect Engn, Johor Baharu, Kagawa, Malaysia
[4] Taif Univ, Saudi Univ, Turabah Univ Coll, Comp Sci Program, At Taif, Saudi Arabia
[5] Jamia Millia Islamia, Dept Elect & Commun, New Delhi 11025, India
关键词
artificial intelligence; convolutional neural networks; deep neural network; machine learning; accelerators; CNN;
D O I
10.1111/exsy.12831
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The availability of huge structured and unstructured data, advanced highly dense memory and high performance computing machines have provided a strong push for the development in artificial intelligence (AI) and machine learning (ML) domains. AI and machine learning has rekindled the hope of efficiently solving complex problems which was not possible in the recent past. The generation and availability of big-data is a strong driving force for the development of AI/ML applications, however, several challenges need to be addressed, like processing speed, memory requirement, high bandwidth, low latency memory access, and highly conductive and flexible connections between processing units and memory blocks. The conventional computing platforms are unable to address these issues with machine learning and AI. Deep neural networks (DNNs) are widely employed for machine learning and AI applications, like speech recognition, computer vison, robotics, and so forth, efficiently and accurately. However, accuracy is achieved at the cost of high computational complexity, sacrificing energy efficiency and throughput like performance measuring parameters along with high latency. To address the problems of latency, energy efficiency, complexity, power consumption, and so forth, a lot of state of the art DNN accelerators have been designed and implemented in the form of application specific integrated circuits (ASICs) and field programmable gate arrays (FPGAs). This work provides the state of the art of all these DNN accelerators which have been developed recently. Various DNN architectures, their computing units, emerging technologies used in improving the performance of DNN accelerators will be discussed. Finally, we will try to explore the scope for further improvement in these accelerator designs, various opportunities and challenges for the future research.
引用
收藏
页数:23
相关论文
共 50 条
  • [21] Using Libraries of Approximate Circuits in Design of Hardware Accelerators of Deep Neural Networks
    Mrazek, Vojtech
    Sekanina, Lukas
    Vasicek, Zdenek
    2020 2ND IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS 2020), 2020, : 243 - 247
  • [22] Design Considerations for Efficient Deep Neural Networks on Processing-in-Memory Accelerators
    Yang, Tien-Ju
    Sze, Vivienne
    2019 IEEE INTERNATIONAL ELECTRON DEVICES MEETING (IEDM), 2019,
  • [23] Optics-informed Neural Networks: Bridging Deep Learning with Photonic Accelerators
    Moralis-Pegios, M.
    Tsakyridis, A.
    Pappas, C.
    Moschos, T.
    Giamougiannis, G.
    Kovaios, S.
    Roumpos, I.
    Kirtas, M.
    Passalis, N.
    Tefas, A.
    Pleros, N.
    2024 OPTICAL FIBER COMMUNICATIONS CONFERENCE AND EXHIBITION, OFC, 2024,
  • [24] Tile-Based Architecture Exploration for Convolutional Accelerators in Deep Neural Networks
    Chen, Yang-Tsai
    Yen, Yu-Xiang
    Chen, Chun-Tse
    Chen, Tzu-Yu
    Huang, Chih-Tsun
    Liou, Jing-Jia
    Lu, Juin-Ming
    2021 IEEE 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE CIRCUITS AND SYSTEMS (AICAS), 2021,
  • [25] A Review of Convolutional Neural Networks Hardware Accelerators for AIoT Edge Computing
    Wu, Fei
    Zhao, Neng
    Liu, Ye
    Chang, Liang
    Zhou, Liang
    Zhou, Jun
    2021 6TH INTERNATIONAL CONFERENCE ON UK-CHINA EMERGING TECHNOLOGIES (UCET 2021), 2021, : 71 - 76
  • [26] Deep Learning Inferencing with High-performance Hardware Accelerators
    Kljucaric, Luke
    George, Alan D.
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (04)
  • [27] Deciphering the Feature Representation of Deep Neural Networks for High-Performance AI
    Islam, Md Tauhidul
    Xing, Lei
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (08) : 5273 - 5287
  • [28] Automatic Generation of Fast and Accurate Performance Models for Deep Neural Network Accelerators
    Luebeck, Konstantin
    Jung, Alexander Louis-Ferdinand
    Wedlich, Felix
    Mueller, Mika Markus
    Peccia, Federico Nicolas
    Thoemmes, Felix
    Steinmetz, Jannik
    Biermaier, Valentin
    Frischknecht, Adrian
    Bernardo, Paul Palomero
    Bringmann, Oliver
    ACM TRANSACTIONS ON EMBEDDED COMPUTING SYSTEMS, 2025, 24 (02)
  • [29] Performance of Deep Neural Networks in Audio Surveillance
    Arslan, Yuksel
    Canbolat, Huseyin
    2018 6TH INTERNATIONAL CONFERENCE ON CONTROL ENGINEERING & INFORMATION TECHNOLOGY (CEIT), 2018,
  • [30] Review of Lightweight Deep Convolutional Neural Networks
    Chen, Fanghui
    Li, Shouliang
    Han, Jiale
    Ren, Fengyuan
    Yang, Zhen
    ARCHIVES OF COMPUTATIONAL METHODS IN ENGINEERING, 2024, 31 (04) : 1915 - 1937