FPGA-Accelerated Hadoop Cluster for Deep Learning Computations

被引:12
|
作者
Alhamali, Abdulrahman [1 ]
Salha, Nibal [1 ]
Morcel, Raghid [1 ]
Ezzeddine, Mazen [1 ]
Hamdan, Omar [1 ]
Akkary, Haitham [1 ]
Hajj, Hazem [1 ]
机构
[1] Amer Univ Beirut, Elect & Comp Engn Dept, Beirut, Lebanon
关键词
deep learning; convolutional neural network; Hadoop; FPGA; map-reduce;
D O I
10.1109/ICDMW.2015.148
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep learning algorithms have received significant attention in the last few years. Their popularity is due to their ability to achieve higher accuracy than conventional machine learning in many research areas such as speech recognition, image processing and natural language processing. Deep learning algorithms rely on multiple cascaded layers of non-linear processing units, typically composed of hidden artificial neural networks for feature extraction and transformation. However, deep learning algorithms require a large amount of computational power and significant amount of time to train. Fortunately, the training and inference algorithms of deep learning architectures expose abundant data-parallelism. We aim in this work to develop technology that exploits deep learning data parallelism in 2 ways: 1) by distributing deep computation into a Hadoop cluster or cloud of computing nodes, and 2) by using field programmable gate arrays (FPGA) hardware acceleration to speed up computationally intensive deep learning kernels. In this paper, we describe a hardware prototype of our accelerated Hadoop deep learning system architecture and report initial performance and energy reduction results. By accelerating the convolutional layers of deep learning Convolutional Neural Network, we have observed a potential speed-up of 12.6 times and an energy reduction of 87.5% on a 6-node FPGA accelerated Hadoop cluster.
引用
收藏
页码:565 / 574
页数:10
相关论文
共 50 条
  • [1] FA3C: FPGA-Accelerated Deep Reinforcement Learning
    Cho, Hyungmin
    Oh, Pyeongseok
    Park, Jiyoung
    Jung, Wookeun
    Lee, Jaejin
    TWENTY-FOURTH INTERNATIONAL CONFERENCE ON ARCHITECTURAL SUPPORT FOR PROGRAMMING LANGUAGES AND OPERATING SYSTEMS (ASPLOS XXIV), 2019, : 499 - 513
  • [2] A Proposal for FPGA-Accelerated Deep Learning Ensembles in MPSoC Platforms Applied to Malware Detection
    Cilardo, Alessandro
    Maisto, Vincenzo
    Mazzocca, Nicola
    di Torrepadula, Franca Rocco
    QUALITY OF INFORMATION AND COMMUNICATIONS TECHNOLOGY, QUATIC 2022, 2022, 1621 : 239 - 249
  • [3] A Methodology for Evaluating and Analyzing FPGA-Accelerated, Deep-Learning Applications for Onboard Space Processing
    Sabogal, Sebastian
    George, Alan
    2021 IEEE SPACE COMPUTING CONFERENCE (SCC), 2021, : 143 - 154
  • [4] FPGA-accelerated Quantum Transport Measurements
    Haarman, Timo
    de Almeida, Antonio Sousa
    Heskes, Amber
    Zwanenburg, Floris
    Alachiotis, Nikolaos
    2023 INTERNATIONAL CONFERENCE ON FIELD PROGRAMMABLE TECHNOLOGY, ICFPT, 2023, : 44 - 52
  • [5] FPGA-accelerated simulation of computer systems
    1600, Morgan and Claypool Publishers (09):
  • [6] FPGA-Accelerated Spreading for Global Placement
    Dhar, Shounak
    Singhal, Love
    Iyer, Mahesh A.
    Pan, David Z.
    2019 IEEE HIGH PERFORMANCE EXTREME COMPUTING CONFERENCE (HPEC), 2019,
  • [7] FPGA-accelerated simulation of computer systems
    1600, Morgan and Claypool Publishers (29):
  • [8] FPGA-accelerated Complex Event Processing
    Takenaka, Takashi
    Inoue, Hiroaki
    Hosomi, Takeo
    Nakamura, Yuichi
    2015 SYMPOSIUM ON VLSI CIRCUITS (VLSI CIRCUITS), 2015,
  • [9] Trust in FPGA-accelerated Cloud Computing
    Turan, Furkan
    Verbauwhede, Ingrid
    ACM COMPUTING SURVEYS, 2021, 53 (06)
  • [10] FPGA-Accelerated Simulation of Computer Systems
    Angepat, Hari
    Chiou, Derek
    Chung, Eric S.
    Hoe, James C.
    Synthesis Lectures on Computer Architecture, 2014, 29 : 1 - 82