Research on Parallel Acceleration for Deep Learning Inference Based on Many-Core ARM Platform

被引:0
|
作者
Zhu, Keqian [1 ]
Jiang, Jingfei [1 ]
机构
[1] Natl Univ Def Technol, Natl Lab Parallel & Distributed Proc, Changsha, Hunan, Peoples R China
来源
关键词
Parallel acceleration; Deep learning inference; Many-core ARM;
D O I
10.1007/978-981-13-2423-9_3
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning is one of the hottest research directions in the field of artificial intelligence. It has achieved results which subvert these of traditional methods. However, the demand for computing ability of hardware platform is also increasing. The academia and industry mainly use heterogeneous GPUs to accelerating computation. ARM is relatively more open than GPUs. The purpose of this paper is to study the performance and related acceleration techniques of ThunderX high-performance many-core ARM chips under large-scale inference tasks. In order to study the computational performance of the target platform objectively, several deep models are adapted for acceleration. Through the selection of computational libraries, adjustment of parallel strategies, application of various performance optimization techniques, we have excavated the computing ability of many-core ARM platforms deeply. The final experimental results show that the performance of single-chip ThunderX is equivalent to that of the i7 7700 K chip, and the overall performance of dual-chip can reach 1.77 times that of the latter. In terms of energy efficiency, the former is inferior to the latter. Stronger cooling system or bad power management may lead to more power consumption. Overall, high-performance ARM chips can be deployed in the cloud to complete large-scale deep learning inference tasks which requiring high throughput.
引用
收藏
页码:30 / 41
页数:12
相关论文
共 50 条
  • [1] Characterizing and Optimizing Transformer Inference on ARM Many-core Processor
    Jiang, Jiazhi
    Du, Jiangsu
    Huang, Dan
    Li, Dongsheng
    Zheng, Jiang
    Lu, Yutong
    51ST INTERNATIONAL CONFERENCE ON PARALLEL PROCESSING, ICPP 2022, 2022,
  • [2] GPU Acceleration for Simulating Massively Parallel Many-Core Platforms
    Raghav, Shivani
    Ruggiero, Martino
    Marongiu, Andrea
    Pinto, Christian
    Atienza, David
    Benini, Luca
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2015, 26 (05) : 1336 - 1349
  • [3] Parallel simulation of many-core processor and many-core clusters
    Lü, Huiwei
    Cheng, Yuan
    Bai, Lu
    Chen, Mingyu
    Fan, Dongrui
    Sun, Ninghui
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2013, 50 (05): : 1110 - 1117
  • [4] Massively Parallel Neural Signal Processing on a Many-Core Platform
    Chen, Dan
    Wang, Lizhe
    Ouyang, Gaoxiang
    Li, Xiaoli
    COMPUTING IN SCIENCE & ENGINEERING, 2011, 13 (06) : 42 - 51
  • [5] SCC: A FLEXIBLE ARCHITECTURE FOR MANY-CORE PLATFORM RESEARCH
    Gries, Matthias
    Hoffmann, Ulrich
    Konow, Michael
    Riepen, Michael
    COMPUTING IN SCIENCE & ENGINEERING, 2011, 13 (06) : 79 - 83
  • [6] A many-core based parallel tabu search
    Lam, Yuet M.
    Luk, Wayne
    International Journal of Computers and Applications, 2014, 36 (01) : 15 - 22
  • [7] Full-Stack Optimizing Transformer Inference on ARM Many-Core CPU
    Jiang, Jiazhi
    Du, Jiangsu
    Huang, Dan
    Chen, Zhiguang
    Lu, Yutong
    Liao, Xiangke
    IEEE TRANSACTIONS ON PARALLEL AND DISTRIBUTED SYSTEMS, 2023, 34 (07) : 2221 - 2235
  • [8] Highly Parallel Framework for HEVC Motion Estimation on Many-core Platform
    Yan, Chenggang
    Zhang, Yongdong
    Dai, Feng
    Li, Liang
    2013 DATA COMPRESSION CONFERENCE (DCC), 2013, : 63 - 72
  • [9] Efficient Parallel Framework for HEVC Deblocking Filter on Many-core Platform
    Yan, Chenggang
    Zhang, Yongdong
    Dai, Feng
    Li, Liang
    2013 DATA COMPRESSION CONFERENCE (DCC), 2013, : 530 - 530
  • [10] Federated Learning Platform on Embedded Many-core Processor with Flower
    Hasumi, Masahiro
    Azumi, Takuya
    2024 IEEE 3RD REAL-TIME AND INTELLIGENT EDGE COMPUTING WORKSHOP, RAGE 2024, 2024, : 37 - 42