Work-in-Progress: Furion: Alleviating Overheads for Deep Learning Framework On Single Machine

被引:0
|
作者
Jin, Lihui [1 ]
Wang, Chao [1 ]
Gong, Lei [1 ]
Xu, Chongchong [1 ]
Hu, Yahui [1 ]
Tan, Luchao [1 ]
Zhou, Xuehai [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Peoples R China
关键词
Deep Learning; Overhead; Throughput;
D O I
暂无
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Deep learning has been successful at solving many kinds of tasks. Hardware accelerators with high performance and parallelism have become mainstream to implement deep neural networks. In order to increase hardware utilization, multiple applications will share the same compute resource. However, different applications may use different deep learning frameworks and occupy different amounts of resources. If there are no scheduling platforms that are compatible with different frameworks, resources competition will result in longer response time, run out of memory, and other errors. When the resources of the system cannot satisfy all the applications at the same time, application switching overhead will be excessive without reasonable resource management strategy. In this paper, we propose Furion - a middleware alleviates overheads for deep learning framework on a single machine. Furion schedules tasks, overlaps the execution of different computing resource, and batches unknown inputs to increase the hardware accelerator utilization. It dynamically manages memory usage for each application to alleviate the overhead of application switching and make a complex model enable implement in a low-end GPU. Our experiment proved that Furion achieves 2.2x-2.7x speedup on the GTX1060.
引用
收藏
页数:2
相关论文
共 50 条
  • [1] A Fast Design Space Exploration Framework for the Deep Learning Accelerators: Work-in-Progress
    Colucci, Alessio
    Marchisio, Alberto
    Bussolino, Beatrice
    Mrazek, Voitech
    Martina, Maurizio
    Masera, Guido
    Shafique, Muhammad
    [J]. PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON HARDWARE/SOFTWARE CODESIGN AND SYSTEM SYNTHESIS (CODES+ISSS), 2019, : 34 - 36
  • [2] Work-In-Progress: Making Machine Learning Real-Time Predictable
    Xu, Hang
    Mueller, Frank
    [J]. 2018 39TH IEEE REAL-TIME SYSTEMS SYMPOSIUM (RTSS 2018), 2018, : 157 - 160
  • [3] Work-In-Progress: A Deep Learning Strategy for I/O Scheduling in Storage Systems
    Farhangi, Ashkan
    Bian, Jiang
    Wang, Jun
    Guo, Zhishan
    [J]. 2019 IEEE 40TH REAL-TIME SYSTEMS SYMPOSIUM (RTSS 2019), 2019, : 568 - 571
  • [4] Work-in-Progress: Snapshot-based Offloading for Machine Learning Web App
    Jeong, InChang
    Jeong, Hyuk-Jin
    Moon, Soo-Mook
    [J]. 2017 INTERNATIONAL CONFERENCE ON EMBEDDED SOFTWARE (EMSOFT), 2017,
  • [5] Work-in-Progress: Machine Development Using Virtual Commissioning
    Smajic, Hasan
    Bosco, Jean
    [J]. CROSS REALITY AND DATA SCIENCE IN ENGINEERING, 2021, 1231 : 614 - 623
  • [6] Work-in-Progress: Design of an Online Learning Coach
    DePiero, Fred W.
    [J]. 2013 ASEE ANNUAL CONFERENCE, 2013,
  • [7] A Lifelong Health Monitoring Framework in Processors: Work-in-Progress
    Hu, Xiao
    Wang, Yaohua
    [J]. PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON COMPILERS, ARCHITECTURE, AND SYNTHESIS FOR EMBEDDED SYSTEMS (CASES), 2020, : 6 - 8
  • [8] Work-in-Progress: Cloud-based Machine Learning for IoT Devices with Better Privacy
    Jeong, Hyuk-Jin
    Lee, Hyeon-Jae
    Moon, Soo-Mook
    [J]. 2017 INTERNATIONAL CONFERENCE ON EMBEDDED SOFTWARE (EMSOFT), 2017,
  • [9] A Machine Learning based Approximate Computing Approach on Data Flow Graphs: Work-in-Progress
    Wang, Ye
    Dong, Jian
    Liu, Yanxin
    Wang, Chunpei
    Qu, Gang
    [J]. PROCEEDINGS OF THE 2020 INTERNATIONAL CONFERENCE ON EMBEDDED SOFTWARE (EMSOFT), 2020, : 37 - 39
  • [10] Work-in-Progress: The Road to Learning, Using Gamification.
    Oropeza Hernandez, Maria Ivonne
    Munoz Lezama, Rodrigo
    Madero Gomez, Sergio
    [J]. PROCEEDINGS OF THE 2021 IEEE GLOBAL ENGINEERING EDUCATION CONFERENCE (EDUCON), 2021, : 1399 - 1403