Processing Natural Language on Embedded Devices: How Well Do Transformer Models Perform?

被引:0
|
作者
Sarkar, Souvika [1 ]
Babar, Mohammad Fakhruddin [2 ]
Hassan, Md Mahadi [1 ]
Hasan, Monowar [2 ]
Santu, Shubhra Kanti Karmaker [1 ]
机构
[1] Auburn Univ, Auburn, AL 36849 USA
[2] Washington State Univ, Pullman, WA USA
来源
PROCEEDINGS OF THE 15TH ACM/SPEC INTERNATIONAL CONFERENCE ON PERFORMANCE ENGINEERING, ICPE 2024 | 2024年
基金
美国国家科学基金会;
关键词
Transformers; Embedded Systems; NLP; Language Models;
D O I
10.1145/3629526.3645054
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Transformer-based language models such as BERT and its variants are primarily developed with compute-heavy servers in mind. Despite the great performance of BERT models across various NLP tasks, their large size and numerous parameters pose substantial obstacles to offline computation on embedded systems. Lighter replacements of such language models (e.g., DistilBERT and TinyBERT) often sacrifice accuracy, particularly for complex NLP tasks. Until now, it is still unclear (a) whether the state-of-the-art language models, viz., BERT and its variants are deployable on embedded systems with a limited processor, memory, and battery power and (b) if they do, what are the "right" set of configurations and parameters to choose for a given NLP task. This paper presents a performance study of transformer language models under different hardware configurations and accuracy requirements and derives empirical observations about these resource/accuracy trade-offs. In particular, we study how the most commonly used BERT-based language models (viz., BERT, RoBERTa, DistilBERT, and TinyBERT) perform on embedded systems. We tested them on four off-the-shelf embedded platforms (Raspberry Pi, Jetson, UP2, and UDOO) with 2 GB and 4 GB memory (i.e., a total of eight hardware configurations) and four datasets (i.e., HuRIC, GoEmotion, CoNLL, WNUT17) running various NLP tasks. Our study finds that executing complex NLP tasks (such as "sentiment" classification) on embedded systems is feasible even without any GPUs (e.g., Raspberry Pi with 2 GB of RAM). Our findings can help designers understand the deployability and performance of transformer language models, especially those based on BERT architectures.
引用
收藏
页码:211 / 222
页数:12
相关论文
共 50 条
  • [1] How Well Do Large Language Models Perform on Faux Pas Tests?
    Shapira, Natalie
    Zwirn, Guy
    Goldberg, Yoav
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023), 2023, : 10438 - 10451
  • [2] How well do CPU, GPU and Hybrid Graph Processing Frameworks Perform?
    Aasawat, Tanuj Kr
    Reza, Tahsin
    Ripeanu, Matei
    2018 IEEE INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM WORKSHOPS (IPDPSW 2018), 2018, : 458 - 466
  • [3] African Dust Deposition to Florida: How well do dust models perform?
    Prospero, J. M.
    Landing, W.
    WMO/GEO EXPERT MEETING ON AN INTERNATIONAL SAND AND DUST STORM WARNING SYSTEM, 2009, 7
  • [4] DESIGNATED JUDGES - HOW WELL DO THEY PERFORM
    GREEN, JJ
    ATKINS, BM
    JUDICATURE, 1978, 61 (08) : 358 - 370
  • [5] How well do your crews perform?
    Muntz, Hans
    Better Roads, 1996, 66 (06):
  • [6] HOW WELL DO MULTINATIONAL COMPANIES PERFORM
    STICH, RS
    MANAGEMENT INTERNATIONAL REVIEW, 1971, 11 (4-5) : 33 - 44
  • [7] Alternative cleaners. How well do they perform?
    Munson, Terry
    Surface mount technology, 1994, 8 (03): : 107 - 108
  • [8] How Well Do Neurochecks Perform After Stroke?
    Benedetti, Andres M. De Leon
    Bhatia, Ritwik
    Ancheta, Selena R.
    Romano, Jose G.
    Koch, Sebastian
    STROKE, 2021, 52 (03) : 1094 - 1097
  • [9] POISON CENTERS IN AMERICA - HOW WELL DO THEY PERFORM
    GELLER, RJ
    FISHER, JG
    LEEPER, JD
    TOOSON, JD
    RANGANATHAN, S
    VETERINARY AND HUMAN TOXICOLOGY, 1990, 32 (03) : 240 - 245
  • [10] How Well do Graph-Processing Platforms Perform? An Empirical Performance Evaluation and Analysis
    Guo, Yong
    Biczak, Marcin
    Varbanescu, Ana Lucia
    Iosup, Alexandru
    Martella, Claudio
    Willke, Theodore L.
    2014 IEEE 28TH INTERNATIONAL PARALLEL AND DISTRIBUTED PROCESSING SYMPOSIUM, 2014,