Processing Natural Language on Embedded Devices: How Well Do Transformer Models Perform?

被引:0
|
作者
Sarkar, Souvika [1 ]
Babar, Mohammad Fakhruddin [2 ]
Hassan, Md Mahadi [1 ]
Hasan, Monowar [2 ]
Santu, Shubhra Kanti Karmaker [1 ]
机构
[1] Auburn Univ, Auburn, AL 36849 USA
[2] Washington State Univ, Pullman, WA USA
来源
PROCEEDINGS OF THE 15TH ACM/SPEC INTERNATIONAL CONFERENCE ON PERFORMANCE ENGINEERING, ICPE 2024 | 2024年
基金
美国国家科学基金会;
关键词
Transformers; Embedded Systems; NLP; Language Models;
D O I
10.1145/3629526.3645054
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Transformer-based language models such as BERT and its variants are primarily developed with compute-heavy servers in mind. Despite the great performance of BERT models across various NLP tasks, their large size and numerous parameters pose substantial obstacles to offline computation on embedded systems. Lighter replacements of such language models (e.g., DistilBERT and TinyBERT) often sacrifice accuracy, particularly for complex NLP tasks. Until now, it is still unclear (a) whether the state-of-the-art language models, viz., BERT and its variants are deployable on embedded systems with a limited processor, memory, and battery power and (b) if they do, what are the "right" set of configurations and parameters to choose for a given NLP task. This paper presents a performance study of transformer language models under different hardware configurations and accuracy requirements and derives empirical observations about these resource/accuracy trade-offs. In particular, we study how the most commonly used BERT-based language models (viz., BERT, RoBERTa, DistilBERT, and TinyBERT) perform on embedded systems. We tested them on four off-the-shelf embedded platforms (Raspberry Pi, Jetson, UP2, and UDOO) with 2 GB and 4 GB memory (i.e., a total of eight hardware configurations) and four datasets (i.e., HuRIC, GoEmotion, CoNLL, WNUT17) running various NLP tasks. Our study finds that executing complex NLP tasks (such as "sentiment" classification) on embedded systems is feasible even without any GPUs (e.g., Raspberry Pi with 2 GB of RAM). Our findings can help designers understand the deployability and performance of transformer language models, especially those based on BERT architectures.
引用
收藏
页码:211 / 222
页数:12
相关论文
共 50 条
  • [31] Colorectal Cancer Detection: How Well Do Appropriateness Criteria Perform?
    Gonvers, Jean-Jacques
    Burnand, Bernard
    Vader, John-Paul
    Froehlich, Florian
    CLINICAL GASTROENTEROLOGY AND HEPATOLOGY, 2009, 7 (07) : 807 - 808
  • [32] HOW WELL DO WOOD TRUSSES REALLY PERFORM DURING A FIRE
    SCHAFFER, EL
    FIRE JOURNAL, 1988, 82 (02): : 57 - &
  • [33] Death Telling-How Well Do EM Residents Perform?
    Hall, R., V
    Forstater, A. T.
    Shin, N.
    Lopez, B. L.
    ANNALS OF EMERGENCY MEDICINE, 2012, 60 (05) : S185 - S185
  • [34] How well do large language models know dentistry? AI takes the test
    Memon, Maaz Anwer
    BRITISH DENTAL JOURNAL, 2025, 238 (01) : 33 - 33
  • [35] POISON CENTERS IN AMERICA .2. HOW WELL DO THEY PERFORM
    GELLER, RJ
    FISHER, JG
    LEEPER, JD
    VETERINARY AND HUMAN TOXICOLOGY, 1987, 29 (06) : 458 - 459
  • [36] HOW WELL DO WOOD TRUSSES REALLY PERFORM DURING A FIRE?
    Schaffer, Erwin L.
    Fire journal Boston, Mass., 1988, 82 (02): : 57 - 61
  • [37] Plagiarism-Detection Services: How Well Do They Actually Perform?
    Fiedler, Rebecca L.
    Kaner, Cem
    IEEE TECHNOLOGY AND SOCIETY MAGAZINE, 2010, 29 (04) : 37 - 43
  • [38] Backdoor Learning of Language Models in Natural Language Processing
    University of Michigan
    1600,
  • [39] Natural language processing in the era of large language models
    Zubiaga, Arkaitz
    FRONTIERS IN ARTIFICIAL INTELLIGENCE, 2024, 6
  • [40] How do we evaluate others' memories? A behavioural and natural language processing study
    Gamoran, Avi
    Gilead, Michael
    Sadeh, Talya
    BRAIN AND COGNITION, 2023, 170