Processing Natural Language on Embedded Devices: How Well Do Transformer Models Perform?

被引:0
|
作者
Sarkar, Souvika [1 ]
Babar, Mohammad Fakhruddin [2 ]
Hassan, Md Mahadi [1 ]
Hasan, Monowar [2 ]
Santu, Shubhra Kanti Karmaker [1 ]
机构
[1] Auburn Univ, Auburn, AL 36849 USA
[2] Washington State Univ, Pullman, WA USA
来源
PROCEEDINGS OF THE 15TH ACM/SPEC INTERNATIONAL CONFERENCE ON PERFORMANCE ENGINEERING, ICPE 2024 | 2024年
基金
美国国家科学基金会;
关键词
Transformers; Embedded Systems; NLP; Language Models;
D O I
10.1145/3629526.3645054
中图分类号
TP3 [计算技术、计算机技术];
学科分类号
0812 ;
摘要
Transformer-based language models such as BERT and its variants are primarily developed with compute-heavy servers in mind. Despite the great performance of BERT models across various NLP tasks, their large size and numerous parameters pose substantial obstacles to offline computation on embedded systems. Lighter replacements of such language models (e.g., DistilBERT and TinyBERT) often sacrifice accuracy, particularly for complex NLP tasks. Until now, it is still unclear (a) whether the state-of-the-art language models, viz., BERT and its variants are deployable on embedded systems with a limited processor, memory, and battery power and (b) if they do, what are the "right" set of configurations and parameters to choose for a given NLP task. This paper presents a performance study of transformer language models under different hardware configurations and accuracy requirements and derives empirical observations about these resource/accuracy trade-offs. In particular, we study how the most commonly used BERT-based language models (viz., BERT, RoBERTa, DistilBERT, and TinyBERT) perform on embedded systems. We tested them on four off-the-shelf embedded platforms (Raspberry Pi, Jetson, UP2, and UDOO) with 2 GB and 4 GB memory (i.e., a total of eight hardware configurations) and four datasets (i.e., HuRIC, GoEmotion, CoNLL, WNUT17) running various NLP tasks. Our study finds that executing complex NLP tasks (such as "sentiment" classification) on embedded systems is feasible even without any GPUs (e.g., Raspberry Pi with 2 GB of RAM). Our findings can help designers understand the deployability and performance of transformer language models, especially those based on BERT architectures.
引用
收藏
页码:211 / 222
页数:12
相关论文
共 50 条
  • [21] When Language Models Fall in Love: Animacy Processing in Transformer Language Models
    Hanna, Michael
    Belinkov, Yonatan
    Pezzelle, Sandro
    2023 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2023), 2023, : 12120 - 12135
  • [23] Unnatural language processing: How do language models handle machine-generated prompts?
    Kervadec, Corentin
    Franzon, Francesca
    Baroni, Marco
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EMNLP 2023), 2023, : 14377 - 14392
  • [24] HOW WELL DO MEDICAL DEVICES WORK
    FIELDER, J
    IEEE ENGINEERING IN MEDICINE AND BIOLOGY MAGAZINE, 1994, 13 (03): : 420 - 422
  • [25] Dementia Detection using Transformer-Based Deep Learning and Natural Language Processing Models
    Saltz, Ploypaphat
    Lin, Shih Yin
    Cheng, Sunny Chieh
    Si, Dong
    2021 IEEE 9TH INTERNATIONAL CONFERENCE ON HEALTHCARE INFORMATICS (ICHI 2021), 2021, : 509 - 510
  • [26] How Much Do Modifications to Transformer Language Models Affect Their Ability to Learn Linguistic Knowledge?
    Sun, Simeng
    Dillon, Brian
    Iyyer, Mohit
    PROCEEDINGS OF THE THIRD WORKSHOP ON INSIGHTS FROM NEGATIVE RESULTS IN NLP (INSIGHTS 2022), 2022, : 46 - 53
  • [27] Pilot mental workload: how well do pilots really perform?
    Morris, Charles H.
    Leung, Ying K.
    ERGONOMICS, 2006, 49 (15) : 1581 - 1596
  • [28] Routine serum creatinine measurements: how well do we perform?
    Hoste, Liesbeth
    Deiteren, Kathleen
    Pottel, Hans
    Callewaert, Nico
    Martens, Frank
    BMC NEPHROLOGY, 2015, 16
  • [29] Routine serum creatinine measurements: how well do we perform?
    Liesbeth Hoste
    Kathleen Deiteren
    Hans Pottel
    Nico Callewaert
    Frank Martens
    BMC Nephrology, 16
  • [30] How well do dingoes, Canis dingo, perform on the detour task?
    Smith, Bradley Philip
    Litchfield, Carla Anita
    ANIMAL BEHAVIOUR, 2010, 80 (01) : 155 - 162