Fast compression and optimization of deep learning models for natural language processing

被引:3
|
作者
Pietron, Marcin [1 ]
Karwatowski, Michal [1 ]
Wielgosz, Maciej [1 ]
Duda, Jerzy [2 ]
机构
[1] AGH Univ Sci & Technol, Dept Comp Sci, Krakow, Poland
[2] AGH Univ Sci & Technol, Dept Management, Krakow, Poland
关键词
NLP; deep learning; recurrent neural networks; pruning; quantization;
D O I
10.1109/CANDARW.2019.00036
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Nowadays, recurrent neural networks (RNN) and convolutional neural networks (CNN) play a major role in a lot of natural language domains like text document categorization, part of speech tagging, chatbots, language modeling or language translation. Very often RNN networks have a few stacked layers with several megabytes of memory, the same is in case of CNN networks. In many domains like automatic speech recognition the real time inference is a crucial factor to achieve satisfactory quality of service. Compressing the network layer parameters and outputs into a suitable precision formats and applying pruning process can reduce the required storage and computation cycles in embedded devices. It can drastically reduce the consumed power and the memory capacity. In this article, we present pruning and quantization on deep learning models used for sentiment analysis, language modelling and language translation. All of them with a minor degradation of performance metric compared to full floating-point version.
引用
收藏
页码:162 / 168
页数:7
相关论文
共 50 条
  • [1] Deep Learning in Natural Language Processing
    Feng, Haoda
    Shi, Feng
    [J]. NATURAL LANGUAGE ENGINEERING, 2021, 27 (03) : 373 - 375
  • [2] Deep learning of the natural language processing
    Allauzen, Alexandre
    Schuetze, Hinrich
    [J]. TRAITEMENT AUTOMATIQUE DES LANGUES, 2018, 59 (02): : 7 - 14
  • [3] An introduction to Deep Learning in Natural Language Processing: Models, techniques, and tools
    Lauriola, Ivano
    Lavelli, Alberto
    Aiolli, Fabio
    [J]. NEUROCOMPUTING, 2022, 470 : 443 - 456
  • [4] Deep Learning for Natural Language Processing and Language Modelling
    Klosowski, Piotr
    [J]. 2018 SIGNAL PROCESSING: ALGORITHMS, ARCHITECTURES, ARRANGEMENTS, AND APPLICATIONS (SPA), 2018, : 223 - 228
  • [5] On the Explainability of Natural Language Processing Deep Models
    El Zini, Julia
    Awad, Mariette
    [J]. ACM COMPUTING SURVEYS, 2023, 55 (05)
  • [6] Backdoor Learning of Language Models in Natural Language Processing
    University of Michigan
    [J]. 1600,
  • [7] Adversarial Attacks on Deep-learning Models in Natural Language Processing: A Survey
    Zhang, Wei Emma
    Sheng, Quan Z.
    Alhazmi, Ahoud
    Li, Chenliang
    [J]. ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2020, 11 (03)
  • [8] Context-Sensitive Visualization of Deep Learning Natural Language Processing Models
    Dunn, Andrew
    Inkpen, Diana
    Andonie, Razvan
    [J]. 2021 25TH INTERNATIONAL CONFERENCE INFORMATION VISUALISATION (IV): AI & VISUAL ANALYTICS & DATA SCIENCE, 2021, : 170 - 175
  • [9] Deep Learning on Graphs for Natural Language Processing
    Wu, Lingfei
    Chen, Yu
    Ji, Heng
    Liu, Bang
    [J]. KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 4084 - 4085
  • [10] Deep Learning Methods in Natural Language Processing
    Flores, Alexis Stalin Alulema
    [J]. APPLIED TECHNOLOGIES (ICAT 2019), PT II, 2020, 1194 : 92 - 107