Applications of Pruning Methods in Natural Language Processing

被引:0
|
作者
Touheed, Marva [1 ]
Zubair, Urooj [1 ]
Sabir, Dilshad [1 ]
Hassan, Ali [2 ]
Butt, Muhammad Fasih Uddin [1 ,3 ]
Riaz, Farhan [4 ]
Abdul, Wadood [5 ]
Ayub, Rashid [6 ]
机构
[1] COMSATS Univ Islamabad, Dept Elect & Comp Engn, Islamabad 45550, Pakistan
[2] Natl Univ Sci & Technol, Coll Elect & Mech Engn, Dept Comp & Software Engn, Islamabad 44000, Pakistan
[3] COMSATS Univ Islamabad, Next Generat Commun Res Grp, Islamabad 45550, Pakistan
[4] Univ Lincoln, Coll Hlth & Sci, Sch Comp Sci, Lincoln LN6 7DL, England
[5] King Saud Univ, Coll Comp & Informat Sci, Dept Comp Engn, Riyadh 11543, Saudi Arabia
[6] King Saud Univ, Coll Comp & Informat Sci, Dept Sci Technol & Innovat, Riyadh 11543, Saudi Arabia
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Computational modeling; Data models; Task analysis; Neural networks; Training; Sentiment analysis; Natural language processing; Convolutional neural networks; Artificial neural networks; Pruning; convolution neural networks; natural language processing; DNN; model compression; acceleration;
D O I
10.1109/ACCESS.2024.3411776
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep neural networks (DNN) are in high demand because of their widespread applications in natural language processing, image processing, and a lot of other domains. However, due to their computational expense, over-parameterization, and large memory requirements, DNN applications often require the use of substantial model resources. This strict requirement of latency and limited memory availability are hurdles in the device deployment of these technologies. Therefore, a common idea could be to mitigate the DNN-based models' size without any performance degradation using different compression techniques. During the last few years, a great deal of progress has been made in the field of Natural Language Processing (NLP) using deep learning approaches. The objective of this research is to offer a thorough overview of the various pruning methods applied in the context of NLP. In this paper, we review several recent pruning-based schemes used for converting standard networks into their compact and accelerated versions. Traditionally, pruning is a technique for improving latency, reducing model size, and computational complexity which is a viable approach to deal with the above-mentioned challenges. In general, these techniques are divided into two main categories: structural and unstructured pruning methods. Structural pruning methods are further classified into filter, channel, layer, block, and movement pruning. Whereas, neuron, magnitude-based, and iterative pruning lie in the category of unstructured pruning. For each method, we discuss the related metrics and benchmarks. Then recent work on each method is discussed in detail, which provides insightful analysis of the performance, related applications, and pros and cons. Then, a comparative analysis is provided to analyze the differences among approaches. Finally, the paper concludes with possible future directions and some technical challenges.
引用
收藏
页码:89418 / 89438
页数:21
相关论文
共 50 条
  • [1] Advances in Pruning and Quantization for Natural Language Processing
    Bibi, Ummara
    Mazhar, Mahrukh
    Sabir, Dilshad
    Fasih Uddin Butt, Muhammad
    Hassan, Ali
    Ali Ghazanfar, Mustansar
    Ali Khan, Arshad
    Abdul, Wadood
    [J]. IEEE Access, 2024, 12 : 139113 - 139128
  • [2] Editorial: Methods and applications of natural language processing in psychiatry research
    Wang, Li
    Li, Shuyan
    Chen, Hui
    Zhou, Yunyun
    [J]. FRONTIERS IN PSYCHIATRY, 2022, 13
  • [3] Applications of natural language processing
    Blandon Andrade, Juan Carlos
    [J]. ENTRE CIENCIA E INGENIERIA, 2022, 16 (31): : 7 - 8
  • [4] Systematic review of current natural language processing methods and applications in cardiology
    Turchioe, Meghan Reading
    Volodarskiy, Alexander
    Pathak, Jyotishman
    Wright, Drew N.
    Tcheng, James Enlou
    Slotwiner, David
    [J]. HEART, 2022, 108 (12) : 909 - 916
  • [5] Applications of natural language processing in construction
    Ding, Yuexiong
    Ma, Jie
    Luo, Xiaowei
    [J]. AUTOMATION IN CONSTRUCTION, 2022, 136
  • [6] Special issue "Deep Learning for Natural Language Processing: Emerging methods and applications"
    Esposito, Massimo
    Fujita, Hamido
    Minutolo, Aniello
    Pota, Marco
    [J]. ARRAY, 2022, 14
  • [7] Tolerance-based granular methods: Foundations and applications in natural language processing
    Ramanna, Sheela
    [J]. INTELLIGENT DECISION TECHNOLOGIES-NETHERLANDS, 2023, 17 (01): : 139 - 158
  • [8] Efficient Methods for Natural Language Processing: A Survey
    Treviso, Marcos
    Lee, Ji-Ung
    Ji, Tianchu
    van Aken, Betty
    Cao, Qingqing
    Ciosici, Manuel R.
    Hassid, Michael
    Heafield, Kenneth
    Hooker, Sara
    Raffel, Colin
    Martins, Pedro H.
    Martins, Andre F. T.
    Forde, Jessica Zosa
    Milder, Peter
    Simpson, Edwin
    Slonim, Noam
    Dodge, Jesse
    Strubell, Emma
    Balasubramanian, Niranjan
    Derczynski, Leon
    Gurevych, Iryna
    Schwartz, Roy
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2023, 11 : 826 - 860
  • [9] Efficient Methods for Natural Language Processing: A Survey
    Treviso, Marcos
    Ji, Tianchu
    Lee, Ji-Ung
    van Aken, Betty
    Cao, Qingqing
    R. Ciosici, Manuel
    Hassid, Michael
    Heafield, Kenneth
    Hooker, Sara
    H. Martins, Pedro
    F. T. Martins, Andre
    Milder, Peter
    Raffel, Colin
    Simpson, Edwin
    Slonim, Noam
    Dodge, Jesse
    Strubell, Emma
    Balasubramanian, Niranjan
    Derczynski, Leon
    Gurevych, Iryna
    Schwartz, Roy
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2023, 11 (826-860) : 826 - 860
  • [10] Deep Learning Methods in Natural Language Processing
    Flores, Alexis Stalin Alulema
    [J]. APPLIED TECHNOLOGIES (ICAT 2019), PT II, 2020, 1194 : 92 - 107