PAL-BERT: An Improved Question Answering Model

被引:51
|
作者
Zheng, Wenfeng [1 ]
Lu, Siyu [1 ]
Cai, Zhuohang [1 ]
Wang, Ruiyang [1 ]
Wang, Lei [2 ]
Yin, Lirong [2 ]
机构
[1] Univ Elect Sci & Technol China, Sch Automat, Chengdu 610054, Peoples R China
[2] Louisiana State Univ, Dept Geog & Anthropol, Baton Rouge, LA 70803 USA
来源
关键词
PAL-BERT; question answering model; pretraining language models; ALBERT; pruning model; network pruning; TextCNN; BiLSTM;
D O I
10.32604/cmes.2023.046692
中图分类号
T [工业技术];
学科分类号
08 ;
摘要
In the field of natural language processing (NLP), there have been various pre-training language models in recent years, with question answering systems gaining significant attention. However, as algorithms, data, and computing power advance, the issue of increasingly larger models and a growing number of parameters has surfaced. Consequently, model training has become more costly and less efficient. To enhance the efficiency and accuracy of the training process while reducing the model volume, this paper proposes a first-order pruning model PAL-BERT based on the ALBERT model according to the characteristics of question-answering (QA) system and language model. Firstly, a first-order network pruning method based on the ALBERT model is designed, and the PAL-BERT model is formed. Then, the parameter optimization strategy of the PAL-BERT model is formulated, and the Mish function was used as an activation function instead of ReLU to improve the performance. Finally, after comparison experiments with traditional deep learning models TextCNN and BiLSTM, it is confirmed that PALBERT is a pruning model compression method that can significantly reduce training time and optimize training efficiency. Compared with traditional models, PAL-BERT significantly improves the NLP task's performance.
引用
下载
收藏
页码:2729 / 2745
页数:17
相关论文
共 50 条
  • [1] DPAL-BERT: A Faster and Lighter Question Answering Model
    Yin, Lirong
    Wang, Lei
    Cai, Zhuohang
    Lu, Siyu
    Wang, Ruiyang
    Alsanad, Ahmed
    Alqahtani, Salman A.
    Chen, Xiaobing
    Yin, Zhengtong
    Li, Xiaolu
    Zheng, Wenfeng
    CMES-COMPUTER MODELING IN ENGINEERING & SCIENCES, 2024, 141 (01): : 771 - 786
  • [2] BERT Representations for Video Question Answering
    Yang, Zekun
    Garcia, Noa
    Chu, Chenhui
    Otani, Mayu
    Nakashima, Yuta
    Takemura, Haruo
    2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 1545 - 1554
  • [3] A BERT-Based Model for Question Answering on Construction Incident Reports
    Hassan, Hebatallah A. Mohamed
    Marengo, Elisa
    Nutt, Werner
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2022), 2022, 13286 : 215 - 223
  • [4] CSA-BERT: Video Question Answering
    Jenni, Kommineni
    Srinivas, M.
    Sannapu, Roshni
    Perumal, Murukessan
    2023 IEEE STATISTICAL SIGNAL PROCESSING WORKSHOP, SSP, 2023, : 532 - 536
  • [5] Multi-passage BERT: A Globally Normalized BERT Model for Open-domain Question Answering
    Wang, Zhiguo
    Ng, Patrick
    Ma, Xiaofei
    Nallapati, Ramesh
    Xiang, Bing
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 5878 - 5882
  • [6] Question Answering Model Based Conversational Chatbot using BERT Model and Google Dialogflow
    Kanodia, Nikita
    Ahmed, Khandakar
    Miao, Yuan
    2021 31ST INTERNATIONAL TELECOMMUNICATION NETWORKS AND APPLICATIONS CONFERENCE (ITNAC), 2021, : 19 - 22
  • [7] Real Life Application of a Question Answering System Using BERT Language Model
    Alloatti, Francesca
    Di Caro, Luigi
    Sportelli, Gianpiero
    20TH ANNUAL MEETING OF THE SPECIAL INTEREST GROUP ON DISCOURSE AND DIALOGUE (SIGDIAL 2019), 2019, : 250 - 253
  • [8] A Legal Multi-Choice Question Answering Model Based on BERT and Attention
    Chen, Guibin
    Luo, Xudong
    Zhu, Junlin
    KNOWLEDGE SCIENCE, ENGINEERING AND MANAGEMENT, PT IV, KSEM 2023, 2023, 14120 : 250 - 266
  • [9] BERT with History Answer Embedding for Conversational Question Answering
    Qu, Chen
    Yang, Liu
    Qiu, Minghui
    Croft, W. Bruce
    Zhang, Yongfeng
    Iyyer, Mohit
    PROCEEDINGS OF THE 42ND INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '19), 2019, : 1133 - 1136
  • [10] MMFT-BERT: Multimodal Fusion Transformer with BERT Encodings for Visual Question Answering
    Khan, Aisha Urooj
    Mazaheri, Amir
    Lobo, Niels Da Vitoria
    Shah, Mubarak
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 4648 - 4660