HBert: A Long Text Processing Method Based on BERT and Hierarchical Attention Mechanisms

被引:3
|
作者
Lv, Xueqiang [1 ]
Liu, Zhaonan [1 ]
Zhao, Ying [1 ]
Xu, Ge [2 ]
You, Xindong [1 ]
机构
[1] Beijing Informat Sci & Technol Univ, Beijing, Peoples R China
[2] Minjiang Univ, Fuzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
BERT; Hierarchical Attention; Long Text Processing;
D O I
10.4018/IJSWIS.322769
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
With the emergence of a large-scale pre-training model based on the transformer model, the effect of all-natural language processing tasks has been pushed to a new level. However, due to the high complexity of the transformer's self-attention mechanism, these models have poor processing ability for long text. Aiming at solving this problem, a long text processing method named HBert based on Bert and hierarchical attention neural network is proposed. Firstly, the long text is divided into multiple sentences whose vectors are obtained through the word encoder composed of Bert and the word attention layer. And the article vector is obtained through the sentence encoder that is composed of transformer and sentence attention. Then the article vector is used to complete the subsequent tasks. The experimental results show that the proposed HBert method achieves good results in text classification and QA tasks. The F1 value is 95.7% in longer text classification tasks and 75.2% in QA tasks, which are better than the state-of-the-art model longformer.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Sentiment analysis method of consumer comment text based on BERT and hierarchical attention in e-commerce big data environment
    Chang, Wanjun
    Zhu, Mingdong
    JOURNAL OF INTELLIGENT SYSTEMS, 2023, 32 (01)
  • [2] Enhancing text classification with attention matrices based on BERT
    Yu, Zhiyi
    Li, Hong
    Feng, Jialin
    EXPERT SYSTEMS, 2024, 41 (03)
  • [3] A Long-Text Classification Method of Chinese News Based on BERT and CNN
    Chen, Xinying
    Cong, Peimin
    Lv, Shuo
    IEEE ACCESS, 2022, 10 : 34046 - 34057
  • [4] A Multiscale Interactive Attention Short Text Classification Model Based on BERT
    Zhou, Lu
    Wang, Peng
    Zhang, Huijun
    Wu, Shengbo
    Zhang, Tao
    IEEE ACCESS, 2024, 12 : 160992 - 161001
  • [5] TEXT RECOGNITION IN IMAGES BASED ON TRANSFORMER WITH HIERARCHICAL ATTENTION
    Zhu, Yiwei
    Wang, Shilin
    Huang, Zheng
    Chen, Kai
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 1945 - 1949
  • [6] A Hierarchical Neural Attention-based Text Classifier
    Sinha, Koustuv
    Dong, Yue
    Cheung, Jackie C. K.
    Ruths, Derek
    2018 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP 2018), 2018, : 817 - 823
  • [7] Language Models for Hierarchical Classification of Radiology Reports With Attention Mechanisms, BERT, and GPT-4
    Olivato, Matteo
    Putelli, Luca
    Arici, Nicola
    Emilio Gerevini, Alfonso
    Lavelli, Alberto
    Serina, Ivan
    IEEE ACCESS, 2024, 12 : 69710 - 69727
  • [8] BVMHA: Text classification model with variable multihead hybrid attention based on BERT
    Peng, Bo
    Zhang, Tao
    Han, Kundong
    Zhang, Zhe
    Ma, Yuquan
    Ma, Mengnan
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2024, 46 (01) : 1443 - 1454
  • [9] Research on News Text Classification Based on BERT-BiLSTM-TextCNN-Attention
    Wang, Jia
    Li, Zongting
    Ma, Chenyang
    PROCEEDINGS OF 2024 3RD INTERNATIONAL CONFERENCE ON CYBER SECURITY, ARTIFICIAL INTELLIGENCE AND DIGITAL ECONOMY, CSAIDE 2024, 2024, : 295 - 298
  • [10] BERT for the Processing of Radiological Reports: An Attention-based Natural Language Processing Algorithm
    Soffer, Shelly
    Glicksberg, Benjamin S.
    Zimlichman, Eyal
    Klang, Eyal
    ACADEMIC RADIOLOGY, 2022, 29 (04) : 634 - 635