APTrans: Transformer-Based Multilayer Semantic and Locational Feature Integration for Efficient Text Classification

被引:0
|
作者
Ji, Gaoyang [1 ]
Chen, Zengzhao [1 ,2 ]
Liu, Hai [1 ,2 ]
Liu, Tingting [3 ]
Wang, Bing [1 ]
机构
[1] Cent China Normal Univ, Fac Artificial Intelligence Educ, Wuhan 430079, Peoples R China
[2] Cent China Normal Univ, Natl Engn Res Ctr Elearning, Wuhan 430079, Peoples R China
[3] Hubei Univ, Sch Educ, Wuhan 430072, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 11期
基金
中国国家自然科学基金;
关键词
text classification; feature fusion; T-PTLM; semantic information; deep learning;
D O I
10.3390/app14114863
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
Text classification is not only a prerequisite for natural language processing work, such as sentiment analysis and natural language reasoning, but is also of great significance for screening massive amounts of information in daily life. However, the performance of classification algorithms is always affected due to the diversity of language expressions, inaccurate semantic information, colloquial information, and many other problems. We identify three clues in this study, namely, core relevance information, semantic location associations, and the mining characteristics of deep and shallow networks for different information, to cope with these challenges. Two key insights about the text are revealed based on these three clues: key information relationship and word group inline relationship. We propose a novel attention feature fusion network, Attention Pyramid Transformer (APTrans), which is capable of learning the core semantic and location information from sentences using the above-mentioned two key insights. Specially, a hierarchical feature fusion module, Feature Fusion Connection (FFCon), is proposed to merge the semantic features of higher layers with positional features of lower layers. Thereafter, a Transformer-based XLNet network is used as the backbone to initially extract the long dependencies from statements. Comprehensive experiments show that APTrans can achieve leading results on the THUCNews Chinese dataset, AG News, and TREC-QA English dataset, outperforming most excellent pre-trained models. Furthermore, extended experiments are carried out on a self-built Chinese dataset theme analysis of teachers' classroom corpus. We also provide visualization work, further proving that APTrans has good potential in text classification work.
引用
收藏
页数:24
相关论文
共 50 条
  • [1] Transformer-based Bug/Feature Classification
    Ozturk, Ceyhun E.
    Yilmaz, Eyup Halit
    Koksal, Omer
    [J]. 2023 31ST SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE, SIU, 2023,
  • [2] BertSRC: transformer-based semantic relation classification
    Lee, Yeawon
    Son, Jinseok
    Song, Min
    [J]. BMC MEDICAL INFORMATICS AND DECISION MAKING, 2022, 22 (01)
  • [3] Practical Transformer-based Multilingual Text Classification
    Wang, Cindy
    Banko, Michele
    [J]. 2021 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, NAACL-HLT 2021, 2021, : 121 - 129
  • [4] BertSRC: transformer-based semantic relation classification
    Yeawon Lee
    Jinseok Son
    Min Song
    [J]. BMC Medical Informatics and Decision Making, 22
  • [5] Efficient feature integration with Wikipedia-based semantic feature extraction for Turkish text summarization
    Guran, Aysun
    Bayazit, Nilgun Guler
    Gurbuz, Mustafa Zahid
    [J]. TURKISH JOURNAL OF ELECTRICAL ENGINEERING AND COMPUTER SCIENCES, 2013, 21 (05) : 1411 - 1425
  • [6] Transformer-Based Composite Language Models for Text Evaluation and Classification
    Skoric, Mihailo
    Utvic, Milos
    Stankovic, Ranka
    [J]. MATHEMATICS, 2023, 11 (22)
  • [7] An Effective, Efficient, and Scalable Confidence-Based Instance Selection Framework for Transformer-Based Text Classification
    Cunha, Washington
    Franca, Celso
    Fonseca, Guilherme
    Rocha, Leonardo
    Goncalves, Marcos Andre
    [J]. PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 665 - 674
  • [8] Development of a Text Classification Framework using Transformer-based Embeddings
    Yeasmin, Sumona
    Afrin, Nazia
    Saif, Kashfia
    Huq, Mohammad Rezwanul
    [J]. PROCEEDINGS OF THE 11TH INTERNATIONAL CONFERENCE ON DATA SCIENCE, TECHNOLOGY AND APPLICATIONS (DATA), 2022, : 74 - 82
  • [9] Online Feature Classification and Clustering for Transformer-based Visual Tracker
    Zou, Zhuojun
    Hao, Jie
    Shu, Lin
    [J]. 2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 3514 - 3521
  • [10] MPT-SFANet: Multiorder Pooling Transformer-Based Semantic Feature Aggregation Network for SAR Image Classification
    Ni, Kang
    Yuan, Chunyang
    Zheng, Zhizhong
    Zhang, Bingbing
    Wang, Peng
    [J]. IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2024, 60 (04) : 4923 - 4938