Pay attention to the hidden semanteme

被引:1
|
作者
Tang, Huanling [1 ,2 ,3 ]
Liu, Xiaoyan [1 ]
Wang, Yulin [1 ]
Dou, Quansheng [1 ,2 ,3 ]
Lu, Mingyu [4 ]
机构
[1] Shandong Technol & Business Univ, Sch Comp Sci & Technol, Yantai 264005, Shandong, Peoples R China
[2] Coinnovat Ctr Shandong Coll & Univ Future Intellig, Yantai 264005, Shandong, Peoples R China
[3] Shandong Technol & Business Univ, Key Lab Intelligent Informat Proc Univ Shandong, Yantai 264005, Shandong, Peoples R China
[4] Dalian Maritime Univ, Informat Sci & Technol Coll, Dalian 116026, Liaoning, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature representation; Attention mechanism; Deep learning; Modeling lightly; Natural language processing;
D O I
10.1016/j.ins.2023.119076
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
With the capability of modeling lighter, MLP-based models like the pNLP-Mixer and the HyperMixer demonstrate the potential for diverse tasks in NLP. However, these linguistic models are not optimized for the regularity of textual hierarchical abstraction. Here, this paper proposes the hidden bias attention (HBA), a novel attention mechanism that is lighter than self-attention and focuses on extracting hidden (topic) semanteme. Additionally, this paper introduces a series of lightweight deep learning architectures, HBA-Mixer based on HBA and MHBA-Mixers based on multi-head HBA, which both outperforms pNLP-Mixer and HyperMixer in accuracy with fewer parameters on 3 tasks, including text classification, natural language inference, and sentiment analysis. Compared with large pre-trained models, MHBA-Mixers achieve over 90% of their accuracy with one-thousandth of the parameters.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Does It Pay to Pay Attention?
    Gargano, Antonio
    Rossi, Alberto G.
    [J]. REVIEW OF FINANCIAL STUDIES, 2018, 31 (12): : 4595 - 4649
  • [2] Pay Attention or Pay the Price
    Suchy, Adam
    [J]. AGRESIVITA NA CESTACH, 2009, : 37 - +
  • [3] PAY ATTENTION
    HERMAN, G
    [J]. MICROCOMPUTING, 1983, 7 (08): : 28 - 28
  • [4] Pay Attention
    Jaekl, Phil
    [J]. SCIENTIST, 2018, 32 (12): : 15 - 17
  • [5] Pay attention
    [J]. Aviat Week Space Technol (New York), 2006, 20 (35-36):
  • [6] Pay attention
    Flanagan, PR
    [J]. LAB ANIMAL, 1998, 27 (02) : 20 - 20
  • [7] PAY ATTENTION
    TRABASSO, T
    [J]. PSYCHOLOGY TODAY, 1968, 2 (05) : 30 - 36
  • [8] Pay attention
    Wieder, S
    [J]. NEW SCIENTIST, 2002, 174 (2340) : 60 - 61
  • [9] Pay Attention
    不详
    [J]. ART IN AMERICA, 2022, 110 (12): : 8 - 8
  • [10] Pay Attention
    Mountford, Peter
    [J]. PARIS REVIEW, 2017, (223): : 49 - 68