Duplicate Question Detection based on Neural Networks and Multi-head Attention

被引:0
|
作者
Zhang, Heng [1 ]
Chen, Liangyu [1 ]
机构
[1] East China Normal Univ, Shanghai Key Lab Trustworthy Comp, Shanghai, Peoples R China
关键词
deep learning; multi-head attention; ensemble learning;
D O I
10.1109/ialp48816.2019.9037671
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
It is well known that using only one neural network can not get a satisfied accuracy for the problem of Duplicate Question Detection. In order to break through this dilemma, different neural networks are ensembled serially to strive for better accuracy. However, many problems, such as vanishing gradient or exploding gradient, will be encountered if the depth of neural network is blindly increased. Worse, the serial integration may be poor in computational performance since it is less parallelizable and needs more time to train. To solve these problems, we use ensemble learning with treating different neural networks as individual learners, calculating in parallel, and proposing a new voting mechanism to get better detection accuracy. In addition to the classical models based on recurrent or convolutional neural network, Multi Head Attention is also integrated to reduce the correlation and the performance gap between different models. The experimental results in Quora question pairs dataset show that the accuracy of our method can reach 89.3 %.
引用
收藏
页码:13 / 18
页数:6
相关论文
共 50 条
  • [1] Self Multi-Head Attention-based Convolutional Neural Networks for fake news detection
    Fang, Yong
    Gao, Jian
    Huang, Cheng
    Peng, Hua
    Wu, Runpu
    [J]. PLOS ONE, 2019, 14 (09):
  • [2] Combining Multi-Head Attention and Sparse Multi-Head Attention Networks for Session-Based Recommendation
    Zhao, Zhiwei
    Wang, Xiaoye
    Xiao, Yingyuan
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [3] Hierarchical Hybrid Neural Networks With Multi-Head Attention for Document Classification
    Huang, Weihao
    Chen, Jiaojiao
    Cai, Qianhua
    Liu, Xuejie
    Zhang, Yudong
    Hu, Xiaohui
    [J]. INTERNATIONAL JOURNAL OF DATA WAREHOUSING AND MINING, 2022, 18 (01)
  • [4] Improving Multi-head Attention with Capsule Networks
    Gu, Shuhao
    Feng, Yang
    [J]. NATURAL LANGUAGE PROCESSING AND CHINESE COMPUTING (NLPCC 2019), PT I, 2019, 11838 : 314 - 326
  • [5] A Study of Visual Question Answering Techniques Based on Collaborative Multi-Head Attention
    Yang, Yingli
    Jin, Jingxuan
    Li, De
    [J]. 2023 3RD ASIA-PACIFIC CONFERENCE ON COMMUNICATIONS TECHNOLOGY AND COMPUTER SCIENCE, ACCTCS, 2023, : 552 - 555
  • [6] Multi-head multi-order graph attention networks
    Ben, Jie
    Sun, Qiguo
    Liu, Keyu
    Yang, Xibei
    Zhang, Fengjun
    [J]. APPLIED INTELLIGENCE, 2024, 54 (17-18) : 8092 - 8107
  • [7] Detecting Credit Card Fraud by Generative Adversarial Networks and Multi-head Attention Neural Networks
    Meng, Zhaorui
    Xie, Yanqi
    Sun, Jinhua
    [J]. IAENG International Journal of Computer Science, 2023, 50 (02)
  • [8] Sarcasm Detection Using Multi-Head Attention Based Bidirectional LSTM
    Kumar, Avinash
    Narapareddy, Vishnu Teja
    Aditya Srikanth, Veerubhotla
    Malapati, Aruna
    Neti, Lalita Bhanu Murthy
    [J]. IEEE ACCESS, 2020, 8 : 6388 - 6397
  • [9] Enhanced Multi-Head Self-Attention Graph Neural Networks for Session-based Recommendation
    Pan, Wenhao
    Yang, Kai
    [J]. ENGINEERING LETTERS, 2022, 30 (01) : 37 - 44
  • [10] Acoustic Scene Analysis with Multi-head Attention Networks
    Wang, Weimin
    Wang, Weiran
    Sun, Ming
    Wang, Chao
    [J]. INTERSPEECH 2020, 2020, : 1191 - 1195