Co-Stack Residual Affinity Networks with Multi-level Attention Refinement for Matching Text Sequences

被引:0
|
作者
Tay, Yi [1 ]
Luu Anh Tuan [2 ]
Hui, Siu Cheung [1 ]
机构
[1] Nanyang Technol Univ, Singapore, Singapore
[2] ASTAR, Inst Infocomm Res, Singapore, Singapore
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning a matching function between two text sequences is a long standing problem in NLP research. This task enables many potential applications such as question answering and paraphrase identification. This paper proposes Co-Stack Residual Affinity Networks (CSRAN), a new and universal neural architecture for this problem. CSRAN is a deep architecture, involving stacked (multi-layered) recurrent encoders. Stacked/Deep architectures are traditionally difficult to train, due to the inherent weaknesses such as difficulty with feature propagation and vanishing gradients. CSRAN incorporates two novel components to take advantage of the stacked architecture. Firstly, it introduces a new bidirectional alignment mechanism that learns affinity weights by fusing sequence pairs across stacked hierarchies. Secondly, it leverages a multi-level attention refinement component between stacked recurrent layers. The key intuition is that, by leveraging information across all network hierarchies, we can not only improve gradient flow but also improve overall performance. We conduct extensive experiments on six well-studied text sequence matching datasets, achieving state-of-the-art performance on all.
引用
收藏
页码:4492 / 4502
页数:11
相关论文
共 50 条
  • [1] Multi-Level Matching Networks for Text Matching
    Xu, Chunlin
    Lin, Zhiwei
    Wu, Shengli
    Wang, Hui
    [J]. PROCEEDINGS OF THE 42ND INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '19), 2019, : 949 - 952
  • [2] A Multi-level Attention Model for Text Matching
    Sun, Qiang
    Wu, Yue
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2018, PT I, 2018, 11139 : 142 - 153
  • [3] Dilated residual networks with multi-level attention for speaker verification
    Wu, Yanfeng
    Guo, Chenkai
    Gao, Hongcan
    Xu, Jing
    Bai, Guangdong
    [J]. NEUROCOMPUTING, 2020, 412 : 177 - 186
  • [4] Multi-level Residual Attention Network for Speckle Suppression
    Lei, Yu
    Liu, Shuaiqi
    Zhang, Luyao
    Zhao, Ling
    Zhao, Jie
    [J]. PATTERN RECOGNITION AND COMPUTER VISION, PT IV, 2021, 13022 : 288 - 299
  • [5] Multi-Level Compare-Aggregate Model for Text Matching
    Xu, Chunlin
    Wang, Hui
    Lin, Zhiwei
    Wu, Shengli
    [J]. 2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,
  • [6] Visual Attention Dehazing Network with Multi-level Features Refinement and Fusion
    Yin, Shibai
    Yang, Xiaolong
    Wang, Yibin
    Yang, Yee-Hong
    [J]. PATTERN RECOGNITION, 2021, 118
  • [7] app Popularity Prediction with Multi-Level Attention Networks
    Zhang, Yixuan
    Guo, Bin
    Liu, Jiaqi
    Ouyang, Yi
    Yu, Zhiwen
    [J]. Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2020, 57 (05): : 984 - 995
  • [8] Multi-level Attention Networks for Visual Question Answering
    Yu, Dongfei
    Fu, Jianlong
    Mei, Tao
    Rui, Yong
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 4187 - 4195
  • [9] Frame-based Multi-level Semantics Representation for text matching
    Guo, Shaoru
    Guan, Yong
    Li, Ru
    Li, Xiaoli
    Tan, Hongye
    [J]. KNOWLEDGE-BASED SYSTEMS, 2021, 232
  • [10] Part Matching with Multi-level Attention for Person Re-Identification
    Wang, Jiaze
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 1805 - 1814