Building siamese attention-augmented recurrent convolutional neural networks for document similarity scoring

被引:9
|
作者
Han, Sifei [1 ]
Shi, Lingyun [1 ]
Richie, Russell [1 ]
Tsui, Fuchiang R. Rich [1 ,2 ]
机构
[1] Childrens Hosp Philadelphia, Dept Biomed & Hlth Informat, Tsui Lab, 2716 South St, Philadelphia, PA 19104 USA
[2] Univ Penn, Perelman Sch Med, 3400 Spruce St,Suite 680 Dulles, Philadelphia, PA USA
基金
美国国家科学基金会;
关键词
Attention neural network; Deep learning; Machine learning; Natural language processing; Information retrieval; Text similarity;
D O I
10.1016/j.ins.2022.10.032
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Automatically measuring document similarity is imperative in natural language process-ing, with applications ranging from recommendation to duplicate document detection. State-of-the-art approach in document similarity commonly involves deep neural net-works, yet there is little study on how different architectures may be combined. Thus, we introduce the Siamese Attention-augmented Recurrent Convolutional Neural Network (S-ARCNN) that combines multiple neural network architectures. In each subnet-work of S-ARCNN, a document passes through a bidirectional Long Short-Term Memory (bi-LSTM) layer, which sends representations to local and global document modules. A local document module uses convolution, pooling, and attention layers, whereas a global document module uses last states of the bi-LSTM. Both local and global features are con-catenated to form a single document representation. Using the Quora Question Pairs data -set, we evaluated S-ARCNN, Siamese convolutional neural networks (S-CNNs), Siamese LSTM, and two BERT models. While S-CNNs (82.02% F1) outperformed S-ARCNN (79.83% F1) overall, S-ARCNN slightly outperformed S-CNN on duplicate question pairs with more than 50 words (39.96% vs. 39.42% accuracy). With the potential advantage of S-ARCNN for processing longer documents, S-ARCNN may help researchers identify collaborators with similar research interests, help editors find potential reviewers, or match resumes with job descriptions.(c) 2022 Published by Elsevier Inc.
引用
收藏
页码:90 / 102
页数:13
相关论文
共 50 条
  • [41] Temporal Attention-Augmented Graph Convolutional Network for Efficient Skeleton-Based Human Action Recognition
    Heidari, Negar
    Iosifidis, Alexandros
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7907 - 7914
  • [42] Convolutional Neural Networks for Document Image Classification
    Kang, Le
    Kumar, Jayant
    Ye, Peng
    Li, Yi
    Doermann, David
    2014 22ND INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2014, : 3168 - 3172
  • [43] Quantum Similarity Testing with Convolutional Neural Networks
    Wu, Ya-Dong
    Zhu, Yan
    Bai, Ge
    Wang, Yuexuan
    Chiribella, Giulio
    PHYSICAL REVIEW LETTERS, 2023, 130 (21)
  • [44] Convolutional Recurrent Neural Networks with a Self-Attention Mechanism for Personnel Performance Prediction
    Xue, Xia
    Feng, Jun
    Gao, Yi
    Liu, Meng
    Zhang, Wenyu
    Sun, Xia
    Zhao, Aiqi
    Guo, Shouxi
    ENTROPY, 2019, 21 (12)
  • [45] Speech Emotion Recognition Using Convolutional-Recurrent Neural Networks with Attention Model
    Mu, Yawei
    Gomez, Hernandez
    Cano Montes, Antonio
    Alcaraz Martinez, Carlos
    Wang, Xuetian
    Gao, Hongmin
    2ND INTERNATIONAL CONFERENCE ON COMPUTER ENGINEERING, INFORMATION SCIENCE AND INTERNET TECHNOLOGY, CII 2017, 2017, : 341 - 350
  • [46] Convolutional-Recurrent Neural Networks With Multiple Attention Mechanisms for Speech Emotion Recognition
    Jiang, Pengxu
    Xu, Xinzhou
    Tao, Huawei
    Zhao, Li
    Zou, Cairong
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2022, 14 (04) : 1564 - 1573
  • [47] Focal Cosine Metric and Adaptive Attention Module for Remote Sensing Scene Classification With Siamese Convolutional Neural Networks
    Min, Lei
    Gao, Kun
    Wang, Hong
    Liu, Yutong
    Zhang, Zhenzhou
    Hu, Zibo
    Zhang, Xiaodian
    IEEE ACCESS, 2022, 10 : 84212 - 84226
  • [48] Neural Architecture Search for Convolutional Neural Networks with Attention
    Nakai, Kohei
    Matsubara, Takashi
    Uehara, Kuniaki
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2021, E104D (02) : 312 - 321
  • [49] Scoring Summaries Using Recurrent Neural Networks
    Ruseti, Stefan
    Dascalu, Mihai
    Johnson, Amy M.
    McNamara, Danielle S.
    Balyan, Renu
    McCarthy, Kathryn S.
    Trausan-Matu, Stefan
    INTELLIGENT TUTORING SYSTEMS, ITS 2018, 2018, 10858 : 191 - 201
  • [50] Recurrent networks with attention and convolutional networks for sentence representation and classification
    Liu, Tengfei
    Yu, Shuangyuan
    Xu, Baomin
    Yin, Hongfeng
    APPLIED INTELLIGENCE, 2018, 48 (10) : 3797 - 3806