Pre-trained Transformer-Based Citation Context-Aware Citation Network Embeddings

被引:5
|
作者
Ohagi, Masaya [1 ]
Aizawa, Akiko [1 ,2 ]
机构
[1] Univ Tokyo, Tokyo, Japan
[2] Natl Inst Informat, Tokyo, Japan
关键词
Network embedding; citation context; citation recommendation;
D O I
10.1145/3529372.3533290
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Academic papers form citation networks wherein each paper is a node and citation relationships between papers are edges. The embeddings of each paper obtained by projecting the citation network into a vector space are called citation network embeddings. Thus far, only a limited number of studies have focused on incorporating information regarding the intent of one paper to cite another paper. We consider citation context, i.e., the text to cite a paper, as a source of information for citation intent, and propose a new method for generating citation context-aware citation network embeddings. We trained SciBERT with our proposed masked paper prediction task in which the model predicts the cited paper from the citing paper and the citation context. In addition, we propose a new loss function that considers not only the citation context but also the neighboring nodes in the citation network. We conducted experiments involving citation-recommendation and paper-classification tasks which we formulated on two existing datasets: FullTextPeerRead and AASC. For both tasks, the proposed method outperformed hyperdoc2vec, an existing method for citation context-aware citation network embedding; further, it achieved a comparable performance to a state-of-the-art citation network embedding that do not utilize any citation context for paper classification.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] Neural Citation Network for Context-Aware Citation Recommendation
    Ebesu, Travis
    Fang, Yi
    [J]. SIGIR'17: PROCEEDINGS OF THE 40TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2017, : 1093 - 1096
  • [2] A survey of transformer-based multimodal pre-trained modals
    Han, Xue
    Wang, Yi-Tong
    Feng, Jun-Lan
    Deng, Chao
    Chen, Zhan-Heng
    Huang, Yu-An
    Su, Hui
    Hu, Lun
    Hu, Peng-Wei
    [J]. NEUROCOMPUTING, 2023, 515 : 89 - 106
  • [3] Pre-trained transformer-based language models for Sundanese
    Wilson Wongso
    Henry Lucky
    Derwin Suhartono
    [J]. Journal of Big Data, 9
  • [4] Pre-trained transformer-based language models for Sundanese
    Wongso, Wilson
    Lucky, Henry
    Suhartono, Derwin
    [J]. JOURNAL OF BIG DATA, 2022, 9 (01)
  • [5] Pre-trained Transformer-based Classification for Automated Patentability Examination
    Lo, Hao-Cheng
    Chu, Jung-Mei
    [J]. 2021 IEEE ASIA-PACIFIC CONFERENCE ON COMPUTER SCIENCE AND DATA ENGINEERING (CSDE), 2021,
  • [6] A LSTM Based Model for Personalized Context-Aware Citation Recommendation
    Yang, Libin
    Zheng, Yu
    Cai, Xiaoyan
    Dai, Hang
    Mu, Dejun
    Guo, Lantian
    Dai, Tao
    [J]. IEEE ACCESS, 2018, 6 : 59618 - 59627
  • [7] Towards a Transformer-Based Pre-trained Model for IoT Traffic Classification
    Bazaluk, Bruna
    Hamdan, Mosab
    Ghaleb, Mustafa
    Gismalla, Mohammed S. M.
    da Silva, Flavio S. Correa
    Batista, Daniel Macedo
    [J]. PROCEEDINGS OF 2024 IEEE/IFIP NETWORK OPERATIONS AND MANAGEMENT SYMPOSIUM, NOMS 2024, 2024,
  • [8] Augmenting context-aware citation recommendations with citation and co-authorship history
    Bhowmick, Anubrata
    Singhal, Ashish
    Wang, Shenghui
    [J]. 18TH INTERNATIONAL CONFERENCE ON SCIENTOMETRICS & INFORMETRICS (ISSI2021), 2021, : 115 - 120
  • [9] Transformer based contextualization of pre-trained word embeddings for irony detection in Twitter
    Angel Gonzalez, Jose
    Hurtado, Lluis-F
    Pla, Ferran
    [J]. INFORMATION PROCESSING & MANAGEMENT, 2020, 57 (04)
  • [10] Multi-task Active Learning for Pre-trained Transformer-based Models
    Rotman, Guy
    Reichart, Roi
    [J]. TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2022, 10 : 1209 - 1228