Data Augmentation Based on Pre-trained Language Model for Event Detection

被引:0
|
作者
Zhang, Meng [1 ]
Xie, Zhiwen [1 ]
Liu, Jin [1 ]
机构
[1] Wuhan Univ, Sch Comp Sci, Wuhan 430072, Peoples R China
来源
CCKS 2021 - EVALUATION TRACK | 2022年 / 1553卷
关键词
Event detection; Pre-trained model; Data augmentation; Pseudo labelling;
D O I
10.1007/978-981-19-0713-5_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Event detection (ED) is an important task which needs to identify the event triggers in the sentence and classify the event types. For the general fine-grained event detection task, we propose an event detection scheme based on pre-trained model, combined with data augmentation and pseudo labelling method, which improves the event detection ability of the model. At the same time, we use voting for model ensemble, so as to effectively utilize the advantages of multiple models. Our model achieves F1 score of 69.86% on the test set of CCKS2021 general fine-grained event detection task and ranks the third place in the competition.
引用
收藏
页码:59 / 68
页数:10
相关论文
共 50 条
  • [21] BSTC: A Fake Review Detection Model Based on a Pre-Trained Language Model and Convolutional Neural Network
    Lu, Junwen
    Zhan, Xintao
    Liu, Guanfeng
    Zhan, Xinrong
    Deng, Xiaolong
    [J]. ELECTRONICS, 2023, 12 (10)
  • [22] EventBERT: A Pre-Trained Model for Event Correlation Reasoning
    Zhou, Yucheng
    Geng, Xiubo
    Shen, Tao
    Long, Guodong
    Jiang, Daxin
    [J]. PROCEEDINGS OF THE ACM WEB CONFERENCE 2022 (WWW'22), 2022, : 850 - 859
  • [23] DFEPT: Data Flow Embedding for Enhancing Pre-Trained Model Based Vulnerability Detection
    Jiang, Zhonghao
    Sun, Weifeng
    Gu, Xiaoyan
    Wu, Jiaxin
    Wen, Tao
    Hu, Haibo
    Yan, Meng
    [J]. PROCEEDINGS OF THE 15TH ASIA-PACIFIC SYMPOSIUM ON INTERNETWARE, INTERNETWARE 2024, 2024, : 95 - 104
  • [24] BERT-Log: Anomaly Detection for System Logs Based on Pre-trained Language Model
    Chen, Song
    Liao, Hai
    [J]. APPLIED ARTIFICIAL INTELLIGENCE, 2022, 36 (01)
  • [25] Surgicberta: a pre-trained language model for procedural surgical language
    Bombieri, Marco
    Rospocher, Marco
    Ponzetto, Simone Paolo
    Fiorini, Paolo
    [J]. INTERNATIONAL JOURNAL OF DATA SCIENCE AND ANALYTICS, 2024, 18 (01) : 69 - 81
  • [26] Schema matching based on energy domain pre-trained language model
    Pan Z.
    Yang M.
    Monti A.
    [J]. Energy Informatics, 2023, 6 (Suppl 1)
  • [27] A teacher action recognition model based on pre-trained language and video model
    Luo, Sen
    Zhou, Juxiang
    Wen, Xiaoyu
    Li, Hao
    [J]. PROCEEDINGS OF THE 15TH INTERNATIONAL CONFERENCE ON EDUCATION TECHNOLOGY AND COMPUTERS, ICETC 2023, 2023, : 335 - 340
  • [28] ViDeBERTa: A powerful pre-trained language model for Vietnamese
    Tran, Cong Dao
    Pham, Nhut Huy
    Nguyen, Anh
    Hy, Truong Son
    Vu, Tu
    [J]. 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1071 - 1078
  • [29] BERTweet: A pre-trained language model for English Tweets
    Dat Quoc Nguyen
    Thanh Vu
    Anh Tuan Nguyen
    [J]. PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING: SYSTEM DEMONSTRATIONS, 2020, : 9 - 14
  • [30] Pre-trained Language Model for Biomedical Question Answering
    Yoon, Wonjin
    Lee, Jinhyuk
    Kim, Donghyeon
    Jeong, Minbyul
    Kang, Jaewoo
    [J]. MACHINE LEARNING AND KNOWLEDGE DISCOVERY IN DATABASES, ECML PKDD 2019, PT II, 2020, 1168 : 727 - 740