Early Rumor Detection based on Data Augmentation and Pre-training Transformer

被引:2
|
作者
Hu, Yanjun [1 ]
Ju, Xinyi [1 ]
Ye, Zhousheng [1 ]
Khan, Sulaiman [1 ]
Yuan, Chengwu [1 ]
Lai, Qiran [1 ]
Liu, Junqiang [1 ]
机构
[1] Zhejiang Gongshang Univ, Sch Informat & Elect Engn, Hangzhou, Peoples R China
基金
中国国家自然科学基金;
关键词
Transformer; data augmentation; rumor detect; deep learning;
D O I
10.1109/CCWC54503.2022.9720776
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
The development of online social networks such as Sina Weibo, Twitter and Facebook has completely changed the way people communicate. While online social media makes it easier for people to get information, it also results in proliferation and wide spread of rumors. Automatic detection of rumors has become a hot topic of research. However, detecting rumors in the early stage is quite challenging. This paper proposes a new model (DAPT) for early rumor detection by analyzing text features, using pre-training technique, and employing data augmentation method. Experiments on BuzzFeed, PolitiFact, and Twitter16 datasets show that the accuracy and F1-score by the proposed model are better than the existing models.
引用
收藏
页码:152 / 158
页数:7
相关论文
共 50 条
  • [1] Neural Language Model Based Training Data Augmentation for Weakly Supervised Early Rumor Detection
    Han, Sooji
    Gao, Jie
    Ciravegna, Fabio
    [J]. PROCEEDINGS OF THE 2019 IEEE/ACM INTERNATIONAL CONFERENCE ON ADVANCES IN SOCIAL NETWORKS ANALYSIS AND MINING (ASONAM 2019), 2019, : 105 - 112
  • [2] Rumor knowledge embedding based data augmentation for imbalanced rumor detection
    Chen, Xiangyan
    Zhu, Duoduo
    Lin, Dazhen
    Cao, Donglin
    [J]. INFORMATION SCIENCES, 2021, 580 : 352 - 370
  • [3] MAE-EEG-Transformer: A transformer-based approach combining masked autoencoder and cross-individual data augmentation pre-training for EEG classification
    Cai, Miao
    Zeng, Yu
    [J]. BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 94
  • [4] Survey: Transformer based video-language pre-training
    Ruan, Ludan
    Jin, Qin
    [J]. AI OPEN, 2022, 3 : 1 - 13
  • [5] TRANSFORMER BASED UNSUPERVISED PRE-TRAINING FOR ACOUSTIC REPRESENTATION LEARNING
    Zhang, Ruixiong
    Wu, Haiwei
    Li, Wubo
    Jiang, Dongwei
    Zou, Wei
    Li, Xiangang
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 6933 - 6937
  • [6] Learning Visual Robotic Control Efficiently with Contrastive Pre-training and Data Augmentation
    Zhan, Albert
    Zhao, Ruihan
    Pinto, Lerrel
    Abbeel, Pieter
    Laskin, Michael
    [J]. 2022 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2022, : 4040 - 4047
  • [7] Pulling Up by the Causal Bootstraps: Causal Data Augmentation for Pre-training Debiasing
    Gowda, Sindhu C. M.
    Joshi, Shalmali
    Zhang, Haoran
    Ghassemi, Marzyeh
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, CIKM 2021, 2021, : 606 - 616
  • [8] LuxemBERT: Simple and Practical Data Augmentation in Language Model Pre-Training for Luxembourgish
    Lothritz, Cedric
    Lebichot, Bertrand
    Allix, Kevin
    Veiber, Lisa
    Bissyande, Tegawende F.
    Klein, Jacques
    Boytsov, Andrey
    Goujon, Anne
    Lefebvre, Clement
    [J]. LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 5080 - 5089
  • [9] Pre-training and Evaluating Transformer-based Language Models for Icelandic
    Daoason, Jon Friorik
    Loftsson, Hrafn
    [J]. LREC 2022: THIRTEEN INTERNATIONAL CONFERENCE ON LANGUAGE RESOURCES AND EVALUATION, 2022, : 7386 - 7391
  • [10] Long Document Extractive Summarization Method Based on Pre-training Model and Transformer
    Zhou, Xinxin
    Guo, Yuechen
    Huang, Yuning
    Yan, Yuming
    Li, Maoyuan
    [J]. Journal of Network Intelligence, 2023, 8 (03): : 913 - 931