TNT: Text Normalization based Pre-training of Transformers for Content Moderation

被引:0
|
作者
Tan, Fei [1 ]
Hu, Yifan [1 ]
Hu, Changwei [1 ]
Li, Keqian [1 ]
Yen, Kevin [1 ]
机构
[1] Yahoo Res, New York, NY 10003 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we present a new language pre-training model TNT (Text Normalization based pre-training of Transformers) for content moderation. Inspired by the masking strategy and text normalization, TNT is developed to learn language representation by training transformers to reconstruct text from four operation types typically seen in text manipulation: substitution, transposition, deletion, and insertion. Furthermore, the normalization involves the prediction of both operation types and token labels, enabling TNT to learn from more challenging tasks than the standard task of masked word recovery. As a result, the experiments demonstrate that TNT outperforms strong baselines on the hate speech classification task. Additional text normalization experiments and case studies show that TNT is a new potential approach to misspelling correction.
引用
收藏
页码:4735 / 4741
页数:7
相关论文
共 50 条
  • [1] Unsupervised Pre-Training for Detection Transformers
    Dai, Zhigang
    Cai, Bolun
    Lin, Yugeng
    Chen, Junying
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (11) : 12772 - 12782
  • [2] Pre-Training Transformers as Energy-Based Cloze Models
    Clark, Kevin
    Luong, Minh-Thang
    Le, Quoc V.
    Manning, Christopher D.
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 285 - 294
  • [3] Evaluation of FractalDB Pre-training with Vision Transformers
    Nakashima K.
    Kataoka H.
    Satoh Y.
    Seimitsu Kogaku Kaishi/Journal of the Japan Society for Precision Engineering, 2023, 89 (01): : 99 - 104
  • [4] RecGPT: Generative Pre-training for Text-based Recommendation
    Mang Ngo
    Dat Quoc Nguyen
    PROCEEDINGS OF THE 62ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 2: SHORT PAPERS, 2024, : 302 - 313
  • [5] Pre-training of Graph Augmented Transformers for Medication Recommendation
    Shang, Junyuan
    Ma, Tengfei
    Xiao, Cao
    Sun, Jimeng
    PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 5953 - 5959
  • [6] Lifting the Curse of Multilinguality by Pre-training Modular Transformers
    Pfeiffer, Jonas
    Goyal, Naman
    Lin, Xi Victoria
    Li, Xian
    Cross, James
    Riedel, Sebastian
    Artetxe, Mikel
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 3479 - 3495
  • [7] Deep Pre-Training Transformers for Scientific Paper Representation
    Wang, Jihong
    Yang, Zhiguang
    Cheng, Zhanglin
    ELECTRONICS, 2024, 13 (11)
  • [8] Unsupervised Extractive Summarization by Pre-training Hierarchical Transformers
    Xu, Shusheng
    Zhang, Xingxing
    Wu, Yi
    Wei, Furu
    Zhou, Ming
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2020, 2020, : 1784 - 1795
  • [9] Factored Phrase-Based Statistical Machine Pre-training with Extended Transformers
    Beyala, Vivien L.
    Li Litet, Perrin
    Nkenlifack, Marcellin J.
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2020, 11 (09) : 51 - 59
  • [10] TUTA: Tree-based Transformers for Generally Structured Table Pre-training
    Wang, Zhiruo
    Dong, Haoyu
    Jia, Ran
    Li, Jia
    Fu, Zhiyi
    Han, Shi
    Zhang, Dongmei
    KDD '21: PROCEEDINGS OF THE 27TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2021, : 1780 - 1790