Harnessing Pre-Trained Neural Networks with Rules for Formality Style Transfer

被引:0
|
作者
Wang, Yunli [1 ]
Wu, Yu [2 ]
Mou, Lili [3 ]
Li, Zhoujun [1 ]
Chao, Wenhan [1 ]
机构
[1] Beihang Univ, State Key Lab Software Dev Environm, Beijing, Peoples R China
[2] Microsoft Res, Beijing, Peoples R China
[3] Univ Alberta, Dept Comp Sci, Edmonton, AB, Canada
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Formality text style transfer plays an important role in various NLP applications, such as non-native speaker assistants and child education. Early studies normalize informal sentences with rules, before statistical and neural models become a prevailing method in the field. While a rule-based system is still a common preprocessing step for formality style transfer in the neural era, it could introduce noise if we use the rules in a naive way such as data preprocessing. To mitigate this problem, we study how to harness rules into a state-of-the-art neural network that is typically pretrained on massive corpora. We propose three fine-tuning methods in this paper and achieve a new state-of-the-art on benchmark datasets.
引用
收藏
页码:3573 / 3578
页数:6
相关论文
共 50 条
  • [21] Investigating the Impact of Pre-trained Word Embeddings on Memorization in Neural Networks
    Thomas, Aleena
    Adelani, David Ifeoluwa
    Davody, Ali
    Mogadala, Aditya
    Klakow, Dietrich
    TEXT, SPEECH, AND DIALOGUE (TSD 2020), 2020, 12284 : 273 - 281
  • [22] Semantic Segmentation of Mammograms Using Pre-Trained Deep Neural Networks
    Prates, Rodrigo Leite
    Gomez-Flores, Wilfrido
    Pereira, Wagner
    2021 18TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, COMPUTING SCIENCE AND AUTOMATIC CONTROL (CCE 2021), 2021,
  • [23] Pre-trained convolutional neural networks as feature extractors for tuberculosis detection
    Lopes, U. K.
    Valiati, J. F.
    COMPUTERS IN BIOLOGY AND MEDICINE, 2017, 89 : 135 - 143
  • [24] Zero time waste in pre-trained early exit neural networks
    Wojcik, Bartosz
    Przewiezlikowski, Marcin
    Szatkowski, Filip
    Wolczyk, Maciej
    Balazy, Klaudia
    Krzepkowski, Bartlomiej
    Podolak, Igor
    Tabor, Jacek
    Smieja, Marek
    Trzcinski, Tomasz
    NEURAL NETWORKS, 2023, 168 : 580 - 601
  • [25] An iVector Extractor Using Pre-trained Neural Networks for Speaker Verification
    Zhang, Shanshan
    Zheng, Rong
    Xu, Bo
    2014 9TH INTERNATIONAL SYMPOSIUM ON CHINESE SPOKEN LANGUAGE PROCESSING (ISCSLP), 2014, : 73 - 77
  • [26] Harnessing pre-trained generalist agents for software engineering tasks
    Mindom, Paulina Stevia Nouwou
    Nikanjam, Amin
    Khomh, Foutse
    EMPIRICAL SOFTWARE ENGINEERING, 2025, 30 (01)
  • [27] Adaptive Prompt Routing for Arbitrary Text Style Transfer with Pre-trained Language Models
    Liu, Qingyi
    Qin, Jinghui
    Ye, Wenxuan
    Mou, Hao
    He, Yuxuan
    Wang, Keze
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 18689 - 18697
  • [28] CONVOLUTIONAL NEURAL NETWORKS FOR OMNIDIRECTIONAL IMAGE QUALITY ASSESSMENT: PRE-TRAINED OR RE-TRAINED?
    Sendjasni, Abderrezzaq
    Larabi, Mohamed-Chaker
    Cheikh, Faouzi Alaya
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 3413 - 3417
  • [29] Transfer Learning for Mammogram Classification Using Pre-Trained Convolutional Neural Network
    Yasuda, K.
    Tsuru, H.
    Ohki, M.
    MEDICAL PHYSICS, 2017, 44 (06) : 3102 - 3102
  • [30] Comparative Analysis of Pre-trained Deep Neural Networks for Plant Disease Classification
    George, Romiyal
    Thuseethan, Selvarajah
    Ragel, Roshan G.
    2024 21ST INTERNATIONAL JOINT CONFERENCE ON COMPUTER SCIENCE AND SOFTWARE ENGINEERING, JCSSE 2024, 2024, : 179 - 186