Multilingual Controllable Transformer-Based Lexical Simplification

被引:0
|
作者
Sheang, Kim Cheng [1 ]
Saggion, Horacio [1 ]
机构
[1] Univ Pompeu Fabra, LaSTUS Grp, TALN Lab, DTIC, Barcelona, Spain
来源
PROCESAMIENTO DEL LENGUAJE NATURAL | 2023年 / 71期
关键词
Multilingual Lexical Simplification; Controllable Lexical Simplification; Text Simplification; Multilinguality;
D O I
10.26342/2023-71-9
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Text is by far the most ubiquitous source of knowledge and information and should be made easily accessible to as many people as possible; however, texts often contain complex words that hinder reading comprehension and accessibility. Therefore, suggesting simpler alternatives for complex words without compromising meaning would help convey the information to a broader audience. This paper proposes mTLS, a multilingual controllable Transformer-based Lexical Simplification (LS) system fined-tuned with the T5 model. The novelty of this work lies in the use of language-specific prefixes, control tokens, and candidates extracted from pretrained masked language models to learn simpler alternatives for complex words. The evaluation results on three well-known LS datasets - LexMTurk, BenchLS, and NNSEval - show that our model outperforms the previous state-of-the-art models like LSBert and ConLS. Moreover, further evaluation of our approach on the part of the recent TSAR-2022 multilingual LS shared-task dataset shows that our model performs competitively when compared with the participating systems for English LS and even outperforms the GPT-3 model on several metrics. Moreover, our model obtains performance gains also for Spanish and Portuguese.
引用
收藏
页码:109 / 123
页数:15
相关论文
共 50 条
  • [41] Swin transformer-based supervised hashing
    Peng, Liangkang
    Qian, Jiangbo
    Wang, Chong
    Liu, Baisong
    Dong, Yihong
    APPLIED INTELLIGENCE, 2023, 53 (14) : 17548 - 17560
  • [42] Transformer-Based Deep Survival Analysis
    Hu, Shi
    Fridgeirsson, Egill A.
    van Wingen, Guido
    Welling, Max
    SURVIVAL PREDICTION - ALGORITHMS, CHALLENGES AND APPLICATIONS, VOL 146, 2021, 146 : 132 - 148
  • [43] Multi-label Discourse Function Classification of Lexical Bundles in Basque and Spanish via transformer-based models
    Goikoetxea, Josu
    Etxabe, Markel
    Garcia, Marcos
    Guzzi, Eleonora
    Alonso, Margarita
    PROCESAMIENTO DEL LENGUAJE NATURAL, 2024, (73): : 29 - 41
  • [44] Transformer-based fall detection in videos
    Nunez-Marcos, Adrian
    Arganda-Carreras, Ignacio
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 132
  • [45] A Compact Transformer-Based GAN Vocoder
    Miao, Chenfeng
    Chen, Ting
    Chen, Minchuan
    Ma, Jun
    Wang, Shaojun
    Xiao, Jing
    INTERSPEECH 2022, 2022, : 1636 - 1640
  • [46] Transformer-Based Parameter Estimation in Statistics
    Yin, Xiaoxin
    Yin, David S.
    MATHEMATICS, 2024, 12 (07)
  • [47] Transformer-Based Maneuvering Target Tracking
    Zhao, Guanghui
    Wang, Zelin
    Huang, Yixiong
    Zhang, Huirong
    Ma, Xiaojing
    SENSORS, 2022, 22 (21)
  • [48] Ferrofluid transformer-based tilt sensor
    DeGraff, Allison
    Rashidi, Reza
    MICROSYSTEM TECHNOLOGIES-MICRO-AND NANOSYSTEMS-INFORMATION STORAGE AND PROCESSING SYSTEMS, 2020, 26 (08): : 2499 - 2506
  • [49] Transformer-based Text Detection in the Wild
    Raisi, Zobeir
    Naiel, Mohamed A.
    Younes, Georges
    Wardell, Steven
    Zelek, John S.
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3156 - 3165
  • [50] Transformer-Based Point Cloud Classification
    Wu, Xianfeng
    Liu, Xinyi
    Wang, Junfei
    Wu, Xianzu
    Lai, Zhongyuan
    Zhou, Jing
    Liu, Xia
    ARTIFICIAL INTELLIGENCE AND ROBOTICS, ISAIR 2022, PT I, 2022, 1700 : 218 - 225