Robust-to-Noise Models in Natural Language Processing Tasks

被引:0
|
作者
Malykh, Valentin [1 ]
机构
[1] Steklov Math Inst St Petersburg, Samsung PDMI Joint Ctr, Moscow Inst Phys & Technol, Neural Syst & Deep Learning Lab, St Petersburg, Russia
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
There are a lot of noisy texts surrounding a person in modern life. A traditional approach is to use spelling correction, yet the existing solutions are far from perfect. We propose a robust to noise word embeddings model which outperforms existing commonly used models like fasttext and word2vec in different tasks. In addition, we investigate the noise robustness of current models in different natural language processing tasks. We propose extensions for modern models in three downstream tasks, i.e. text classification, named entity recognition and aspect extraction, these extensions show improvement in noise robustness over existing solutions.
引用
收藏
页码:10 / 16
页数:7
相关论文
共 50 条
  • [1] Unsupervised multi-sense language models for natural language processing tasks
    Roh, Jihyeon
    Park, Sungjin
    Kim, Bo-Kyeong
    Oh, Sang-Hoon
    Lee, Soo-Young
    [J]. NEURAL NETWORKS, 2021, 142 : 397 - 409
  • [2] Solving Hungarian natural language processing tasks with multilingual generative models
    Yang, Zijian Gyozo
    Laki, Laszlo Janos
    [J]. ANNALES MATHEMATICAE ET INFORMATICAE, 2023, 57 : 92 - 106
  • [3] Solving Hungarian natural language processing tasks with multilingual generative models
    Yang, Zijian Gyozo
    Laki, Laszlo Janos
    [J]. ANNALES MATHEMATICAE ET INFORMATICAE, 2023, 57 : 92 - 106
  • [4] Robust Speech and Natural Language Processing Models for Depression Screening
    Lu, Y.
    Harati, A.
    Rutowski, T.
    Oliveira, R.
    Chlebek, P.
    Shriberg, E.
    [J]. 2020 IEEE SIGNAL PROCESSING IN MEDICINE AND BIOLOGY SYMPOSIUM, 2020,
  • [5] ScispaCy: Fast and Robust Models for Biomedical Natural Language Processing
    Neumann, Mark
    King, Daniel
    Beltagy, Iz
    Ammar, Waleed
    [J]. SIGBIOMED WORKSHOP ON BIOMEDICAL NATURAL LANGUAGE PROCESSING (BIONLP 2019), 2019, : 319 - 327
  • [6] Robust Natural Language Dialogues for Instruction Tasks
    Scheutz, Matthias
    [J]. UNMANNED SYSTEMS TECHNOLOGY XII, 2010, 7692
  • [7] Analysis of sentence embedding models using prediction tasks in natural language processing
    Adi, Y.
    Kermany, E.
    Belinkov, Y.
    Lavi, O.
    Goldberg, Y.
    [J]. IBM JOURNAL OF RESEARCH AND DEVELOPMENT, 2017, 61 (4-5)
  • [8] Empirical learning of natural language processing tasks
    Daelemans, W
    van den Bosch, A
    Weijters, T
    [J]. MACHINE LEARNING : ECML-97, 1997, 1224 : 337 - 344
  • [9] Quantifying Uncertainties in Natural Language Processing Tasks
    Xiao, Yijun
    Wang, William Yang
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 7322 - 7329
  • [10] On the Role of Relevance in Natural Language Processing Tasks
    Sauchuk, Artsiom
    Thorne, James
    Halevy, Alon
    Tonellotto, Nicola
    Silvestri, Fabrizio
    [J]. PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 1785 - 1789