An Empirical Investigation of Word Class-Based Features for Natural Language Understanding

被引:4
|
作者
Celikyilmaz, Asli [1 ]
Sarikaya, Ruhi [1 ]
Jeong, Minwoo [1 ]
Deoras, Anoop [2 ,3 ]
机构
[1] Microsoft Corp, Intent Sci Team, Redmond, WA 98052 USA
[2] Microsoft Corp, Conversat Understanding Sci, Redmond, WA 98052 USA
[3] Netflix, Algorithms Res & Engn Grp, Los Gatos, CA 95032 USA
关键词
Class-based features; conditional random fields; exponential models; natural language understanding; regularization; shrinkage features; NETWORKS;
D O I
10.1109/TASLP.2015.2511925
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
There are many studies that show using class-based features improves the performance of natural language processing (NLP) tasks such as syntactic part-of-speech tagging, dependency parsing, sentiment analysis, and slot filling in natural language understanding (NLU), but not much has been reported on the underlying reasons for the performance improvements. In this paper, we investigate the effects of the word class-based features for the exponential family of models specifically focusing on NLU tasks, and demonstrate that the performance improvements could be attributed to the regularization effect of the class-based features on the underlying model. Our hypothesis is based on empirical observation that shrinking the sum of parameter magnitudes in an exponential model tends to improve performance. We show on several semantic tagging tasks that there is a positive correlation between the model size reduction by the addition of the class-based features and the model performance on a held-out dataset. We also demonstrate that class-based features extracted from different data sources using alternate word clustering methods can individually contribute to the performance gain. Since the proposed features are generated in an unsupervised manner without significant computational overhead, the improvements in performance largely come for free and we show that such features provide gains for a wide range of tasks from semantic classification and slot tagging in NLU to named entity recognition (NER).
引用
收藏
页码:994 / 1005
页数:12
相关论文
共 50 条
  • [1] RNN language model with word clustering and class-based output layer
    Yongzhe Shi
    Wei-Qiang Zhang
    Jia Liu
    Michael T Johnson
    [J]. EURASIP Journal on Audio, Speech, and Music Processing, 2013
  • [2] RNN language model with word clustering and class-based output layer
    Shi, Yongzhe
    Zhang, Wei-Qiang
    Liu, Jia
    Johnson, Michael T.
    [J]. EURASIP JOURNAL ON AUDIO SPEECH AND MUSIC PROCESSING, 2013,
  • [3] A class-based approach to word alignment
    Ker, SJ
    Chang, JS
    [J]. COMPUTATIONAL LINGUISTICS, 1997, 23 (02) : 313 - 343
  • [4] Enhancement of the word2vec Class-Based Language Modeling by Optimizing the Features Vector using PCA
    Abdulhameed, Tiba Zaki
    Zitouni, Imed
    Abdel-Qader, Ikhlas
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ELECTRO/INFORMATION TECHNOLOGY (EIT), 2018, : 866 - 870
  • [5] BAYESIAN CLASS-BASED LANGUAGE MODELS
    Su, Yi
    [J]. 2011 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2011, : 5564 - 5567
  • [6] A class-based logic language for ontologies
    Benslimane, D
    Hacid, MS
    Terzi, E
    Toumani, F
    [J]. FLEXIBLE QUERY ANSWERING SYSTEMS, PROCEEDINGS, 2002, 2522 : 56 - 70
  • [7] EXPLOITING DIFFERENT WORD CLUSTERINGS FOR CLASS-BASED RNN LANGUAGE MODELING IN SPEECH RECOGNITION
    Song, Minguang
    Zhao, Yunxin
    Wang, Shaojun
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 5735 - 5739
  • [8] Dynamically extensible objects in a class-based language
    Schmidt, RW
    [J]. TOOLS 23 - TECHNOLOGY OF OBJECT-ORIENTED LANGUAGES AND SYSTEMS, PROCEEDINGS, 1998, : 294 - 305
  • [9] A Spectral Algorithm for Learning Class-Based n-gram Models of Natural Language
    Stratos, Karl
    Kim, Do-kyum
    Collins, Michael
    Hsu, Daniel
    [J]. UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2014, : 762 - 771
  • [10] Generic class-based statistical language models for robust speech understanding in directed dialog applications
    Hebert, Matthieu
    [J]. INTERSPEECH 2007: 8TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION, VOLS 1-4, 2007, : 73 - 76