ON THE USE OF N-GRAM TRANSDUCERS FOR DIALOGUE ANNOTATION

被引:2
|
作者
Tamarit, Vicent [1 ]
Martinez-Hinarejos, Carlos-D. [1 ]
Benedi, Jose-Miguel [1 ]
机构
[1] Univ Politecn Valencia, Inst Tecnol Informat, Valencia, Spain
来源
SPOKEN DIALOGUE SYSTEMS: TECHNOLOGY AND DESIGN | 2011年
关键词
Statistical models; Dialogue annotation;
D O I
10.1007/978-1-4419-7934-6_11
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
The implementation of dialogue systems is one of the most interesting applications of language technologies. Statistical models can be used in this implementation, allowing for a more flexible approach than when using rules defined by a human expert. However, statistical models require large amounts of dialogues annotated with dialogue-function labels (usually Dialogue Acts), and the annotation process is hard and time-consuming. Consequently, the use of other statistical models to obtain faster annotations is really interesting for the development of dialogue systems. In this work we compare two statistical models for dialogue annotation, a more classical Hidden Markov Model (HMM) based model and the new N-gram Transducers (NGT) model. This comparison is performed on two corpora of different nature, the well-known SwitchBoard corpus and the DIHANA corpus. The results show that the NGT model produces a much more accurate annotation that the HMM-based model (even 11% less error in the SwitchBoard corpus).
引用
收藏
页码:255 / 276
页数:22
相关论文
共 50 条
  • [21] Differentially Private n-gram Extraction
    Kim, Kunho
    Gopi, Sivakanth
    Kulkarni, Janardhan
    Yekhanin, Sergey
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [22] Text mining with n-gram variables
    Schonlau, Matthias
    Guenther, Nick
    Sucholutsky, Ilia
    STATA JOURNAL, 2017, 17 (04): : 866 - 881
  • [23] Uniquely decodable n-gram embeddings
    Kontorovich, L
    THEORETICAL COMPUTER SCIENCE, 2004, 329 (1-3) : 271 - 284
  • [24] Semantic N-Gram Topic Modeling
    Kherwa, Pooja
    Bansal, Poonam
    EAI ENDORSED TRANSACTIONS ON SCALABLE INFORMATION SYSTEMS, 2020, 7 (26) : 1 - 12
  • [25] N-gram Analysis of a Mongolian Text
    Altangerel, Khuder
    Tsend, Ganbat
    Jalsan, Khash-Erdene
    IFOST 2008: PROCEEDING OF THE THIRD INTERNATIONAL FORUM ON STRATEGIC TECHNOLOGIES, 2008, : 258 - 259
  • [26] On compressing n-gram language models
    Hirsimaki, Teemu
    2007 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL IV, PTS 1-3, 2007, : 949 - 952
  • [27] N-GRAM ANALYSIS IN THE ENGINEERING DOMAIN
    Leary, Martin
    Pearson, Geoff
    Burvill, Colin
    Mazur, Maciej
    Subic, Aleksandar
    PROCEEDINGS OF THE 18TH INTERNATIONAL CONFERENCE ON ENGINEERING DESIGN (ICED 11): IMPACTING SOCIETY THROUGH ENGINEERING DESIGN, VOL 6: DESIGN INFORMATION AND KNOWLEDGE, 2011, 6 : 414 - 423
  • [28] Supervised N-gram Topic Model
    Kawamae, Noriaki
    WSDM'14: PROCEEDINGS OF THE 7TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, 2014, : 473 - 482
  • [29] Discriminative n-gram language modeling
    Roark, Brian
    Saraclar, Murat
    Collins, Michael
    COMPUTER SPEECH AND LANGUAGE, 2007, 21 (02): : 373 - 392
  • [30] Similar N-gram Language Model
    Gillot, Christian
    Cerisara, Christophe
    Langlois, David
    Haton, Jean-Paul
    11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 3 AND 4, 2010, : 1824 - 1827