MLP emulation of N-gram models as a first step to connectionist language modeling

被引:0
|
作者
Castro, MJ [1 ]
Prat, F [1 ]
Casacuberta, F [1 ]
机构
[1] Univ Politecn Valencia, Dept Sistemes Informat & Computacio, Valencia, Spain
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In problems such as automatic speech recognition and machine translation, where the system response must be a sentence in a given language, language models are employed in order to improve system performance. These language models are usually N-gram models (for instance, bigram or trigram models) which are estimated from large text databases using the occurrence frequencies of these N-grams. In 1989, Nakamura and Shikano empirically showed how multilayer perceptrons can emulate trigram model predictive capabilities with additional generalization features. Our paper discusses Nakamura and Shikano's work, provides new empirical evidence on multilayer perceptron capability to emulate N-gram models, and proposes new directions for extending neural network-based language models. The experimental work we present here compares connectionist phonological bigram models with a conventional one using different measures, which include recognition performances in a Spanish acoustic-phonetic decoding task.
引用
收藏
页码:910 / 915
页数:6
相关论文
共 50 条
  • [41] Semantic N-Gram Topic Modeling
    Kherwa, Pooja
    Bansal, Poonam
    EAI ENDORSED TRANSACTIONS ON SCALABLE INFORMATION SYSTEMS, 2020, 7 (26) : 1 - 12
  • [42] n-gram Effect in Malware Detection Using Multilayer Perceptron (MLP)
    Purnama, Benni
    Stiawan, Deris
    Hanapi, Darmawijoyo
    Winanto, Eko Arip
    Budiarto, Rahmat
    Bin Idris, Mohd Yazid
    2021 8TH INTERNATIONAL CONFERENCE ON ELECTRICAL ENGINEERING, COMPUTERSCIENCE AND INFORMATICS (EECSI) 2021, 2021, : 45 - 49
  • [43] A New Estimate of the n-gram Language Model
    Aouragh, Si Lhoussain
    Yousfi, Abdellah
    Laaroussi, Saida
    Gueddah, Hicham
    Nejja, Mohammed
    AI IN COMPUTATIONAL LINGUISTICS, 2021, 189 : 211 - 215
  • [44] A variant of n-gram based language classification
    Tomovic, Andrija
    Janicic, Predrag
    AI(ASTERISK)IA 2007: ARTIFICIAL INTELLIGENCE AND HUMAN-ORIENTED COMPUTING, 2007, 4733 : 410 - +
  • [45] Pruning Sparse Non-negative Matrix N-gram Language Models
    Pelemans, Joris
    Shazeer, Noam
    Chelba, Ciprian
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 1433 - 1437
  • [46] A Distributed System for Large-scale n-gram Language Models at Tencent
    Long, Qiang
    Wang, Wei
    Deng, Jinfu
    Liu, Song
    Huang, Wenhao
    Chen, Fangying
    Liu, Sifan
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2019, 12 (12): : 2206 - 2217
  • [47] Development of the N-gram Model for Azerbaijani Language
    Bannayeva, Aliya
    Aslanov, Mustafa
    2020 IEEE 14TH INTERNATIONAL CONFERENCE ON APPLICATION OF INFORMATION AND COMMUNICATION TECHNOLOGIES (AICT2020), 2020,
  • [48] Use of statistical N-gram models in natural language generation for machine translation
    Liu, FH
    Gu, L
    Gao, YQ
    Picheny, M
    2003 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, VOL I, PROCEEDINGS: SPEECH PROCESSING I, 2003, : 636 - 639
  • [49] URL-Based Web Page Classification: With n-Gram Language Models
    Abdallah, Tarek Amr
    de La Iglesia, Beatriz
    KNOWLEDGE DISCOVERY, KNOWLEDGE ENGINEERING AND KNOWLEDGE MANAGEMENT, IC3K 2014, 2015, 553 : 19 - 33
  • [50] Improving N-gram Language Modeling for Code-switching Speech Recognition
    Zeng, Zhiping
    Xu, Haihua
    Chong, Tze Yuang
    Chng, Eng-Siong
    Li, Haizhou
    2017 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC 2017), 2017, : 1546 - 1551