A Vietnamese Language Model Based on Recurrent Neural Network

被引:0
|
作者
Viet-Trung Tran [1 ]
Kiem-Hieu Nguyen [1 ]
Duc-Hanh Bui [1 ]
机构
[1] Hanoi Univ Sci & Technol, Hanoi, Vietnam
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Language modeling plays a critical role in many natural language processing (NLP) tasks such as text prediction, machine translation and speech recognition. Traditional statistical language models (e.g. n-gram models) can only offer words that have been seen before and can not capture long word context. Neural language model provides a promising solution to surpass this shortcoming of statistical language model. This paper investigates Recurrent Neural Networks (RNNs) language model for Vietnamese, at character and syllable-levels. Experiments were conducted on a large dataset of 24M syllables, constructed from 1,500 movie subtitles. The experimental results show that our RNN-based language models yield reasonable performance on the movie subtitle dataset. Concretely, our models outperform n-gram language models in term of perplexity score.
引用
收藏
页码:274 / 278
页数:5
相关论文
共 50 条
  • [1] Recurrent neural network based language model
    Mikolov, Tomas
    Karafiat, Martin
    Burget, Lukas
    Cernocky, Jan Honza
    Khudanpur, Sanjeev
    11TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION 2010 (INTERSPEECH 2010), VOLS 1-2, 2010, : 1045 - 1048
  • [2] Multiscale recurrent neural network based language model
    Morioka, Tsuyoshi
    Iwata, Tomoharu
    Hori, Takaaki
    Kobayashi, Tetsunori
    16TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2015), VOLS 1-5, 2015, : 2366 - 2370
  • [3] The Prediction of Character Based on Recurrent Neural Network Language Model
    Shi, Zejian
    Shi, Minyong
    Li, Chunfang
    2017 16TH IEEE/ACIS INTERNATIONAL CONFERENCE ON COMPUTER AND INFORMATION SCIENCE (ICIS 2017), 2017, : 613 - 616
  • [4] FPGA Acceleration of Recurrent Neural Network based Language Model
    Li, Sicheng
    Wu, Chunpeng
    Li, Hai
    Li, Boxun
    Wang, Yu
    Qiu, Qinru
    2015 IEEE 23RD ANNUAL INTERNATIONAL SYMPOSIUM ON FIELD-PROGRAMMABLE CUSTOM COMPUTING MACHINES (FCCM), 2015, : 111 - 118
  • [5] Recurrent Neural Network Based Language Model Personalization by Social Network Crowdsourcing
    Wen, Tsung-Hsien
    Heidel, Aaron
    Lee, Hung-yi
    Tsao, Yu
    Lee, Lin -Shan
    14TH ANNUAL CONFERENCE OF THE INTERNATIONAL SPEECH COMMUNICATION ASSOCIATION (INTERSPEECH 2013), VOLS 1-5, 2013, : 2702 - 2706
  • [6] Personalizing Recurrent-Neural-Network-Based Language Model by Social Network
    Lee, Hung-Yi
    Tseng, Bo-Hsiang
    Wen, Tsung-Hsien
    Tsao, Yu
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2017, 25 (03) : 519 - 530
  • [7] MARKOV RECURRENT NEURAL NETWORK LANGUAGE MODEL
    Chien, Jen-Tzung
    Kuo, Che-Yu
    2019 IEEE AUTOMATIC SPEECH RECOGNITION AND UNDERSTANDING WORKSHOP (ASRU 2019), 2019, : 807 - 813
  • [8] BAYESIAN RECURRENT NEURAL NETWORK LANGUAGE MODEL
    Chien, Jen-Tzung
    Ku, Yuan-Chu
    2014 IEEE WORKSHOP ON SPOKEN LANGUAGE TECHNOLOGY SLT 2014, 2014, : 206 - 211
  • [9] EXTENSIONS OF RECURRENT NEURAL NETWORK LANGUAGE MODEL
    Mikolov, Tomas
    Kombrink, Stefan
    Burget, Lukas
    Cernocky, Jan Honza
    Khudanpur, Sanjeev
    2011 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2011, : 5528 - 5531
  • [10] An Improved Recurrent Neural Network Language Model for Programming Language
    Wu, Liwei
    Wu, Youhua
    Li, Fei
    Zheng, Tao
    2019 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2019,