Character-Level Quantum Mechanical Approach for a Neural Language Model

被引:0
|
作者
Zhihao Wang
Min Ren
Xiaoyan Tian
Xia Liang
机构
[1] Shandong University of Finance and Economics,School of Management Science and Engineering
[2] Shandong University of Finance and Economics,School of Mathematics and Quantitative Economics
[3] Shandong Police College,undefined
关键词
Character-level; Quantum theory; Network-in-network; Language model;
D O I
暂无
中图分类号
学科分类号
摘要
This article proposes a character-level neural language model (NLM) that is based on quantum theory. The input of the model is the character-level coding represented by the quantum semantic space model. Our model integrates a convolutional neural network (CNN) that is based on network-in-network (NIN). We assessed the effectiveness of our model through extensive experiments based on the English-language Penn Treebank dataset. The experiments results confirm that the quantum semantic inputs work well for the language models. For example, the PPL of our model is 10%–30% less than the states of the arts, while it keeps the relatively smaller number of parameters (i.e., 6 m).
引用
收藏
页码:1613 / 1621
页数:8
相关论文
共 50 条
  • [41] Deep learning based Character-level approach for Morphological Inflection Generation
    Prasad, Vidya K.
    Premjith, B.
    Chandran, Chandni, V
    Soman, K. P.
    Poornachandran, Prabaharan
    [J]. PROCEEDINGS OF THE 2019 INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING AND CONTROL SYSTEMS (ICCS), 2019, : 1423 - 1427
  • [42] Character-Level Chinese Dependency Parsing
    Zhang, Meishan
    Zhang, Yue
    Che, Wanxiang
    Liu, Ting
    [J]. PROCEEDINGS OF THE 52ND ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, VOL 1, 2014, : 1326 - 1336
  • [43] Applying the Transformer to Character-level Transduction
    Wu, Shijie
    Cotterell, Ryan
    Hulden, Mans
    [J]. 16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1901 - 1907
  • [44] Quasi Character-Level Transformers to Improve Neural Machine Translation on Small Datasets
    Carrion, Salvador
    Casacuberta, Francisco
    [J]. 2021 EIGHTH INTERNATIONAL CONFERENCE ON SOCIAL NETWORK ANALYSIS, MANAGEMENT AND SECURITY (SNAMS), 2021, : 184 - 189
  • [45] Character-level Adversarial Examples in Arabic
    Alshemali, Basemah
    Kalita, Jugal
    [J]. 20TH IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS (ICMLA 2021), 2021, : 9 - 14
  • [46] Character-Level Attention Convolutional Neural Networks for Short-Text Classification
    Yin, Feiyang
    Yao, Zhilin
    Liu, Jia
    [J]. HUMAN CENTERED COMPUTING, 2019, 11956 : 560 - 567
  • [47] Character-level Adversarial Samples Generation Approach for Chinese Text Classification
    Zhang, Shunxiang
    Wu, Houyue
    Zhu, Guangli
    Xu, Xin
    Su, Mingxing
    [J]. JOURNAL OF ELECTRONICS & INFORMATION TECHNOLOGY, 2023, 45 (06) : 2226 - 2235
  • [48] Character-level recurrent neural networks in practice: comparing training and sampling schemes
    De Boom, Cedric
    Demeester, Thomas
    Dhoedt, Bart
    [J]. NEURAL COMPUTING & APPLICATIONS, 2019, 31 (08): : 4001 - 4017
  • [49] Enhanced character-level deep convolutional neural networks for cardiovascular disease prediction
    Zhang, Zhichang
    Qiu, Yanlong
    Yang, Xiaoli
    Zhang, Minyu
    [J]. BMC MEDICAL INFORMATICS AND DECISION MAKING, 2020, 20 (Suppl 3)
  • [50] Weakly-supervised character-level convolutional neural networks for text classification
    Liu, Yongsheng
    Chen, Wenyu
    Niyongabo, Rubungo Andre
    Qu, Hong
    Miao, Kebin
    Wei, Feng
    [J]. DEVELOPMENTS OF ARTIFICIAL INTELLIGENCE TECHNOLOGIES IN COMPUTATION AND ROBOTICS, 2020, 12 : 701 - 708