Performance Improvement on Traditional Chinese Task-Oriented Dialogue Systems With Reinforcement Learning and Regularized Dropout Technique

被引:3
|
作者
Sheu, Jeng-Shin [1 ]
Wu, Siang-Ru [1 ]
Wu, Wen-Hung [2 ]
机构
[1] Natl Yunlin Univ Sci & Technol, Dept Comp Sci & Informat Engn, Yunlin 640002, Taiwan
[2] Ponddy Educ Taiwan Ltd, New Taipei 231, Taiwan
关键词
Task analysis; Reinforcement learning; Computational modeling; Artificial intelligence; Tokenization; Data models; NLP; regularized dropout; reinforcement learning; task-oriented dialogue;
D O I
10.1109/ACCESS.2023.3248796
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The development of conversational voice assistant applications has been in full swing around the world. This paper aims to develop traditional Chinese multi-domain task-oriented dialogue (TOD) systems. It is typically implemented using pipeline approach, where submodules are optimized independently, resulting in inconsistencies with each other. Instead, this paper implements end-to-end multi-domain TOD models using pre-trained deep neural networks (DNNs). This allows us to integrate all the submodules into one single DNN model to solve the inconsistencies. Data shortages are common in conversational natural language processing (NLP) tasks using DNN models. In this regard, dropout regularization has been widely used to improve overfitting caused by insufficient training dataset. However, the randomness it introduces leads to non-negligible discrepancies between training and inference. On the other hand, pre-trained language models have successfully provided effective regularization for NLP tasks. An inherent disadvantage is that fine-tuning the pre-trained language model suffers from exposure bias and loss-evaluation mismatch. To this end, we propose a reinforcement learning (RL) approach to address both issues. Furthermore, we adopt a method called regularized dropout (R-Drop) to improve the inconsistency in dropout layers of DNNs. Experimental results show that both our proposed RL approach and the R-Drop technique can significantly improve the joint target accuracy (JGA) score and combined score of traditional Chinese TOD system in tasks of dialogue state tracking (DST) and end-to-end sentence prediction, respectively.
引用
收藏
页码:19849 / 19862
页数:14
相关论文
共 50 条
  • [31] A Task-oriented Chatbot Based on LSTM and Reinforcement Learning
    Chou, Tai-Liang
    Hsueh, Yu-Ling
    NLPIR 2019: 2019 3RD INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING AND INFORMATION RETRIEVAL, 2019, : 87 - 91
  • [32] Memory-to-Sequence learning with LSTM joint decoding for task-oriented dialogue systems
    Yu, Bing
    Ren, Fuji
    Bao, Yanwei
    PROCEEDINGS OF THE 2019 14TH IEEE CONFERENCE ON INDUSTRIAL ELECTRONICS AND APPLICATIONS (ICIEA 2019), 2019, : 200 - 204
  • [33] High-Quality Diversification for Task-Oriented Dialogue Systems
    Tang, Zhiwen
    Kulkarni, Hrishikesh
    Yang, Grace Hui
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 1861 - 1872
  • [34] Simulating User Satisfaction for the Evaluation of Task-oriented Dialogue Systems
    Sun, Weiwei
    Zhang, Shuo
    Balog, Krisztian
    Ren, Zhaochun
    Ren, Pengjie
    Chen, Zhumin
    de Rijke, Maarten
    SIGIR '21 - PROCEEDINGS OF THE 44TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2021, : 2499 - 2506
  • [35] Metaphorical User Simulators for Evaluating Task-oriented Dialogue Systems
    Sun, Weiwei
    Guo, Shuyu
    Zhang, Shuo
    Ren, Pengjie
    Chen, Zhumin
    de Rijke, Maarten
    Ren, Zhaochun
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2024, 42 (01)
  • [36] Training Neural Response Selection for Task-Oriented Dialogue Systems
    Henderson, Matthew
    Vulic, Ivan
    Gerz, Daniela
    Casanueva, Inigo
    Budzianowski, Pawel
    Coope, Sam
    Spithourakis, Georgios
    Wen, Tsung-Hsien
    Mrksic, Nikola
    Su, Pei-Hao
    57TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2019), 2019, : 5392 - 5404
  • [37] Multi-task Learning for Natural Language Generation in Task-Oriented Dialogue
    Zhu, Chenguang
    Zeng, Michael
    Huang, Xuedong
    2019 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING AND THE 9TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (EMNLP-IJCNLP 2019): PROCEEDINGS OF THE CONFERENCE, 2019, : 1261 - 1266
  • [38] Deep Reinforcement Learning Based Task-Oriented Communication in Multi-Agent Systems
    He, Guojun
    Feng, Mingjie
    Zhang, Yu
    Liu, Guanghua
    Dai, Yueyue
    Jiang, Tao
    IEEE WIRELESS COMMUNICATIONS, 2023, 30 (03) : 112 - 119
  • [39] Task-oriented reinforcement learning for continuous tasks in dynamic environment
    Kamal, MAS
    Murata, J
    Hirasawa, K
    SICE 2002: PROCEEDINGS OF THE 41ST SICE ANNUAL CONFERENCE, VOLS 1-5, 2002, : 829 - 832
  • [40] Multi-task learning with graph attention networks for multi-domain task-oriented dialogue systems
    Zhao, Meng
    Wang, Lifang
    Jiang, Zejun
    Li, Ronghan
    Lu, Xinyu
    Hu, Zhongtian
    KNOWLEDGE-BASED SYSTEMS, 2023, 259