Performance Improvement on Traditional Chinese Task-Oriented Dialogue Systems With Reinforcement Learning and Regularized Dropout Technique

被引:3
|
作者
Sheu, Jeng-Shin [1 ]
Wu, Siang-Ru [1 ]
Wu, Wen-Hung [2 ]
机构
[1] Natl Yunlin Univ Sci & Technol, Dept Comp Sci & Informat Engn, Yunlin 640002, Taiwan
[2] Ponddy Educ Taiwan Ltd, New Taipei 231, Taiwan
关键词
Task analysis; Reinforcement learning; Computational modeling; Artificial intelligence; Tokenization; Data models; NLP; regularized dropout; reinforcement learning; task-oriented dialogue;
D O I
10.1109/ACCESS.2023.3248796
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The development of conversational voice assistant applications has been in full swing around the world. This paper aims to develop traditional Chinese multi-domain task-oriented dialogue (TOD) systems. It is typically implemented using pipeline approach, where submodules are optimized independently, resulting in inconsistencies with each other. Instead, this paper implements end-to-end multi-domain TOD models using pre-trained deep neural networks (DNNs). This allows us to integrate all the submodules into one single DNN model to solve the inconsistencies. Data shortages are common in conversational natural language processing (NLP) tasks using DNN models. In this regard, dropout regularization has been widely used to improve overfitting caused by insufficient training dataset. However, the randomness it introduces leads to non-negligible discrepancies between training and inference. On the other hand, pre-trained language models have successfully provided effective regularization for NLP tasks. An inherent disadvantage is that fine-tuning the pre-trained language model suffers from exposure bias and loss-evaluation mismatch. To this end, we propose a reinforcement learning (RL) approach to address both issues. Furthermore, we adopt a method called regularized dropout (R-Drop) to improve the inconsistency in dropout layers of DNNs. Experimental results show that both our proposed RL approach and the R-Drop technique can significantly improve the joint target accuracy (JGA) score and combined score of traditional Chinese TOD system in tasks of dialogue state tracking (DST) and end-to-end sentence prediction, respectively.
引用
收藏
页码:19849 / 19862
页数:14
相关论文
共 50 条
  • [21] Transfer Learning based Task-oriented Dialogue Policy for Multiple Domains using Hierarchical Reinforcement Learning
    Saha, Tulika
    Saha, Sriparna
    Bhattacharyya, Pushpak
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [22] A Review of the Research on Dialogue Management of Task-Oriented Systems
    Zhao, Yin Jiang
    Li, Yan Ling
    Lin, Min
    2019 3RD INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE, AUTOMATION AND CONTROL TECHNOLOGIES (AIACT 2019), 2019, 1267
  • [23] EasyDial: A tool for task-oriented dialogue systems on the telephone
    Moisa, L
    Pinton, C
    Popovici, C
    NINTH INTERNATIONAL WORKSHOP ON DATABASE AND EXPERT SYSTEMS APPLICATIONS, PROCEEDINGS, 1998, : 176 - 181
  • [24] Understanding User Satisfaction with Task-oriented Dialogue Systems
    Siro, Clemencia
    Aliannejadi, Mohammad
    de Rijke, Maarten
    PROCEEDINGS OF THE 45TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL (SIGIR '22), 2022, : 2018 - 2023
  • [25] Building Task-Oriented Dialogue Systems for Online Shopping
    Yan, Zhao
    Duan, Nan
    Chen, Peng
    Zhou, Ming
    Zhou, Jianshe
    Li, Zhoujun
    THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 4618 - 4625
  • [26] Memory-Augmented Dialogue Management for Task-Oriented Dialogue Systems
    Zhang, Zheng
    Huang, Minlie
    Zhao, Zhongzhou
    Ji, Feng
    Chen, Haiqing
    Zhu, Xiaoyan
    ACM TRANSACTIONS ON INFORMATION SYSTEMS, 2019, 37 (03)
  • [27] Personality-aware Natural Language Generation for Task-oriented Dialogue using Reinforcement Learning
    Guo, Ao
    Ohashi, Atsumoto
    Chiba, Yuya
    Tsunomori, Yuiko
    Hirai, Ryu
    Higashinaka, Ryuichiro
    2023 32ND IEEE INTERNATIONAL CONFERENCE ON ROBOT AND HUMAN INTERACTIVE COMMUNICATION, RO-MAN, 2023, : 1823 - 1828
  • [28] Improvement of performance in small task-oriented groups
    Hoyos, CG
    ZEITSCHRIFT FUR ARBEITS-UND ORGANISATIONSPSYCHOLOGIE, 2003, 47 (01): : 52 - 54
  • [29] Task-Oriented Reinforcement Learning with Interest State Representation
    Li, Ziyi
    Hu, Xiangtao
    Zhang, Yongle
    Zhou, Fujie
    2024 INTERNATIONAL CONFERENCE ON ADVANCED ROBOTICS AND MECHATRONICS, ICARM 2024, 2024, : 721 - 728
  • [30] A Task-oriented Chatbot Based on LSTM and Reinforcement Learning
    Hsueh, Yu-Ling
    Chou, Tai-Liang
    ACM TRANSACTIONS ON ASIAN AND LOW-RESOURCE LANGUAGE INFORMATION PROCESSING, 2023, 22 (01)