Deep Reinforcement Learning for On-line Dialogue State Tracking

被引:2
|
作者
Chen, Zhi [1 ]
Chen, Lu [1 ]
Zhou, Xiang [1 ]
Yu, Kai [1 ]
机构
[1] Shanghai Jiao Tong Univ, AI Inst, MoE Key Lab Artificial Intelligence, X LANCE Lab,Dept Comp Sci & Engn, Shanghai, Peoples R China
关键词
Task-oriented Dialogue System; Joint Training; Reinforcement Learning;
D O I
10.1007/978-981-99-2401-1_25
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Dialogue state tracking (DST) is a crucial module in dialogue management. It is usually cast as a supervised training problem, which is not convenient for on-line optimization. In this paper, a novel companion teaching based deep reinforcement learning (DRL) framework for on-line DST optimization is proposed. To the best of our knowledge, this is the first effort to optimize the DST module within DRL framework for on-line task-oriented spoken dialogue systems. In addition, dialogue policy can be further jointly updated. Experiments show that on-line DST optimization can effectively improve the dialogue manager performance while keeping the flexibility of using predefined policy. Joint training of both DST and policy can further improve the performance.
引用
收藏
页码:278 / 292
页数:15
相关论文
共 50 条
  • [31] Deep Reinforcement Learning with Iterative Shift for Visual Tracking
    Ren, Liangliang
    Yuan, Xin
    Lu, Jiwen
    Yang, Ming
    Zhou, Jie
    COMPUTER VISION - ECCV 2018, PT IX, 2018, 11213 : 697 - 713
  • [32] Visual Tracking via Hierarchical Deep Reinforcement Learning
    Zhang, Dawei
    Zheng, Zhonglong
    Jia, Riheng
    Li, Minglu
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 3315 - 3323
  • [33] On the Use of Deep Reinforcement Learning for Visual Tracking: A Survey
    Cruciata, Giorgio
    Lo Presti, Liliana
    La Cascia, Marco
    IEEE ACCESS, 2021, 9 : 120880 - 120900
  • [34] Exploring Deep Reinforcement Learning for Autonomous Powerline Tracking
    Pienroj, Panin
    Schonborn, Sandro
    Birke, Robert
    IEEE CONFERENCE ON COMPUTER COMMUNICATIONS WORKSHOPS (IEEE INFOCOM 2019 WKSHPS), 2019, : 496 - 501
  • [35] Deep Reinforcement Learning for Data Association in Cell Tracking
    Wang Junjie
    Su Xiaohong
    Zhao Lingling
    Zhang Jun
    FRONTIERS IN BIOENGINEERING AND BIOTECHNOLOGY, 2020, 8
  • [36] Drift-Proof Tracking With Deep Reinforcement Learning
    Chen, Zhongze
    Li, Jing
    Wu, Jia
    Chang, Jun
    Xiao, Yafu
    Wang, Xiaoting
    IEEE TRANSACTIONS ON MULTIMEDIA, 2022, 24 : 609 - 624
  • [37] Decision Controller for Object Tracking With Deep Reinforcement Learning
    Zhong, Zhao
    Yang, Zichen
    Feng, Weitao
    Wu, Wei
    Hu, Yangyang
    Liu, Cheng-Lin
    IEEE ACCESS, 2019, 7 : 28069 - 28079
  • [38] End-to-end on-line rescheduling from Gantt chart images using deep reinforcement learning
    Andres Palombarini, Jorge
    Carlos Martinez, Ernesto
    INTERNATIONAL JOURNAL OF PRODUCTION RESEARCH, 2022, 60 (14) : 4434 - 4463
  • [39] On-line simultaneous learning and tracking of visual feature graphs
    Declercq, Arnaud
    Piater, Justus H.
    2007 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-8, 2007, : 3320 - +
  • [40] POLICY ADAPTATION FOR DEEP REINFORCEMENT LEARNING-BASED DIALOGUE MANAGEMENT
    Chen, Lu
    Chang, Cheng
    Chen, Zhi
    Tan, Bowen
    Gasic, Milica
    Yu, Kai
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 6074 - 6078