Open Domain Dialogue Generation with Latent Images

被引:0
|
作者
Yang, Ze [1 ]
Wu, Wei [2 ]
Hu, Huang [3 ]
Xu, Can [3 ]
Wang, Wei [4 ]
Li, Zhoujun [1 ]
机构
[1] Beihang Univ, State Key Lab Software Dev Environm, Beijing, Peoples R China
[2] Meituan, Beijing, Peoples R China
[3] Microsoft, Beijing, Peoples R China
[4] China Resources Grp, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider grounding open domain dialogues with images. Existing work assumes that both an image and a textual context are available, but image-grounded dialogues by nature are more difficult to obtain than textual dialogues. Thus, we propose learning a response generation model with both image-grounded dialogues and textual dialogues by assuming that the visual scene information at the time of a conversation can be represented by an image, and trying to recover the latent images of the textual dialogues through text-to-image generation techniques. The likelihood of the two types of dialogues is then formulated by a response generator and an image reconstructor that are learned within a conditional variational auto-encoding framework. Empirical studies are conducted in both image-grounded conversation and text-based conversation. In the first scenario, image-grounded dialogues, especially under a low-resource setting, can be effectively augmented by textual dialogues with latent images; while in the second scenario, latent images can enrich the content of responses and at the same time keep them relevant to contexts.
引用
下载
收藏
页码:14239 / 14247
页数:9
相关论文
共 50 条
  • [1] Interpreting open-domain dialogue generation by disentangling latent feature representations
    Wang, Ye
    Liao, Jingbo
    Yu, Hong
    Wang, Guoyin
    Zhang, Xiaoxia
    Liu, Li
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (28): : 20855 - 20867
  • [2] Interpreting open-domain dialogue generation by disentangling latent feature representations
    Ye Wang
    Jingbo Liao
    Hong Yu
    Guoyin Wang
    Xiaoxia Zhang
    Li Liu
    Neural Computing and Applications, 2023, 35 : 20855 - 20867
  • [3] Enhancing the Open-Domain Dialogue Evaluation in Latent Space
    Chan, Zhangming
    Liu, Lemao
    Li, Juntao
    Zhang, Haisong
    Zhao, Dongyan
    Shi, Shuming
    Yan, Rui
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, ACL-IJCNLP 2021, 2021, : 4889 - 4900
  • [4] Adversarial Evaluation for Open-Domain Dialogue Generation
    Bruni, Elia
    Fernandez, Raquel
    18TH ANNUAL MEETING OF THE SPECIAL INTEREST GROUP ON DISCOURSE AND DIALOGUE (SIGDIAL 2017), 2017, : 284 - 288
  • [5] Towards Holistic and Automatic Evaluation of Open-Domain Dialogue Generation
    Pang, Bo
    Nijkamp, Erik
    Han, Wenjuan
    Zhou, Linqi
    Liu, Yixian
    Tu, Kewei
    58TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2020), 2020, : 3619 - 3629
  • [6] A Randomized Link Transformer for Diverse Open-Domain Dialogue Generation
    Lee, Jing Yang
    Lee, Kong Aik
    Gan, Woon Seng
    PROCEEDINGS OF THE 4TH WORKSHOP ON NLP FOR CONVERSATIONAL AI, 2022, : 1 - 11
  • [7] Leveraging Context for Neural Question Generation in Open-domain Dialogue Systems
    Ling, Yanxiang
    Cai, Fei
    Chen, Honghui
    de Rijke, Maarten
    WEB CONFERENCE 2020: PROCEEDINGS OF THE WORLD WIDE WEB CONFERENCE (WWW 2020), 2020, : 2486 - 2492
  • [8] Text is NOT Enough: Integrating Visual Impressions into Open-domain Dialogue Generation
    Shen, Lei
    Zhan, Haolan
    Shen, Xin
    Song, Yonghao
    Zhao, Xiaofang
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 4287 - 4296
  • [9] Establishing the Informational Requirements for Modelling Open Domain Dialogue and Prototyping a Retrieval Open Domain Dialogue System
    Meier, Trent
    Pimenidis, Elias
    COMPUTATIONAL COLLECTIVE INTELLIGENCE (ICCCI 2021), 2021, 12876 : 655 - 667
  • [10] Emotion-and-knowledge grounded response generation in an open-domain dialogue setting
    Varshney, Deeksha
    Ekbal, Asif
    Cambria, Erik
    KNOWLEDGE-BASED SYSTEMS, 2024, 284