Open Domain Dialogue Generation with Latent Images

被引:0
|
作者
Yang, Ze [1 ]
Wu, Wei [2 ]
Hu, Huang [3 ]
Xu, Can [3 ]
Wang, Wei [4 ]
Li, Zhoujun [1 ]
机构
[1] Beihang Univ, State Key Lab Software Dev Environm, Beijing, Peoples R China
[2] Meituan, Beijing, Peoples R China
[3] Microsoft, Beijing, Peoples R China
[4] China Resources Grp, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider grounding open domain dialogues with images. Existing work assumes that both an image and a textual context are available, but image-grounded dialogues by nature are more difficult to obtain than textual dialogues. Thus, we propose learning a response generation model with both image-grounded dialogues and textual dialogues by assuming that the visual scene information at the time of a conversation can be represented by an image, and trying to recover the latent images of the textual dialogues through text-to-image generation techniques. The likelihood of the two types of dialogues is then formulated by a response generator and an image reconstructor that are learned within a conditional variational auto-encoding framework. Empirical studies are conducted in both image-grounded conversation and text-based conversation. In the first scenario, image-grounded dialogues, especially under a low-resource setting, can be effectively augmented by textual dialogues with latent images; while in the second scenario, latent images can enrich the content of responses and at the same time keep them relevant to contexts.
引用
下载
收藏
页码:14239 / 14247
页数:9
相关论文
共 50 条
  • [11] LEARNING TO SELECT CONTEXT IN A HIERARCHICAL AND GLOBAL PERSPECTIVE FOR OPEN-DOMAIN DIALOGUE GENERATION
    Shen, Lei
    Zhan, Haolan
    Shen, Xin
    Feng, Yang
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7438 - 7442
  • [12] Generative commonsense knowledge subgraph retrieval for open-domain dialogue response generation
    Wu, Sixing
    Yu, Jiong
    Chen, Jiahao
    Zhou, Wei
    NEURAL NETWORKS, 2024, 180
  • [13] SIDECONTROL: Controlled Open-domain Dialogue Generation via Additive Side Networks
    Du, Wanyu
    Ji, Yangfeng
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EMNLP 2021, 2021, : 2175 - 2194
  • [14] Sequential or jumping: context-adaptive response generation for open-domain dialogue systems
    Ling, Yanxiang
    Liang, Zheng
    Wang, Tianqi
    Cai, Fei
    Chen, Honghui
    APPLIED INTELLIGENCE, 2023, 53 (09) : 11251 - 11266
  • [15] Open-Domain Dialogue Generation Grounded with Dynamic Multi-form Knowledge Fusion
    Xu, Feifei
    Zhou, Shanlin
    Ma, Yunpu
    Wang, Xinpeng
    Zhang, Wenkai
    Li, Zhisong
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS, DASFAA 2022, PT III, 2022, : 101 - 116
  • [16] MixEI: Mixing explicit and implicit commonsense knowledge in open-domain dialogue response generation
    Wu, Sixing
    Yu, Jiong
    Zhou, Wei
    Neurocomputing, 2025, 618
  • [17] Sequential or jumping: context-adaptive response generation for open-domain dialogue systems
    Yanxiang Ling
    Zheng Liang
    Tianqi Wang
    Fei Cai
    Honghui Chen
    Applied Intelligence, 2023, 53 : 11251 - 11266
  • [18] Dialogue Distillation: Open-Domain Dialogue Augmentation Using Unpaired Data
    Zhang, Rongsheng
    Zheng, Yinhe
    Shao, Jianzhi
    Mao, Xiaoxi
    Xi, Yadong
    Huang, Minlie
    PROCEEDINGS OF THE 2020 CONFERENCE ON EMPIRICAL METHODS IN NATURAL LANGUAGE PROCESSING (EMNLP), 2020, : 3449 - 3460
  • [19] RADE: Reference-Assisted Dialogue Evaluation for Open-Domain Dialogue
    Shi, Zhengliang
    Sun, Weiwei
    Zhang, Shuo
    Zhang, Zhen
    Ren, Pengjie
    Ren, Zhaochun
    PROCEEDINGS OF THE 61ST ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2023): LONG PAPERS, VOL 1, 2023, : 12856 - 12875
  • [20] Improving Open-Domain Dialogue Response Generation with Multi-Source Multilingual Commonsense Knowledge
    Wu, Sixing
    Yu, Jiong
    Chen, Jiahao
    Deng, Xiaofan
    Zhou, Wei
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 17, 2024, : 19252 - 19260