Personalized Showcases: Generating Multi-Modal Explanations for Recommendations

被引:4
|
作者
Yan, An [1 ]
He, Zhankui [1 ]
Li, Jiacheng [1 ]
Zhang, Tianyang [1 ]
McAuley, Julian [1 ]
机构
[1] Univ Calif San Diego, La Jolla, CA USA
关键词
Datasets; Text Generation; Multi-Modality; Contrastive Learning;
D O I
10.1145/3539618.3592036
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Existing explanation models generate only text for recommendations but still struggle to produce diverse contents. In this paper, to further enrich explanations, we propose a new task named personalized showcases, in which we provide both textual and visual information to explain our recommendations. Specifically, we first select a personalized image set that is the most relevant to a user's interest toward a recommended item. Then, natural language explanations are generated accordingly given our selected images. For this new task, we collect a large-scale dataset from Google Maps and construct a high-quality subset for generating multi-modal explanations. We propose a personalized multi-modal framework which can generate diverse and visually-aligned explanations via contrastive learning. Experiments show that our framework benefits from different modalities as inputs, and is able to produce more diverse and expressive explanations compared to previous methods on a variety of evaluation metrics. (1)
引用
收藏
页码:2251 / 2255
页数:5
相关论文
共 50 条
  • [31] Multi-modal Perception
    Kondo, T.
    Denshi Joho Tsushin Gakkai Shi/Journal of the Institute of Electronics, Information and Communications Engineers, 78 (12):
  • [32] Multi-modal perception
    BT Technol J, 1 (35-46):
  • [33] Multi-modal mapping
    Yates, Darran
    NATURE REVIEWS NEUROSCIENCE, 2016, 17 (09) : 536 - 536
  • [34] Multi-modal Fusion
    Liu, Huaping
    Hussain, Amir
    Wang, Shuliang
    INFORMATION SCIENCES, 2018, 432 : 462 - 462
  • [35] Generating Physically Realistic and Directable Human Motions from Multi-modal Inputs
    Shrestha, Aayam
    Liu, Pan
    Ros, German
    Yuan, Kai
    Fern, Alan
    COMPUTER VISION - ECCV 2024, PT LXII, 2025, 15120 : 1 - 17
  • [36] Multi-modal perception
    Hollier, MP
    Rimell, AN
    Hands, DS
    Voelcker, RM
    BT TECHNOLOGY JOURNAL, 1999, 17 (01) : 35 - 46
  • [37] An imitation learning framework for generating multi-modal trajectories from unstructured demonstrations
    Peng, Jian-Wei
    Hu, Min-Chun
    Chu, Wei-Ta
    NEUROCOMPUTING, 2022, 500 : 712 - 723
  • [38] Multi-modal mapping
    Darran Yates
    Nature Reviews Neuroscience, 2016, 17 : 536 - 536
  • [39] Hadamard matrix-guided multi-modal hashing for multi-modal retrieval
    Yu, Jun
    Huang, Wei
    Li, Zuhe
    Shu, Zhenqiu
    Zhu, Liang
    DIGITAL SIGNAL PROCESSING, 2022, 130
  • [40] Multi-Modal Knowledge Graph Transformer Framework for Multi-Modal Entity Alignment
    Li, Qian
    Ji, Cheng
    Guo, Shu
    Liang, Zhaoji
    Wang, Lihong
    Li, Jianxin
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS - EMNLP 2023, 2023, : 987 - 999