Personalized Showcases: Generating Multi-Modal Explanations for Recommendations

被引:4
|
作者
Yan, An [1 ]
He, Zhankui [1 ]
Li, Jiacheng [1 ]
Zhang, Tianyang [1 ]
McAuley, Julian [1 ]
机构
[1] Univ Calif San Diego, La Jolla, CA USA
关键词
Datasets; Text Generation; Multi-Modality; Contrastive Learning;
D O I
10.1145/3539618.3592036
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Existing explanation models generate only text for recommendations but still struggle to produce diverse contents. In this paper, to further enrich explanations, we propose a new task named personalized showcases, in which we provide both textual and visual information to explain our recommendations. Specifically, we first select a personalized image set that is the most relevant to a user's interest toward a recommended item. Then, natural language explanations are generated accordingly given our selected images. For this new task, we collect a large-scale dataset from Google Maps and construct a high-quality subset for generating multi-modal explanations. We propose a personalized multi-modal framework which can generate diverse and visually-aligned explanations via contrastive learning. Experiments show that our framework benefits from different modalities as inputs, and is able to produce more diverse and expressive explanations compared to previous methods on a variety of evaluation metrics. (1)
引用
收藏
页码:2251 / 2255
页数:5
相关论文
共 50 条
  • [21] Multi-modal anchor adaptation learning for multi-modal summarization
    Chen, Zhongfeng
    Lu, Zhenyu
    Rong, Huan
    Zhao, Chuanjun
    Xu, Fan
    NEUROCOMPUTING, 2024, 570
  • [22] Explanation as a Process: User-Centric Construction of Multi-level and Multi-modal Explanations
    Finzel, Bettina
    Tafler, David E.
    Scheele, Stephan
    Schmid, Ute
    ADVANCES IN ARTIFICIAL INTELLIGENCE, KI 2021, 2021, 12873 : 80 - 94
  • [23] Phase Analysis of Bifilar Helix Antenna Generating Multi-modal OAM
    Zhang, Yunqi
    Yang, Rui
    Li, Xuping
    2022 IEEE 10TH ASIA-PACIFIC CONFERENCE ON ANTENNAS AND PROPAGATION, APCAP, 2022,
  • [24] A NOVEL METHOD FOR AUTOMATICALLY GENERATING MULTI-MODAL DIALOGUE FROM TEXT
    Prendinger, Helmut
    Piwek, Paul
    Ishizuka, Mitsuru
    INTERNATIONAL JOURNAL OF SEMANTIC COMPUTING, 2007, 1 (03) : 319 - 334
  • [25] Complementary Recommendations Using Deep Multi-modal Embeddings For Online Retail
    Ahsan, Unaiza
    Cui, Xiquan
    West, Rebecca
    Guo, Mingming
    Wu, San He
    Al Jadda, Khalifeh
    2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA (BIG DATA), 2020, : 1774 - 1779
  • [26] Flexible Dual Multi-Modal Hashing for Incomplete Multi-Modal Retrieval
    Wei, Yuhong
    An, Junfeng
    INTERNATIONAL JOURNAL OF IMAGE AND GRAPHICS, 2024,
  • [27] Multi-Modal 2020: Multi-Modal Argumentation 30 Years Later
    Gilbert, Michael A.
    INFORMAL LOGIC, 2022, 42 (03): : 487 - 506
  • [28] Towards Understandable Personalized Recommendations: Hybrid Explanations
    Svrcek, Martin
    Kompan, Michal
    Bielikova, Maria
    COMPUTER SCIENCE AND INFORMATION SYSTEMS, 2019, 16 (01) : 179 - 203
  • [29] Hydra: A Personalized and Context-Aware Multi-Modal Transportation Recommendation System
    Liu, Hao
    Tong, Yongxin
    Zhang, Panpan
    Lu, Xinjiang
    Duan, Jianguo
    Xiong, Hui
    KDD'19: PROCEEDINGS OF THE 25TH ACM SIGKDD INTERNATIONAL CONFERENCCE ON KNOWLEDGE DISCOVERY AND DATA MINING, 2019, : 2314 - 2324
  • [30] Aggregation Design for Personalized Federated Multi-Modal Learning Over Wireless Networks
    Yin, Benshun
    Chen, Zhiyong
    Tao, Meixia
    IEEE COMMUNICATIONS LETTERS, 2024, 28 (08) : 1850 - 1854