The Art of Asking: Prompting Large Language Models for Serendipity Recommendations

被引:0
|
作者
Fu, Zhe [1 ]
Niu, Xi [1 ]
机构
[1] Univ North Carolina Charlotte, Coll Comp & Informat, Charlotte, NC 28223 USA
基金
美国国家科学基金会;
关键词
Serendipity; Large Language Models; Prompt Learning;
D O I
10.1145/3664190.3672521
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Serendipity means an unexpected but valuable discovery. Its elusive nature makes it susceptible to modeling. In this paper, we address the challenge of modeling serendipity in recommender systems using Large Language Models (LLMs), a recent breakthrough in AI technologies. We leveraged LLMs' prompting mechanisms to convert a problem of serendipity recommendations into a problem of formulating a prompt to elicit serendipity recommendations. The formulated prompt is called SerenPrompt. We designed three types of SerenPrompt: discrete with natural words, continuous with trainable tokens, and hybrid that combines the previous two types. In the meanwhile, for each type of SerenPrompt, we also designed two styles: direct and indirect, to investigate whether it is feasible to directly ask an LLM a question on whether an item is a serendipity, or we should breakdown the question into several sub-questions. Extensive experiments have demonstrated the effectiveness of SerenPrompt in generating serendipity recommendations, compared to the state-of-the-art models. The combination of the hybrid type and the indirect style achieves the best performance, with relatively low sacrifice to computational efficiency. The results demonstrate that natural words and virtual tokens, as building blocks of LLM prompts, each have their own advantages. The better performance of the indirect style speaks to the effectiveness of decomposing the direct question on serendipity.
引用
收藏
页码:157 / 166
页数:10
相关论文
共 50 条
  • [21] Chain-of-event prompting for multi-document summarization by large language models
    Bao, Songlin
    Li, Tiantian
    Cao, Bin
    INTERNATIONAL JOURNAL OF WEB INFORMATION SYSTEMS, 2024, 20 (03) : 229 - 247
  • [22] Prompting or Fine-tuning? A Comparative Study of Large Language Models for Taxonomy Construction
    Chen, Boqi
    Yi, Fandi
    Varro, Daniel
    2023 ACM/IEEE INTERNATIONAL CONFERENCE ON MODEL DRIVEN ENGINEERING LANGUAGES AND SYSTEMS COMPANION, MODELS-C, 2023, : 588 - 596
  • [23] Impact of Contradicting Subtle Emotion Cues on Large Language Models with Various Prompting Techniques
    Huda, Noor Ul
    Sahito, Sanam Fayaz
    Gilal, Abdul Rehman
    Abro, Ahsanullah
    Alshanqiti, Abdullah
    Alsughayyir, Aeshah
    Palli, Abdul Sattar
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2024, 15 (04) : 407 - 414
  • [24] Benchmarking large language models for biomedical natural language processing applications and recommendations
    Qingyu Chen
    Yan Hu
    Xueqing Peng
    Qianqian Xie
    Qiao Jin
    Aidan Gilson
    Maxwell B. Singer
    Xuguang Ai
    Po-Ting Lai
    Zhizheng Wang
    Vipina K. Keloth
    Kalpana Raja
    Jimin Huang
    Huan He
    Fongci Lin
    Jingcheng Du
    Rui Zhang
    W. Jim Zheng
    Ron A. Adelman
    Zhiyong Lu
    Hua Xu
    Nature Communications, 16 (1)
  • [25] Large Language Models for Intent-Driven Session Recommendations
    Sun, Zhu
    Liu, Hongyang
    Qu, Xinghua
    Feng, Kaidong
    Wang, Yan
    Ong, Yew Soon
    PROCEEDINGS OF THE 47TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2024, 2024, : 324 - 334
  • [26] InteraRec: Interactive Recommendations Using Multimodal Large Language Models
    Karra, Saketh Reddy
    Tulabandhula, Theja
    TRENDS AND APPLICATIONS IN KNOWLEDGE DISCOVERY AND DATA MINING, PAKDD 2024 WORKSHOPS, RAFDA AND IWTA, 2024, 14658 : 32 - 43
  • [27] Automatic Lesson Plan Generation via Large Language Models with Self-critique Prompting
    Zheng, Ying
    Li, Xueyi
    Huang, Yaying
    Liang, Qianru
    Guo, Teng
    Hou, Mingliang
    Gao, Boyu
    Tian, Mi
    Liu, Zitao
    Luo, Weiqi
    ARTIFICIAL INTELLIGENCE IN EDUCATION: POSTERS AND LATE BREAKING RESULTS, WORKSHOPS AND TUTORIALS, INDUSTRY AND INNOVATION TRACKS, PRACTITIONERS, DOCTORAL CONSORTIUM AND BLUE SKY, AIED 2024, PT I, 2024, 2150 : 163 - 178
  • [28] Investigating the Efficacy of Large Language Models in Reflective Assessment Methods through Chain of Thoughts Prompting
    Masikisiki, Baphumelele
    Marivate, Vukosi
    Hlophe, Yvette
    PROCEEDINGS OF THE 4TH AFRICAN CONFERENCE FOR HUMAN COMPUTER INTERACTION, AFRICHI 2023, 2023, : 44 - 49
  • [29] An active inference strategy for prompting reliable responses from large language models in medical practice
    Roma Shusterman
    Allison C. Waters
    Shannon O’Neill
    Marshall Bangs
    Phan Luu
    Don M. Tucker
    npj Digital Medicine, 8 (1)
  • [30] Prompting Large Language Models with Answer Heuristics for Knowledge-based Visual Question Answering
    Shao, Zhenwei
    Yu, Zhou
    Wang, Meng
    Yu, Jun
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 14974 - 14983