Intrinsically motivated reinforcement learning based recommendation with counterfactual data augmentation

被引:0
|
作者
Xiaocong Chen
Siyu Wang
Lianyong Qi
Yong Li
Lina Yao
机构
[1] University of New South Wales,School of Computer Science and Engineering
[2] China University of Petroleum (East China),College of Computer Science and Technology
[3] Tsinghua University,Department of Electronic Engineering
[4] CSIRO,Data 61
来源
World Wide Web | 2023年 / 26卷
关键词
Recommender systems; Deep reinforcement learning; Counterfactual reasoning;
D O I
暂无
中图分类号
学科分类号
摘要
Deep reinforcement learning (DRL) has shown promising results in modeling dynamic user preferences in RS in recent literature. However, training a DRL agent in the sparse RS environment poses a significant challenge. This is because the agent must balance between exploring informative user-item interaction trajectories and using existing trajectories for policy learning, a known exploration and exploitation trade-off. This trade-off greatly affects the recommendation performance when the environment is sparse. In DRL-based RS, balancing exploration and exploitation is even more challenging as the agent needs to deeply explore informative trajectories and efficiently exploit them in the context of RS. To address this issue, we propose a novel intrinsically motivated reinforcement learning (IMRL) method that enhances the agent’s capability to explore informative interaction trajectories in the sparse environment. We further enrich these trajectories via an adaptive counterfactual augmentation strategy with a customised threshold to improve their efficiency in exploitation. Our approach is evaluated on six offline datasets and three online simulation platforms, demonstrating its superiority over existing state-of-the-art methods. The extensive experiments show that our IMRL method outperforms other methods in terms of recommendation performance in the sparse RS environment.
引用
收藏
页码:3253 / 3274
页数:21
相关论文
共 50 条
  • [21] Significance extraction based on data augmentation for reinforcement learning
    Han, Yuxi
    Li, Dequan
    Yang, Yang
    FRONTIERS OF INFORMATION TECHNOLOGY & ELECTRONIC ENGINEERING, 2025, : 385 - 399
  • [22] Learning-based counterfactual explanations for recommendation
    Jingxuan WEN
    Huafeng LIU
    Liping JING
    Jian YU
    Science China(Information Sciences), 2024, 67 (08) : 115 - 129
  • [23] Learning-based counterfactual explanations for recommendation
    Wen, Jingxuan
    Liu, Huafeng
    Jing, Liping
    Yu, Jian
    SCIENCE CHINA-INFORMATION SCIENCES, 2024, 67 (08)
  • [24] Data-Augmented Counterfactual Learning for Bundle Recommendation
    Zhu, Shixuan
    Shen, Qi
    Cui, Chuan
    Ji, Yu
    Zhang, Yiming
    Dong, Zhenwei
    Wei, Zhihua
    DATABASE SYSTEMS FOR ADVANCED APPLICATIONS. DASFAA 2023 INTERNATIONAL WORKSHOPS, BDMS 2023, BDQM 2023, GDMA 2023, BUNDLERS 2023, 2023, 13922 : 314 - 330
  • [25] Intrinsically Motivated Reinforcement Learning in Socio-Economic Systems: the Dynamical Analysis
    Zgonnikov, Arkady
    Lubashevsky, Ihor
    2013 IEEE THIRD JOINT INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING AND EPIGENETIC ROBOTICS (ICDL), 2013,
  • [26] Counterfactual Adversarial Learning for Recommendation
    Liu, Jialin
    Zhang, Zijian
    Zhao, Xiangyu
    Li, Jun
    PROCEEDINGS OF THE 32ND ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2023, 2023, : 4115 - 4119
  • [27] Data Augmented Sequential Recommendation Based on Counterfactual Thinking
    Chen, Xu
    Wang, Zhenlei
    Xu, Hongteng
    Zhang, Jingsen
    Zhang, Yongfeng
    Zhao, Wayne Xin
    Wen, Ji-Rong
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2023, 35 (09) : 9181 - 9194
  • [28] Autotelic Agents with Intrinsically Motivated Goal-Conditioned Reinforcement Learning: A Short Survey
    Colas C.
    Karch T.
    Sigaud O.
    Oudeyer P.-Y.
    Journal of Artificial Intelligence Research, 2022, 74 : 1159 - 1199
  • [29] Autotelic Agents with Intrinsically Motivated Goal-Conditioned Reinforcement Learning: A Short Survey
    Colas, Cedric
    Karch, Tristan
    Sigaud, Olivier
    Oudeyer, Pierre-Yves
    JOURNAL OF ARTIFICIAL INTELLIGENCE RESEARCH, 2022, 74 : 1159 - 1199
  • [30] Intrinsically motivated reinforcement learning for human-robot interaction in the real-world
    Qureshi, Ahmed Hussain
    Nakamura, Yutaka
    Yoshikawa, Yuichiro
    Ishiguro, Hiroshi
    NEURAL NETWORKS, 2018, 107 : 23 - 33