Intrinsically motivated reinforcement learning based recommendation with counterfactual data augmentation

被引:0
|
作者
Xiaocong Chen
Siyu Wang
Lianyong Qi
Yong Li
Lina Yao
机构
[1] University of New South Wales,School of Computer Science and Engineering
[2] China University of Petroleum (East China),College of Computer Science and Technology
[3] Tsinghua University,Department of Electronic Engineering
[4] CSIRO,Data 61
来源
World Wide Web | 2023年 / 26卷
关键词
Recommender systems; Deep reinforcement learning; Counterfactual reasoning;
D O I
暂无
中图分类号
学科分类号
摘要
Deep reinforcement learning (DRL) has shown promising results in modeling dynamic user preferences in RS in recent literature. However, training a DRL agent in the sparse RS environment poses a significant challenge. This is because the agent must balance between exploring informative user-item interaction trajectories and using existing trajectories for policy learning, a known exploration and exploitation trade-off. This trade-off greatly affects the recommendation performance when the environment is sparse. In DRL-based RS, balancing exploration and exploitation is even more challenging as the agent needs to deeply explore informative trajectories and efficiently exploit them in the context of RS. To address this issue, we propose a novel intrinsically motivated reinforcement learning (IMRL) method that enhances the agent’s capability to explore informative interaction trajectories in the sparse environment. We further enrich these trajectories via an adaptive counterfactual augmentation strategy with a customised threshold to improve their efficiency in exploitation. Our approach is evaluated on six offline datasets and three online simulation platforms, demonstrating its superiority over existing state-of-the-art methods. The extensive experiments show that our IMRL method outperforms other methods in terms of recommendation performance in the sparse RS environment.
引用
收藏
页码:3253 / 3274
页数:21
相关论文
共 50 条
  • [41] Counterfactual based reinforcement learning for graph neural networks
    Pham, David
    Zhang, Yongfeng
    ANNALS OF OPERATIONS RESEARCH, 2022,
  • [42] Automatic Data Augmentation for Generalization in Reinforcement Learning
    Raileanu, Roberta
    Goldstein, Max
    Yarats, Denis
    Kostrikov, Ilya
    Fergus, Rob
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [43] Generalization in Reinforcement Learning by Soft Data Augmentation
    Hansen, Nicklas
    Wang, Xiaolong
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 13611 - 13617
  • [44] Multimodal Counterfactual Learning Network for Multimedia-based Recommendation
    Li, Shuaiyang
    Guo, Dan
    Liu, Kang
    Hong, Richang
    Xue, Feng
    PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023, 2023, : 1539 - 1548
  • [45] Reinforcement Learning in Non-Stationary Environments: An Intrinsically Motivated Stress Based Memory Retrieval Performance (SBMRP) Model
    Tang, Tiong Yew
    Egerton, Simon
    Kubota, Naoyuki
    2014 IEEE INTERNATIONAL CONFERENCE ON FUZZY SYSTEMS (FUZZ-IEEE), 2014, : 1728 - 1735
  • [46] Graph contrastive learning for recommendation with generative data augmentation
    Li, Xiaoge
    Wang, Yin
    Wang, Yihan
    An, Xiaochun
    MULTIMEDIA SYSTEMS, 2024, 30 (04)
  • [47] Counterfactual Embedding Learning for Debiased Recommendation
    Jian, Meng
    Guo, Jingjing
    Xiang, Ye
    Wu, Lifang
    2021 IEEE SEVENTH INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM 2021), 2021, : 69 - 73
  • [48] Plug-and-Play Model-Agnostic Counterfactual Policy Synthesis for Deep Reinforcement Learning-Based Recommendation
    Wang, Siyu
    Chen, Xiaocong
    McAuley, Julian
    Cripps, Sally
    Yao, Lina
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2025, 36 (01) : 1044 - 1055
  • [49] Research on big data personalised recommendation model based on deep reinforcement learning
    Shi H.
    Shang L.
    International Journal of Networking and Virtual Organisations, 2023, 28 (2-4) : 364 - 380
  • [50] Robust Intrinsically Motivated Exploration and Active Learning
    Baranes, Adrien
    Oudeyer, Pierre-Yves
    2009 IEEE 8TH INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING, 2009, : 124 - 129