Intrinsically motivated reinforcement learning based recommendation with counterfactual data augmentation

被引:0
|
作者
Xiaocong Chen
Siyu Wang
Lianyong Qi
Yong Li
Lina Yao
机构
[1] University of New South Wales,School of Computer Science and Engineering
[2] China University of Petroleum (East China),College of Computer Science and Technology
[3] Tsinghua University,Department of Electronic Engineering
[4] CSIRO,Data 61
来源
World Wide Web | 2023年 / 26卷
关键词
Recommender systems; Deep reinforcement learning; Counterfactual reasoning;
D O I
暂无
中图分类号
学科分类号
摘要
Deep reinforcement learning (DRL) has shown promising results in modeling dynamic user preferences in RS in recent literature. However, training a DRL agent in the sparse RS environment poses a significant challenge. This is because the agent must balance between exploring informative user-item interaction trajectories and using existing trajectories for policy learning, a known exploration and exploitation trade-off. This trade-off greatly affects the recommendation performance when the environment is sparse. In DRL-based RS, balancing exploration and exploitation is even more challenging as the agent needs to deeply explore informative trajectories and efficiently exploit them in the context of RS. To address this issue, we propose a novel intrinsically motivated reinforcement learning (IMRL) method that enhances the agent’s capability to explore informative interaction trajectories in the sparse environment. We further enrich these trajectories via an adaptive counterfactual augmentation strategy with a customised threshold to improve their efficiency in exploitation. Our approach is evaluated on six offline datasets and three online simulation platforms, demonstrating its superiority over existing state-of-the-art methods. The extensive experiments show that our IMRL method outperforms other methods in terms of recommendation performance in the sparse RS environment.
引用
收藏
页码:3253 / 3274
页数:21
相关论文
共 50 条
  • [31] Emotionally motivated reinforcement learning based controller
    Ayesh, A
    2004 IEEE INTERNATIONAL CONFERENCE ON SYSTEMS, MAN & CYBERNETICS, VOLS 1-7, 2004, : 874 - 878
  • [32] Motivated reinforcement learning
    Dayan, P
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 14, VOLS 1 AND 2, 2002, 14 : 11 - 18
  • [33] MOCODA: Model-based Counterfactual Data Augmentation
    Pitis, Silviu
    Creager, Elliot
    Mandlekar, Ajay
    Garg, Animesh
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [34] Intrinsically Motivated Multimodal Structure Learning
    Wong, Jay Ming
    Grupen, Roderic A.
    2016 JOINT IEEE INTERNATIONAL CONFERENCE ON DEVELOPMENT AND LEARNING AND EPIGENETIC ROBOTICS (ICDL-EPIROB), 2016, : 260 - 261
  • [35] Learning Path Recommendation Based on Reinforcement Learning
    Li, Ji
    Yu, Simiao
    Zhang, Tiancheng
    ENGINEERING LETTERS, 2024, 32 (09) : 1823 - 1832
  • [36] A novel deep learning motivated data augmentation system based on defect segmentation requirements
    Niu, Shuanlong
    Peng, Yaru
    Li, Bin
    Qiu, Yuanhong
    Niu, Tongzhi
    Li, Weifeng
    JOURNAL OF INTELLIGENT MANUFACTURING, 2024, 35 (02) : 687 - 701
  • [37] A Reinforcement Learning Based Tag Recommendation
    Ge, Feng
    He, Yi
    Liu, Jin
    Lv, Xiaoming
    Zhang, Wensheng
    Li, Yiqun
    PRACTICAL APPLICATIONS OF INTELLIGENT SYSTEMS, 2011, 124 : 251 - +
  • [38] A novel deep learning motivated data augmentation system based on defect segmentation requirements
    Shuanlong Niu
    Yaru Peng
    Bin Li
    Yuanhong Qiu
    Tongzhi Niu
    Weifeng Li
    Journal of Intelligent Manufacturing, 2024, 35 : 687 - 701
  • [39] Using Data Augmentation Based Reinforcement Learning for Daily Stock Trading
    Yuan, Yuyu
    Wen, Wen
    Yang, Jincui
    ELECTRONICS, 2020, 9 (09) : 1 - 13
  • [40] Boundary Data Augmentation for Offline Reinforcement Learning
    SHEN Jiahao
    JIANG Ke
    TAN Xiaoyang
    ZTE Communications, 2023, 21 (03) : 29 - 36