Large-Scale Interactive Recommendation With Tree-Structured Reinforcement Learning

被引:4
|
作者
Chen, Haokun [1 ,2 ]
Zhu, Chenxu [1 ]
Tang, Ruiming [3 ]
Zhang, Weinan [1 ]
He, Xiuqiang [3 ]
Yu, Yong [1 ]
机构
[1] Shanghai Jiao Tong Univ, Shanghai 200240, Peoples R China
[2] Alibaba Grp, Hangzhou 310052, Peoples R China
[3] Huawei Noahs Ark Lab, Shenzhen 518129, Peoples R China
基金
中国国家自然科学基金;
关键词
Reinforcement learning; Predictive models; interactive recommender system; large-scale recommendation; cold-start; ALGORITHMS;
D O I
10.1109/TKDE.2021.3137310
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Although reinforcement learning (RL) techniques are regarded as promising solutions for interactive recommender systems (IRS), such solutions still face three main challenges, namely, i) time inefficiency when handling large discrete action space in IRS, ii) inability to deal with the cold-start scenarios in IRS, iii) data inefficiency during training the RL-based methods. To tackle these challenges, we propose a generic tree-structured RL framework taking both policy-based and value-based approaches into consideration. We propose to construct a balanced tree over representations of the items, such that picking an item is formulated as seeking a suitable path from the root to a leaf node in the balanced tree, which dramatically reduces the time complexity of item recommendation. Further, for cold-start scenarios where prior information of the items is unavailable, we initialize a random balanced tree as the starting point and then refine the tree structure based on the learned item representations. Besides, we also incorporate a user modeling component to explicitly model the environment, which can be utilized in the training phase to improve data efficiency. Extensive experiments on two real-world datasets are conducted and demonstrate that our framework can achieve superior recommendation performance and provide time and data efficiency improvement over state-of-the-art methods in both warm-start and cold-start IRS scenarios.
引用
收藏
页码:4018 / 4032
页数:15
相关论文
共 50 条
  • [21] Blackthorn: Large-Scale Interactive Multimodal Learning
    Zahalka, Jan
    Rudinac, Stevan
    Jonsson, Bjorn Dor
    Koelma, Dennis C.
    Worring, Marcel
    IEEE TRANSACTIONS ON MULTIMEDIA, 2018, 20 (03) : 687 - 698
  • [22] Tractable large-scale deep reinforcement learning
    Sarang, Nima
    Poullis, Charalambos
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2023, 232
  • [23] WebLens: Towards Interactive Large-scale Structured Data Profiling
    Khan, Rituparna
    Gubanov, Michael
    CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 3425 - 3428
  • [24] Tree-structured scale effects in binary and ordinal regression
    Tutz, Gerhard
    Berger, Moritz
    STATISTICS AND COMPUTING, 2021, 31 (02)
  • [25] EFFICIENT CODERS FOR LARGE TREE-STRUCTURED DICTIONARIES OF TILINGS
    Hua, K. -L.
    Zhang, R.
    Comer, M.
    Pollak, I.
    2011 CONFERENCE RECORD OF THE FORTY-FIFTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS (ASILOMAR), 2011, : 230 - 234
  • [26] Tree-structured scale effects in binary and ordinal regression
    Gerhard Tutz
    Moritz Berger
    Statistics and Computing, 2021, 31
  • [27] Predictive learning on hidden tree-structured ising models
    Nikolakakis, Konstantinos E.
    Kalogerias, Dionysios S.
    Sarwate, Anand D.
    Journal of Machine Learning Research, 2021, 22
  • [28] Learning Tree-structured Approximations for Conditional Random Fields
    Skurikhin, Alexei N.
    2014 IEEE APPLIED IMAGERY PATTERN RECOGNITION WORKSHOP (AIPR), 2014,
  • [29] Learning Tree-structured Descriptor Quantizers for Image Categorization
    Krapac, Josip
    Verbeek, Jakob
    Jurie, Frederic
    PROCEEDINGS OF THE BRITISH MACHINE VISION CONFERENCE 2011, 2011,
  • [30] Predictive Learning on Hidden Tree-Structured Ising Models
    Nikolakakis, Konstantinos E.
    Kalogerias, Dionysios S.
    Sarwate, Anand D.
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22