Training Recommenders Over Large Item Corpus With Importance Sampling

被引:0
|
作者
Lian, Defu [1 ]
Gao, Zhenguo [2 ]
Song, Xia [2 ]
Li, Yucheng [1 ]
Liu, Qi [1 ]
Chen, Enhong [1 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Hefei 230052, Anhui, Peoples R China
[2] Shanghai Jiao Tong Univ, Sch Math Sci, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
Personalized ranking; cluster-based sampling; implicit feedback; item recommendation;
D O I
10.1109/TKDE.2023.3344657
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
By predicting a personalized ranking on a set of items, item recommendation helps users determine the information they need. While optimizing a ranking-focused loss is more in line with the objectives of item recommendation, previous studies have indicated that current sampling-based ranking methods don't always surpass non-sampling ones. This is because it is either inefficient to sample a pool of representative negatives for better generalization or challenging to gauge their contributions to ranking-focused losses accurately. To this end, we propose a novel weighted ranking loss, which weights each negative with the softmax probability based on model's predictive score. Our theoretical analysis suggests that optimizing this loss boosts the normalized discounted cumulative gain. Furthermore, it appears that this loss acts as an approximate analytic solution for adversarial training of personalized ranking. To improve optimization efficiency, we approximate the weighted ranking loss with self-normalized importance sampling and show that the loss has good generalization properties. To improve generalization, we further develop efficient cluster-based negative samplers based on clustering over item vectors, to decrease approximation error caused by the divergence between the proposal and the target distribution. Comprehensive evaluations on real-world datasets show that our methods remarkably outperform leading item recommendation algorithms.
引用
收藏
页码:9433 / 9447
页数:15
相关论文
共 50 条
  • [21] The imbalanced training sample problem:: Under or over sampling?
    Barandela, R
    Valdovinos, RM
    Sánchez, JS
    Ferri, FJ
    STRUCTURAL, SYNTACTIC, AND STATISTICAL PATTERN RECOGNITION, PROCEEDINGS, 2004, 3138 : 806 - 814
  • [22] Importance Sampling over Sets: A New Probabilistic Inference Scheme
    Hadjis, Stefan
    Ermon, Stefano
    UNCERTAINTY IN ARTIFICIAL INTELLIGENCE, 2015, : 355 - 364
  • [23] Application of importance sampling to the computation of large deviations in nonequilibrium processes
    Kundu, Anupam
    Sabhapandit, Sanjib
    Dhar, Abhishek
    PHYSICAL REVIEW E, 2011, 83 (03):
  • [24] Solving Large Fault Trees with Importance Sampling and Tree Search
    Zhao, Yunfei
    O'Leary, Joseph
    2024 ANNUAL RELIABILITY AND MAINTAINABILITY SYMPOSIUM, RAMS, 2024,
  • [25] Importance sampling for large ATM-type queueing networks
    LEcuyer, P
    Champoux, Y
    1996 WINTER SIMULATION CONFERENCE PROCEEDINGS, 1996, : 309 - 316
  • [26] Importance Sampling Tree for Large-scale Empirical Expectation
    Canevet, Olivier
    Jose, Cijo
    Fleuret, Francois
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 48, 2016, 48
  • [27] Large deviations for weighted empirical measures arising in importance sampling
    Hult, Henrik
    Nyquist, Pierre
    STOCHASTIC PROCESSES AND THEIR APPLICATIONS, 2016, 126 (01) : 138 - 170
  • [28] Importance sampling and Mean-Square error in neural detector training
    Sanz-González, JL
    Andina, D
    Seijas, J
    NEURAL PROCESSING LETTERS, 2002, 16 (03) : 259 - 276
  • [29] Adaptive importance sampling to accelerate training of a neural probabilistic language model
    Bengio, Yoshua
    Senecal, Jean-Sebastien
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (04): : 713 - 722
  • [30] Importance Sampling and Mean-Square Error in Neural Detector Training
    José L. Sanz-González
    Diego Andina
    Juan Seijas
    Neural Processing Letters, 2002, 16 : 259 - 276