Training Recommenders Over Large Item Corpus With Importance Sampling

被引:0
|
作者
Lian, Defu [1 ]
Gao, Zhenguo [2 ]
Song, Xia [2 ]
Li, Yucheng [1 ]
Liu, Qi [1 ]
Chen, Enhong [1 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, Hefei 230052, Anhui, Peoples R China
[2] Shanghai Jiao Tong Univ, Sch Math Sci, Shanghai 200240, Peoples R China
基金
中国国家自然科学基金;
关键词
Personalized ranking; cluster-based sampling; implicit feedback; item recommendation;
D O I
10.1109/TKDE.2023.3344657
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
By predicting a personalized ranking on a set of items, item recommendation helps users determine the information they need. While optimizing a ranking-focused loss is more in line with the objectives of item recommendation, previous studies have indicated that current sampling-based ranking methods don't always surpass non-sampling ones. This is because it is either inefficient to sample a pool of representative negatives for better generalization or challenging to gauge their contributions to ranking-focused losses accurately. To this end, we propose a novel weighted ranking loss, which weights each negative with the softmax probability based on model's predictive score. Our theoretical analysis suggests that optimizing this loss boosts the normalized discounted cumulative gain. Furthermore, it appears that this loss acts as an approximate analytic solution for adversarial training of personalized ranking. To improve optimization efficiency, we approximate the weighted ranking loss with self-normalized importance sampling and show that the loss has good generalization properties. To improve generalization, we further develop efficient cluster-based negative samplers based on clustering over item vectors, to decrease approximation error caused by the divergence between the proposal and the target distribution. Comprehensive evaluations on real-world datasets show that our methods remarkably outperform leading item recommendation algorithms.
引用
收藏
页码:9433 / 9447
页数:15
相关论文
共 50 条
  • [11] Adaptive Importance Sampling technique for neural detector training
    Sanz-González, JL
    Alvarez-Vaquero, F
    ARTIFICIAL NEURAL NETWORKS - ICANN 2002, 2002, 2415 : 1037 - 1042
  • [12] Importance sampling for TCM scheme over fading channel
    Sakai, T
    Shibata, K
    IEICE TRANSACTIONS ON FUNDAMENTALS OF ELECTRONICS COMMUNICATIONS AND COMPUTER SCIENCES, 2002, E85A (10): : 2272 - 2275
  • [13] Large and moderate deviations for importance sampling in the Heston model
    Geha, Marc
    Jacquier, Antoine
    Zuric, Zan
    ANNALS OF OPERATIONS RESEARCH, 2024, 336 (1-2) : 47 - 92
  • [14] Large and moderate deviations for importance sampling in the Heston model
    Marc Geha
    Antoine Jacquier
    Žan Žurič
    Annals of Operations Research, 2024, 336 : 47 - 92
  • [15] Efficient Large Deviation Estimation Based on Importance Sampling
    Guyader, Arnaud
    Touchette, Hugo
    JOURNAL OF STATISTICAL PHYSICS, 2020, 181 (02) : 551 - 586
  • [16] Efficient Large Deviation Estimation Based on Importance Sampling
    Arnaud Guyader
    Hugo Touchette
    Journal of Statistical Physics, 2020, 181 : 551 - 586
  • [17] State-dependent importance sampling and large deviations
    Statistics Department, Harvard University, Cambridge, MA, United States
    不详
    ACM Int. Conf. Proc. Ser., 2006,
  • [18] Classifier Learning from Imbalanced Corpus by Autoencoded Over-Sampling
    Park, Eunkyung
    Wong, Raymond K.
    Chu, Victor W.
    PRICAI 2019: TRENDS IN ARTIFICIAL INTELLIGENCE, PT I, 2019, 11670 : 16 - 29
  • [19] Sampling Node Pairs Over Large Graphs
    Wang, Pinghui
    Zhao, Junzhou
    Lui, John C. S.
    Towsley, Don
    Guan, Xiaohong
    2013 IEEE 29TH INTERNATIONAL CONFERENCE ON DATA ENGINEERING (ICDE), 2013, : 781 - 792
  • [20] Training Deep Models Faster with Robust, Approximate Importance Sampling
    Johnson, Tyler B.
    Guestrin, Carlos
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31