Gradient descent optimization of smoothed information retrieval metrics

被引:0
|
作者
Olivier Chapelle
Mingrui Wu
机构
[1] Yahoo! Labs,
来源
Information Retrieval | 2010年 / 13卷
关键词
Learning to rank; Gradient descent; Annealing;
D O I
暂无
中图分类号
学科分类号
摘要
Most ranking algorithms are based on the optimization of some loss functions, such as the pairwise loss. However, these loss functions are often different from the criteria that are adopted to measure the quality of the web page ranking results. To overcome this problem, we propose an algorithm which aims at directly optimizing popular measures such as the Normalized Discounted Cumulative Gain and the Average Precision. The basic idea is to minimize a smooth approximation of these measures with gradient descent. Crucial to this kind of approach is the choice of the smoothing factor. We provide various theoretical analysis on that choice and propose an annealing algorithm to iteratively minimize a less and less smoothed approximation of the measure of interest. Results on the Letor benchmark datasets show that the proposed algorithm achieves state-of-the-art performances.
引用
收藏
页码:216 / 235
页数:19
相关论文
共 50 条
  • [21] Beyond Online Balanced Descent: An Optimal Algorithm for Smoothed Online Optimization
    Goel, Gautam
    Lin, Yiheng
    Sun, Haoyuan
    Wierman, Adam
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [22] NOMA Codebook Optimization by Batch Gradient Descent
    Si, Zhongwei
    Wen, Shaoguo
    Dong, Bing
    [J]. IEEE ACCESS, 2019, 7 : 117274 - 117281
  • [23] Stochastic gradient descent for optimization for nuclear systems
    Williams, Austin
    Walton, Noah
    Maryanski, Austin
    Bogetic, Sandra
    Hines, Wes
    Sobes, Vladimir
    [J]. SCIENTIFIC REPORTS, 2023, 13 (01)
  • [24] Distributed Optimization with Gradient Descent and Quantized Communication
    Rikos, Apostolos I.
    Jiang, Wei
    Charalambous, Themistoklis
    Johansson, Karl H.
    [J]. IFAC PAPERSONLINE, 2023, 56 (02): : 5900 - 5906
  • [25] Limitations of Information-Theoretic Generalization Bounds for Gradient Descent Methods in Stochastic Convex Optimization
    Haghifam, Mahdi
    Rodriguez-Galvez, Borja
    Thobaben, Ragnar
    Skoglund, Mikael
    Roy, Daniel M.
    Dziugaite, Gintare Karolina
    [J]. INTERNATIONAL CONFERENCE ON ALGORITHMIC LEARNING THEORY, VOL 201, 2023, 201 : 663 - 706
  • [26] Varying similarity metrics in visual information retrieval
    Jin, JS
    Kurniawati, R
    [J]. PATTERN RECOGNITION LETTERS, 2001, 22 (05) : 583 - 592
  • [27] Objective Metrics and Gradient Descent Algorithms for Adversarial Examples in Machine Learning
    Jang, Uyeong
    Wu, Xi
    Jha, Somesh
    [J]. 33RD ANNUAL COMPUTER SECURITY APPLICATIONS CONFERENCE (ACSAC 2017), 2017, : 262 - 277
  • [28] Online convex optimization in the bandit setting: gradient descent without a gradient
    Flaxman, Abraham D.
    Kalai, Adam Tauman
    McMahan, H. Brendan
    [J]. PROCEEDINGS OF THE SIXTEENTH ANNUAL ACM-SIAM SYMPOSIUM ON DISCRETE ALGORITHMS, 2005, : 385 - 394
  • [29] Kernel gradient descent algorithm for information theoretic learning
    Hu, Ting
    Wu, Qiang
    Zhou, Ding-Xuan
    [J]. JOURNAL OF APPROXIMATION THEORY, 2021, 263
  • [30] Information cut for clustering using a gradient descent approach
    Jenssen, Robert
    Erdogmus, Deniz
    Hild, Kenneth E., II
    Principe, Jose C.
    Eltoft, Torbjorn
    [J]. PATTERN RECOGNITION, 2007, 40 (03) : 796 - 806