Preference-based learning to rank

被引:2
|
作者
Nir Ailon
Mehryar Mohri
机构
[1] Computer Science Faculty,
[2] Technion – Israel Institute of Technology,undefined
[3] Courant Institute of Mathematical Sciences,undefined
来源
Machine Learning | 2010年 / 80卷
关键词
Learning to rank; Machine learning reductions; ROC;
D O I
暂无
中图分类号
学科分类号
摘要
This paper presents an efficient preference-based ranking algorithm running in two stages. In the first stage, the algorithm learns a preference function defined over pairs, as in a standard binary classification problem. In the second stage, it makes use of that preference function to produce an accurate ranking, thereby reducing the learning problem of ranking to binary classification. This reduction is based on the familiar QuickSort and guarantees an expected pairwise misranking loss of at most twice that of the binary classifier derived in the first stage. Furthermore, in the important special case of bipartite ranking, the factor of two in loss is reduced to one. This improved bound also applies to the regret achieved by our ranking and that of the binary classifier obtained.
引用
收藏
页码:189 / 211
页数:22
相关论文
共 50 条
  • [21] Dueling Posterior Sampling for Preference-Based Reinforcement Learning
    Novoseller, Ellen R.
    Wei, Yibing
    Sui, Yanan
    Yue, Yisong
    Burdick, Joel W.
    [J]. CONFERENCE ON UNCERTAINTY IN ARTIFICIAL INTELLIGENCE (UAI 2020), 2020, 124 : 1029 - 1038
  • [22] Preference-based Teaching
    Gao, Ziyuan
    Ries, Christoph
    Simon, Hans U.
    Zilles, Sandra
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2017, 18 : 1 - 32
  • [23] Preference-based unawareness
    Schipper, Burkhard C.
    [J]. MATHEMATICAL SOCIAL SCIENCES, 2014, 70 : 34 - 41
  • [24] APReL: A Library for Active Preference-based Reward Learning Algorithms
    Biyik, Erdem
    Talati, Aditi
    Sadigh, Dorsa
    [J]. PROCEEDINGS OF THE 2022 17TH ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION (HRI '22), 2022, : 613 - 617
  • [25] Preference-Based Assistance Map Learning With Robust Adaptive Oscillators
    Li, Shilei
    Zou, Wulin
    Duan, Pu
    Shi, Ling
    [J]. IEEE TRANSACTIONS ON MEDICAL ROBOTICS AND BIONICS, 2022, 4 (04): : 1000 - 1009
  • [26] Contextual Bandits and Imitation Learning with Preference-Based Active Queries
    Sekhari, Ayush
    Sridharan, Karthik
    Sun, Wen
    Wu, Runzhe
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [27] Efficient Meta Reinforcement Learning for Preference-based Fast Adaptation
    Ren, Zhizhou
    Liu, Anji
    Liang, Yitao
    Peng, Jian
    Ma, Jianzhu
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35, NEURIPS 2022, 2022,
  • [28] A Policy Iteration Algorithm for Learning from Preference-Based Feedback
    Wirth, Christian
    Furnkranz, Johannes
    [J]. ADVANCES IN INTELLIGENT DATA ANALYSIS XII, 2013, 8207 : 427 - 437
  • [29] Active Preference-Based Gaussian Process Regression for Reward Learning
    Biyik, Lirdem
    Huynh, Nicolas
    Kochenderfer, Mykel J.
    Sadigh, Dorsa
    [J]. ROBOTICS: SCIENCE AND SYSTEMS XVI, 2020,
  • [30] Preference-based Reinforcement Learning with Finite-Time Guarantees
    Xu, Yichong
    Wang, Ruosong
    Yang, Lin F.
    Singh, Aarti
    Dubrawski, Artur
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33