Learning with Average Top-k Loss

被引:0
|
作者
Fan, Yanbo [1 ,3 ,4 ]
Lyu, Siwei [1 ]
Ying, Yiming [2 ]
Hu, Bao-Gang [3 ,4 ]
机构
[1] SUNY Albany, Dept Comp Sci, Albany, NY 12222 USA
[2] SUNY Albany, Dept Math & Stat, Albany, NY 12222 USA
[3] CASIA, Natl Lab Pattern Recognit, Beijing, Peoples R China
[4] Univ Chinese Acad Sci, Beijing, Peoples R China
基金
美国国家科学基金会;
关键词
SUPPORT; ALGORITHM;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this work, we introduce the average top-k (AT(k)) loss as a new aggregate loss for supervised learning, which is the average over the k largest individual losses over a training dataset. We show that the AT(k) loss is a natural generalization of the two widely used aggregate losses, namely the average loss and the maximum loss, but can combine their advantages and mitigate their drawbacks to better adapt to different data distributions. Furthermore, it remains a convex function over all individual losses, which can lead to convex optimization problems that can be solved effectively with conventional gradient-based methods. We provide an intuitive interpretation of the AT(k) loss based on its equivalent effect on the continuous individual loss functions, suggesting that it can reduce the penalty on correctly classified data. We further give a learning theory analysis of MAT(k) learning on the classification calibration of the AT(k) loss and the error bounds of AT(k)-SVM. We demonstrate the applicability of minimum average top-k learning for binary classification and regression using synthetic and real datasets.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Average Top-k Aggregate Loss for Supervised Learning
    Lyu, Siwei
    Fan, Yanbo
    Ying, Yiming
    Hu, Bao-Gang
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (01) : 76 - 86
  • [2] Group sparse additive machine with average top-k loss
    Yuan, Peipei
    You, Xinge
    Chen, Hong
    Peng, Qinmu
    Zhao, Yue
    Xu, Zhou
    Jing, Xiao-Yuan
    He, Zhenyu
    [J]. NEUROCOMPUTING, 2020, 395 : 1 - 14
  • [3] Differentiable Top-k Classification Learning
    Petersen, Felix
    Kuehne, Hilde
    Borgelt, Christian
    Deussen, Oliver
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [4] Online Learning to Rank with Top-k Feedback
    Chaudhuri, Sougata
    Tewari, Ambuj
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2017, 18
  • [5] Loss Functions for Top-k Error: Analysis and Insights
    Lapin, Maksim
    Hein, Matthias
    Schiele, Bernt
    [J]. 2016 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2016, : 1468 - 1477
  • [6] Practical Counterfactual Policy Learning for Top-K Recommendations
    Liu, Yaxu
    Yen, Jui-Nan
    Yuan, Bowen
    Shi, Rundong
    Yan, Peng
    Lin, Chih-Jen
    [J]. PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 1141 - 1151
  • [7] Learning Weighted Top-k Support Vector Machine
    Kato, Tsuyoshi
    Hirohashi, Yoshihiro
    [J]. ASIAN CONFERENCE ON MACHINE LEARNING, VOL 101, 2019, 101 : 774 - 789
  • [8] Top-k Learning to Rank: Labeling, Ranking and Evaluation
    Niu, Shuzi
    Guo, Jiafeng
    Lan, Yanyan
    Cheng, Xueqi
    [J]. SIGIR 2012: PROCEEDINGS OF THE 35TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, 2012, : 751 - 760
  • [9] Multisample-Based Contrastive Loss for Top-K Recommendation
    Tang, Hao
    Zhao, Guoshuai
    Wu, Yuxia
    Qian, Xueming
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 (339-351) : 339 - 351
  • [10] Efficient All Top-k Computation-A Unified Solution for All Top-k, Reverse Top-k and Top-m Influential Queries
    Ge, Shen
    U, Leong Hou
    Mamoulis, Nikos
    Cheung, David W.
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2013, 25 (05) : 1015 - 1027