Fine-Grained Distribution-Dependent Learning Curves

被引:0
|
作者
Bousquet, Olivier [1 ]
Hanneke, Steve [2 ]
Moran, Shay [3 ,4 ]
Shafer, Jonathan [5 ]
Tolstikhin, Ilya [1 ]
机构
[1] Google, Brain Team, Mountain View, CA 94043 USA
[2] Purdue Univ, W Lafayette, IN USA
[3] Technion Israel Inst Technol, Haifa, Israel
[4] Google Res, Mountain View, CA USA
[5] Univ Calif Berkeley, Berkeley, CA USA
来源
THIRTY SIXTH ANNUAL CONFERENCE ON LEARNING THEORY, VOL 195 | 2023年 / 195卷
关键词
PAC Learning; Online Learning; Universal Learning; Strong Minimax Lower Bounds;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning curves plot the expected error of a learning algorithm as a function of the number of labeled samples it receives from a target distribution. They are widely used as a measure of an algorithm's performance, but classic PAC learning theory cannot explain their behavior. As observed by Antos and Lugosi (1996, 1998), the classic 'No Free Lunch' lower bounds only trace the upper envelope above all learning curves of specific target distributions. For a concept class with VC dimension d the classic bound decays like d/n, yet it is possible that the learning curve for every specific distribution decays exponentially. In this case, for each n there exists a different 'hard' distribution requiring d/n samples. Antos and Lugosi asked which concept classes admit a 'strong minimax lower bound' - a lower bound of d '/n that holds for a fixed distribution for infinitely many n. We solve this problem in a principled manner, by introducing a combinatorial dimension called VCL that characterizes the best d ' for which d '/n is a strong minimax lower bound. Conceptually, the VCL dimension determines the asymptotic rate of decay of the minimax learning curve, which we call the 'distribution-free trail' of the class. Our characterization strengthens the lower bounds of Bousquet, Hanneke, Moran, van Handel, and Yehudayoff (2021), and it refines their analysis of learning curves, by showing that for classes with finite VCL the learning rate can be decomposed into a linear component that depends only on the hypothesis class and a faster (e.g., exponential) component that depends also on the target distribution. As a corollary, we recover the lower bound of Antos and Lugosi (1996, 1998) for half-spaces in R-d. Finally, to provide another viewpoint on our work and how it compares to traditional PAC learning bounds, we also present an alternative formulation of our results in a language that is closer to the PAC setting.
引用
收藏
页数:35
相关论文
共 50 条
  • [1] Improve Fine-Grained Feature Learning in Fine-Grained DataSet GAI
    Wang, Hai Peng
    Geng, Zhi Qing
    IEEE ACCESS, 2025, 13 : 12777 - 12788
  • [2] Towards Fine-Grained Recognition: Joint Learning for Object Detection and Fine-Grained Classification
    Wang, Qiaosong
    Rasmussen, Christopher
    ADVANCES IN VISUAL COMPUTING, ISVC 2019, PT II, 2019, 11845 : 332 - 344
  • [3] A FINE-GRAINED ANALYSIS ON DISTRIBUTION SHIFT
    Wiles, Olivia
    Gowal, Sven
    Stimberg, Florian
    Rebuffi, Sylvestre-Alvise
    Ktena, Ira
    Dvijotham, Krishnamurthy
    Cemgil, Taylan
    ICLR 2022 - 10th International Conference on Learning Representations, 2022,
  • [4] Fine-Grained Crowdsourcing for Fine-Grained Recognition
    Jia Deng
    Krause, Jonathan
    Li Fei-Fei
    2013 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2013, : 580 - 587
  • [5] Learning to Navigate for Fine-Grained Classification
    Yang, Ze
    Luo, Tiange
    Wang, Dong
    Hu, Zhiqiang
    Gao, Jun
    Wang, Liwei
    COMPUTER VISION - ECCV 2018, PT XIV, 2018, 11218 : 438 - 454
  • [6] Fine-Grained Partial Label Learning
    Chen, Cheng
    Lyu, Yueming
    Yu, Xingrui
    Li, Jing
    Tsang, Ivor W.
    2024 IEEE CONFERENCE ON ARTIFICIAL INTELLIGENCE, CAI 2024, 2024, : 587 - 596
  • [7] The Fine-Grained Impact of Gaming (?) on Learning
    Gong, Yue
    Beck, Joseph E.
    Heffernan, Neil T.
    Forbes-Summers, Elijah
    INTELLIGENT TUTORING SYSTEMS, PT 1, PROCEEDINGS, 2010, 6094 : 194 - 203
  • [8] A Distribution-Dependent Analysis of Meta-Learning
    Konobeev, Mikhail
    Kuzborskij, Ilja
    Szepesvari, Csaba
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [9] Distribution Shift Metric Learning for Fine-Grained Ship Classification in SAR Images
    Xu, Yongjie
    Lang, Haitao
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2020, 13 : 2276 - 2285
  • [10] ADGaze: Anisotropic Gaussian Label Distribution Learning for fine-grained gaze estimation
    Li, Duantengchuan
    Wang, Shutong
    Zhao, Wanli
    Kang, Lingyun
    Dong, Liangshan
    Wang, Jiazhang
    Wang, Xiaoguang
    PATTERN RECOGNITION, 2025, 164