Logistic Regression, AdaBoost and Bregman Distances

被引:2
|
作者
Michael Collins
Robert E. Schapire
Yoram Singer
机构
[1] Shannon Laboratory,AT&T Labs—Research
[2] Hebrew University,School of Computer Science & Engineering
来源
Machine Learning | 2002年 / 48卷
关键词
logistic regression; maximum-entropy methods; boosting; AdaBoost; Bregman distances; convex optimization; iterative scaling; information geometry;
D O I
暂无
中图分类号
学科分类号
摘要
We give a unified account of boosting and logistic regression in which each learning problem is cast in terms of optimization of Bregman distances. The striking similarity of the two problems in this framework allows us to design and analyze algorithms for both simultaneously, and to easily adapt algorithms designed for one problem to the other. For both problems, we give new algorithms and explain their potential advantages over existing methods. These algorithms are iterative and can be divided into two types based on whether the parameters are updated sequentially (one at a time) or in parallel (all at once). We also describe a parameterized family of algorithms that includes both a sequential- and a parallel-update algorithm as special cases, thus showing how the sequential and parallel approaches can themselves be unified. For all of the algorithms, we give convergence proofs using a general formalization of the auxiliary-function proof technique. As one of our sequential-update algorithms is equivalent to AdaBoost, this provides the first general proof of convergence for AdaBoost. We show that all of our algorithms generalize easily to the multiclass case, and we contrast the new algorithms with the iterative scaling algorithm. We conclude with a few experimental results with synthetic data that highlight the behavior of the old and newly proposed algorithms in different settings.
引用
收藏
页码:253 / 285
页数:32
相关论文
共 50 条
  • [1] Logistic regression, AdaBoost and Bregman distances
    Collins, M
    Schapire, RE
    Singer, Y
    [J]. MACHINE LEARNING, 2002, 48 (1-3) : 253 - 285
  • [2] Bregman Distance to L1 Regularized Logistic Regression
    Das Gupta, Mithun
    Huang, Thomas S.
    [J]. 19TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOLS 1-6, 2008, : 2577 - 2580
  • [3] Comparisons of ADABOOST, KNN, SVM and Logistic Regression in Classification of Imbalanced Dataset
    Abd Rahman, Hezlin Aryani
    Wah, Yap Bee
    He, Haibo
    Bulgiba, Awang
    [J]. SOFT COMPUTING IN DATA SCIENCE, SCDS 2015, 2015, 545 : 54 - 64
  • [4] Bregman distances and Chebyshev sets
    Bauschke, Heinz H.
    Wang, Xianfu
    Ye, Jane
    Yuan, Xiaoming
    [J]. JOURNAL OF APPROXIMATION THEORY, 2009, 159 (01) : 3 - 25
  • [5] Bregman distances and Klee sets
    Bauschke, Heinz H.
    Wang, Xianfu
    Ye, Jane
    Yuan, Xiaoming
    [J]. JOURNAL OF APPROXIMATION THEORY, 2009, 158 (02) : 170 - 183
  • [6] A Note on the Approximate Symmetry of Bregman Distances
    Kindermann, Stefan
    [J]. JOURNAL OF CONVEX ANALYSIS, 2019, 26 (03) : 991 - 999
  • [7] On Bregman Distances and Divergences of Probability Measures
    Stummer, Wolfgang
    Vajda, Igor
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2012, 58 (03) : 1277 - 1288
  • [8] Comparison of AdaBoost and Logistic Regression for Detecting Colorectal Cancer Patients with Synchronous Liver Metastasis
    Wen, Jingran
    Zhang, Xiaoyan
    Xu, Ye
    Li, Zuofeng
    Liu, Lei
    [J]. 2009 INTERNATIONAL CONFERENCE ON BIOMEDICAL AND PHARMACEUTICAL ENGINEERING, 2009, : 145 - +
  • [9] Gradient regularization of Newton method with Bregman distances
    Nikita Doikov
    Yurii Nesterov
    [J]. Mathematical Programming, 2024, 204 : 1 - 25
  • [10] A framework for covariate balance using Bregman distances
    Josey, Kevin P.
    Juarez-Colunga, Elizabeth
    Yang, Fan
    Ghosh, Debashis
    [J]. SCANDINAVIAN JOURNAL OF STATISTICS, 2021, 48 (03) : 790 - 816