Sparse multinomial logistic regression: Fast algorithms and generalization bounds

被引:595
|
作者
Krishnapuram, B
Carin, L
Figueiredo, MAT
Hartemink, AJ
机构
[1] Siemens Med Solut USA Inc, Comp Aided Diag & Therapy Grp, Malvern, PA 19355 USA
[2] Duke Univ, Dept Elect Engn, Durham, NC 27708 USA
[3] Inst Super Tecn, Dept Elect & Comp Engn, Inst Telecommun, P-1049001 Lisbon, Portugal
[4] Duke Univ, Dept Comp Sci, Durham, NC 27708 USA
基金
美国国家科学基金会;
关键词
supervised learning; classification; sparsity; Bayesian inference; multinomial logistic regression; bound optimization; expectation maximization (EM); learning theory; generalization bounds;
D O I
10.1109/TPAMI.2005.127
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recently developed methods for learning sparse classifiers are among the state-of-the-art in supervised learning. These methods learn classifiers that incorporate weighted sums of basis functions with sparsity-promoting priors encouraging the weight estimates to be either significantly large or exactly zero. From a learning-theoretic perspective, these methods control the capacity of the learned classifier by minimizing the number of basis functions used, resulting in better generalization. This paper presents three contributions related to learning sparse classifiers. First, we introduce a true multiclass formulation based on multinomial logistic regression. Second, by combining a bound optimization approach with a component-wise update procedure, we derive fast exact algorithms for learning sparse multiclass classifiers that scale favorably in both the number of training samples and the feature dimensionality, making them applicable even to large data sets in high-dimensional feature spaces. To the best of our knowledge, these are the first algorithms to perform exact multinomial logistic regression with a sparsity-promoting prior. Third, we show how nontrivial generalization bounds can be derived for our classifier in the binary case. Experimental results on standard benchmark data sets attest to the accuracy, sparsity, and efficiency of the proposed methods.
引用
收藏
页码:957 / 968
页数:12
相关论文
共 50 条
  • [31] Robust Multinomial Logistic Regression Based on RPCA
    Yin, Ming
    Zeng, Deyu
    Gao, Junbin
    Wu, Zongze
    Xie, Shengli
    [J]. IEEE JOURNAL OF SELECTED TOPICS IN SIGNAL PROCESSING, 2018, 12 (06) : 1144 - 1154
  • [32] Bayesian Lasso and multinomial logistic regression on GPU
    Cesnovar, Rok
    Strumbelj, Erik
    [J]. PLOS ONE, 2017, 12 (06):
  • [33] Covolutional Multinomial Logistic Regression for Face Recognition
    Ongkittikul, Surachai
    Suwatcharakulthorn, Jirawut
    Chutisowan, Kanoksom
    Ratanarangsank, Kongnat
    [J]. 2020 8TH INTERNATIONAL ELECTRICAL ENGINEERING CONGRESS (IEECON), 2020,
  • [34] Mind reading with regularized multinomial logistic regression
    Huttunen, Heikki
    Manninen, Tapio
    Kauppi, Jukka-Pekka
    Tohka, Jussi
    [J]. MACHINE VISION AND APPLICATIONS, 2013, 24 (06) : 1311 - 1325
  • [35] Bayesian multinomial logistic regression for author identification
    Madigan, D
    Genkin, A
    Lewis, DD
    Fradkin, D
    [J]. BAYESIAN INFERENCE AND MAXIMUM ENTROPY METHODS IN SCIENCE AND ENGINEERING, 2005, 803 : 509 - 516
  • [36] Sentiment Analysis Using Multinomial Logistic Regression
    Ramadhan, W. P.
    Novianty, Astri
    Setianingsih, Casi
    [J]. 2017 INTERNATIONAL CONFERENCE ON CONTROL, ELECTRONICS, RENEWABLE ENERGY AND COMMUNICATIONS (ICCREC), 2017, : 46 - 49
  • [37] Multinomial Logistic Regression For Breast Thermogram Classification
    Jha, Rashmi
    Singh, Tripty
    [J]. 2017 INTERNATIONAL CONFERENCE ON COMMUNICATION AND SIGNAL PROCESSING (ICCSP), 2017, : 1266 - 1271
  • [39] Mind reading with regularized multinomial logistic regression
    Heikki Huttunen
    Tapio Manninen
    Jukka-Pekka Kauppi
    Jussi Tohka
    [J]. Machine Vision and Applications, 2013, 24 : 1311 - 1325
  • [40] Multinomial Latent Logistic Regression for Image Understanding
    Xu, Zhe
    Hong, Zhibin
    Zhang, Ya
    Wu, Junjie
    Tsoi, Ah Chung
    Tao, Dacheng
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (02) : 973 - 987