Information Bottleneck and Aggregated Learning

被引:0
|
作者
Soflaei, Masoumeh [1 ]
Zhang, Richong [2 ]
Guo, Hongyu [4 ]
Al-Bashabsheh, Ali [3 ]
Mao, Yongyi [1 ]
机构
[1] Univ Ottawa, Sch Elect Engn & Comp Sci, Ottawa, ON K1N 6N5, Canada
[2] Beihang Univ, Sch Comp Sci & Engn, SKLSDE, Beijing 100191, Peoples R China
[3] Beihang Univ, Beijing 100191, Peoples R China
[4] Natl Res Council Canada, Ottawa, ON K1A 0R6, Canada
基金
国家重点研发计划;
关键词
Aggregated learning; information bottleneck; quantization;
D O I
10.1109/TPAMI.2023.3302150
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the problem of learning a neural network classifier. Under the information bottleneck (IB) principle, we associate with this classification problem a representation learning problem, which we call "IB learning". We show that IB learning is, in fact, equivalent to a special class of the quantization problem. The classical results in rate-distortion theory then suggest that IB learning can benefit from a "vector quantization" approach, namely, simultaneously learning the representations of multiple input objects. Such an approach assisted with some variational techniques, result in a novel learning framework, "Aggregated Learning", for classification with neural network models. In this framework, several objects are jointly classified by a single neural network. The effectiveness of this framework is verified through extensive experiments on standard image recognition and text classification tasks.
引用
收藏
页码:14807 / 14820
页数:14
相关论文
共 50 条
  • [1] Learning and generalization with the information bottleneck
    Shamir, Ohad
    Sabato, Sivan
    Tishby, Naftali
    [J]. THEORETICAL COMPUTER SCIENCE, 2010, 411 (29-30) : 2696 - 2711
  • [2] Learning and Generalization with the Information Bottleneck
    Shamir, Ohad
    Sabato, Sivan
    Tishby, Naftali
    [J]. ALGORITHMIC LEARNING THEORY, PROCEEDINGS, 2008, 5254 : 92 - 107
  • [3] Deep Learning and the Information Bottleneck Principle
    Tishby, Naftali
    Zaslavsky, Noga
    [J]. 2015 IEEE INFORMATION THEORY WORKSHOP (ITW), 2015,
  • [4] On the information bottleneck theory of deep learning
    Saxe, Andrew M.
    Bansal, Yamini
    Dapello, Joel
    Advani, Madhu
    Kolchinsky, Artemy
    Tracey, Brendan D.
    Cox, David D.
    [J]. JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2019, 2019 (12):
  • [5] Information Bottleneck Methods for Distributed Learning
    Farajiparvar, Parinaz
    Beirami, Ahmad
    Nokleby, Matthew
    [J]. 2018 56TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2018, : 24 - 31
  • [6] The Role of the Information Bottleneck in Representation Learning
    Vera, Matias
    Piantanida, Pablo
    Rey Vega, Leonardo
    [J]. 2018 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2018, : 1580 - 1584
  • [7] Successive Information Bottleneck and Applications in Deep Learning
    Yousfi, Yassine
    Akyol, Emrah
    [J]. 2020 54TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2020, : 1210 - 1213
  • [8] Information Bottleneck Approach to Spatial Attention Learning
    Lai, Qiuxia
    Li, Yu
    Zeng, Ailing
    Liu, Minhao
    Sun, Hanqiu
    Xu, Qiang
    [J]. PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 779 - 785
  • [9] Graph Structure Learning with Variational Information Bottleneck
    Sun, Qingyun
    Li, Jianxin
    Peng, Hao
    Wu, Jia
    Fu, Xingcheng
    Ji, Cheng
    Yu, Philip S.
    [J]. THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / THE TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 4165 - 4174
  • [10] Information Bottleneck in Deep Learning - A Semiotic Approach
    Musat, B.
    Andonie, R.
    [J]. INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2022, 17 (01)