Deep Learning and the Information Bottleneck Principle

被引:0
|
作者
Tishby, Naftali [1 ,2 ]
Zaslavsky, Noga [1 ]
机构
[1] Hebrew Univ Jerusalem, Edmond & Lilly Safra Ctr Brain Sci, IL-91905 Jerusalem, Israel
[2] Hebrew Univ Jerusalem, Sch Engn & Comp Sci, IL-91905 Jerusalem, Israel
关键词
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Deep Neural Networks (DNNs) are analyzed via the theoretical framework of the information bottleneck (IB) principle. We first show that any DNN can be quantified by the mutual information between the layers and the input and output variables. Using this representation we can calculate the optimal information theoretic limits of the DNN and obtain finite sample generalization bounds. The advantage of getting closer to the theoretical limit is quantifiable both by the generalization bound and by the network's simplicity. We argue that both the optimal architecture, number of layers and features/connections at each layer, are related to the bifurcation points of the information bottleneck tradeoff, namely, relevant compression of the input layer with respect to the output layer. The hierarchical representations at the layered network naturally correspond to the structural phase transitions along the information curve. We believe that this new insight can lead to new optimality bounds and deep learning algorithms.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] Deep representation learning for domain generalization with information bottleneck principle
    Zhang, Jiao
    Zhang, Xu-Yao
    Wang, Chuang
    Liu, Cheng-Lin
    [J]. PATTERN RECOGNITION, 2023, 143
  • [2] On the information bottleneck theory of deep learning
    Saxe, Andrew M.
    Bansal, Yamini
    Dapello, Joel
    Advani, Madhu
    Kolchinsky, Artemy
    Tracey, Brendan D.
    Cox, David D.
    [J]. JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2019, 2019 (12):
  • [3] Direct Validation of the Information Bottleneck Principle for Deep Nets
    Elad, Adar
    Haviv, Doron
    Blau, Yochai
    Michaeli, Tomer
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 758 - 762
  • [4] Applying the Information Bottleneck Principle to Prosodic Representation Learning
    Zhang, Guangyan
    Qin, Ying
    Tan, Daxin
    Lee, Tan
    [J]. INTERSPEECH 2021, 2021, : 3156 - 3160
  • [5] Successive Information Bottleneck and Applications in Deep Learning
    Yousfi, Yassine
    Akyol, Emrah
    [J]. 2020 54TH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS, AND COMPUTERS, 2020, : 1210 - 1213
  • [6] Information Bottleneck in Deep Learning - A Semiotic Approach
    Musat, B.
    Andonie, R.
    [J]. INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2022, 17 (01)
  • [7] Information Bottleneck: Theory and Applications in Deep Learning
    Geiger, Bernhard C.
    Kubin, Gernot
    [J]. ENTROPY, 2020, 22 (12)
  • [8] On the Difference between the Information Bottleneck and the Deep Information Bottleneck
    Wieczorek, Aleksander
    Roth, Volker
    [J]. ENTROPY, 2020, 22 (02)
  • [9] Multimodal information bottleneck for deep reinforcement learning with multiple sensors
    You, Bang
    Liu, Huaping
    [J]. NEURAL NETWORKS, 2024, 176
  • [10] Imitation Learning for Adaptive Video Streaming With Future Adversarial Information Bottleneck Principle
    Wang, Shuoyao
    Lin, Jiawei
    Ye, Fangwei
    [J]. IEEE Transactions on Mobile Computing, 2024, 23 (12) : 13670 - 13683