Information Bottleneck Theory Based Exploration of Cascade Learning

被引:1
|
作者
Du, Xin [1 ]
Farrahi, Katayoun [1 ]
Niranjan, Mahesan [1 ]
机构
[1] Univ Southampton, Sch Elect & Comp Sci, Southampton SO17 3AS, Hants, England
基金
英国工程与自然科学研究理事会;
关键词
information bottleneck theory; Cascade Learning; neural networks;
D O I
10.3390/e23101360
中图分类号
O4 [物理学];
学科分类号
0702 ;
摘要
In solving challenging pattern recognition problems, deep neural networks have shown excellent performance by forming powerful mappings between inputs and targets, learning representations (features) and making subsequent predictions. A recent tool to help understand how representations are formed is based on observing the dynamics of learning on an information plane using mutual information, linking the input to the representation (I(X;T)) and the representation to the target (I(T;Y)). In this paper, we use an information theoretical approach to understand how Cascade Learning (CL), a method to train deep neural networks layer-by-layer, learns representations, as CL has shown comparable results while saving computation and memory costs. We observe that performance is not linked to information-compression, which differs from observation on End-to-End (E2E) learning. Additionally, CL can inherit information about targets, and gradually specialise extracted features layer-by-layer. We evaluate this effect by proposing an information transition ratio, I(T;Y)/I(X;T), and show that it can serve as a useful heuristic in setting the depth of a neural network that achieves satisfactory accuracy of classification.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] A parallel clustering method combined information bottleneck theory and centroid-based clustering
    Zhanquan Sun
    Geoffrey Fox
    Weidong Gu
    Zhao Li
    The Journal of Supercomputing, 2014, 69 : 452 - 467
  • [42] MVAIBNet: Multiview Disentangled Representation Learning With Information Bottleneck
    Yin, Ming
    Liu, Xin
    Gao, Junli
    Yuan, Haoliang
    Jin, Taisong
    Zhang, Shengwei
    Li, Lingling
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, : 1 - 10
  • [43] β-CapsNet: learning disentangled representation for CapsNet by information bottleneck
    Hu, Ming-fei
    Liu, Jian-wei
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (03): : 2503 - 2525
  • [44] Learning hidden variable networks: The information bottleneck approach
    Elidan, G
    Friedman, N
    JOURNAL OF MACHINE LEARNING RESEARCH, 2005, 6 : 81 - 127
  • [45] Information Bottleneck Based Joint Feedback and Channel Learning in FDD Massive MIMO Systems
    Cao, Jiaqi
    Lian, Lixiang
    2022 IEEE GLOBAL COMMUNICATIONS CONFERENCE (GLOBECOM 2022), 2022, : 1442 - 1447
  • [46] Applying the Information Bottleneck Principle to Prosodic Representation Learning
    Zhang, Guangyan
    Qin, Ying
    Tan, Daxin
    Lee, Tan
    INTERSPEECH 2021, 2021, : 3156 - 3160
  • [47] Learning Representations for Neural Network-Based Classification Using the Information Bottleneck Principle
    Amjad, Rana Ali
    Geiger, Bernhard C.
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2020, 42 (09) : 2225 - 2239
  • [48] SNIB: Improving Spike-Based Machine Learning Using Nonlinear Information Bottleneck
    Yang, Shuangming
    Chen, Badong
    IEEE TRANSACTIONS ON SYSTEMS MAN CYBERNETICS-SYSTEMS, 2023, 53 (12): : 7852 - 7863
  • [49] β-CapsNet: learning disentangled representation for CapsNet by information bottleneck
    Ming-fei Hu
    Jian-wei Liu
    Neural Computing and Applications, 2023, 35 : 2503 - 2525
  • [50] The information bottleneck problem and its applications in machine learning
    Goldfeld Z.
    Polyanskiy Y.
    IEEE Journal on Selected Areas in Information Theory, 2020, 1 (01): : 19 - 38