Learning hidden variable networks: The information bottleneck approach

被引:0
|
作者
Elidan, G [1 ]
Friedman, N [1 ]
机构
[1] Hebrew Univ Jerusalem, Dept Comp Sci & Engn, IL-91904 Jerusalem, Israel
关键词
Bayesian networks; hidden variables; information bottleneck; continuation; variational methods;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A central challenge in learning probabilistic graphical models is dealing with domains that involve hidden variables. The common approach for learning model parameters in such domains is the expectation maximization ( EM) algorithm. This algorithm, however, can easily get trapped in suboptimal local maxima. Learning the model structure is even more challenging. The structural EM algorithm can adapt the structure in the presence of hidden variables, but usually performs poorly without prior knowledge about the cardinality and location of the hidden variables. In this work, we present a general approach for learning Bayesian networks with hidden variables that overcomes these problems. The approach builds on the information bottleneck framework of Tishby et al. (1999). We start by proving formal correspondence between the information bottleneck objective and the standard parametric EM functional. We then use this correspondence to construct a learning algorithm that combines an information-theoretic smoothing term with a continuation procedure. Intuitively, the algorithm bypasses local maxima and achieves superior solutions by following a continuous path from a solution of, an easy and smooth, target function, to a solution of the desired likelihood function. As we show, our algorithmic framework allows learning of the parameters as well as the structure of a network. In addition, it also allows us to introduce new hidden variables during model selection and learn their cardinality. We demonstrate the performance of our procedure on several challenging real-life data sets.
引用
收藏
页码:81 / 127
页数:47
相关论文
共 50 条
  • [1] Information Bottleneck Approach to Spatial Attention Learning
    Lai, Qiuxia
    Li, Yu
    Zeng, Ailing
    Liu, Minhao
    Sun, Hanqiu
    Xu, Qiang
    [J]. PROCEEDINGS OF THE THIRTIETH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2021, 2021, : 779 - 785
  • [2] Information Bottleneck in Deep Learning - A Semiotic Approach
    Musat, B.
    Andonie, R.
    [J]. INTERNATIONAL JOURNAL OF COMPUTERS COMMUNICATIONS & CONTROL, 2022, 17 (01)
  • [3] Predicting Disease Complications Using a Stepwise Hidden Variable Approach for Learning Dynamic Bayesian Networks
    Yousefi, Leila
    Tucker, Allan
    Al-luhaybi, Mashael
    Saachi, Lucia
    Bellazzi, Riccardo
    Chiovato, Luca
    [J]. 2018 31ST IEEE INTERNATIONAL SYMPOSIUM ON COMPUTER-BASED MEDICAL SYSTEMS (CBMS 2018), 2018, : 106 - 111
  • [4] Learning and generalization with the information bottleneck
    Shamir, Ohad
    Sabato, Sivan
    Tishby, Naftali
    [J]. THEORETICAL COMPUTER SCIENCE, 2010, 411 (29-30) : 2696 - 2711
  • [5] Learning and Generalization with the Information Bottleneck
    Shamir, Ohad
    Sabato, Sivan
    Tishby, Naftali
    [J]. ALGORITHMIC LEARNING THEORY, PROCEEDINGS, 2008, 5254 : 92 - 107
  • [6] Information Bottleneck and Aggregated Learning
    Soflaei, Masoumeh
    Zhang, Richong
    Guo, Hongyu
    Al-Bashabsheh, Ali
    Mao, Yongyi
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (12) : 14807 - 14820
  • [7] Learning to Time-Decode in Spiking Neural Networks Through the Information Bottleneck
    Skatchkovsky, Nicolas
    Simeone, Osvaldo
    Jang, Hyeryung
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [8] Deep Learning and the Information Bottleneck Principle
    Tishby, Naftali
    Zaslavsky, Noga
    [J]. 2015 IEEE INFORMATION THEORY WORKSHOP (ITW), 2015,
  • [9] On the information bottleneck theory of deep learning
    Saxe, Andrew M.
    Bansal, Yamini
    Dapello, Joel
    Advani, Madhu
    Kolchinsky, Artemy
    Tracey, Brendan D.
    Cox, David D.
    [J]. JOURNAL OF STATISTICAL MECHANICS-THEORY AND EXPERIMENT, 2019, 2019 (12):
  • [10] Information Bottleneck Methods for Distributed Learning
    Farajiparvar, Parinaz
    Beirami, Ahmad
    Nokleby, Matthew
    [J]. 2018 56TH ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2018, : 24 - 31