Mutual Information Estimation using LSH Sampling

被引:0
|
作者
Spring, Ryan [1 ]
Shrivastava, Anshumali [1 ]
机构
[1] Rice Univ, Houston, TX 77005 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Learning representations in an unsupervised or self-supervised manner is a growing area of research. Current approaches in representation learning seek to maximize the mutual information between the learned representation and original data. One of the most popular ways to estimate mutual information (MI) is based on Noise Contrastive Estimation (NCE). This MI estimate exhibits low variance, but it is upper-bounded by log(N), where N is the number of samples. In an ideal scenario, we would use the entire dataset to get the most accurate estimate. However, using such a large number of samples is computationally prohibitive. Our proposed solution is to decouple the upper-bound for the MI estimate from the sample size. Instead, we estimate the partition function of the NCE loss function for the entire dataset using importance sampling (IS). In this paper, we use locality-sensitive hashing (LSH) as an adaptive sampler and propose an unbiased estimator that accurately approximates the partition function in sublinear (near-constant) time. The samples are correlated and non-normalized, but the derived estimator is unbiased without any assumptions. We show that our LSH sampling estimate provides a superior bias-variance trade-off when compared to other state-of-the-art approaches.
引用
收藏
页码:2807 / 2815
页数:9
相关论文
共 50 条
  • [1] THE MUTUAL INFORMATION ESTIMATION IN THE SAMPLING WITH REPLACEMENT
    GIL, MA
    PEREZ, R
    MARTINEZ, I
    [J]. RAIRO-RECHERCHE OPERATIONNELLE-OPERATIONS RESEARCH, 1986, 20 (03): : 257 - 268
  • [2] THE MUTUAL INFORMATION - ESTIMATION IN THE SAMPLING WITHOUT REPLACEMENT
    GIL, MA
    PEREZ, R
    GIL, P
    [J]. KYBERNETIKA, 1987, 23 (05) : 407 - 419
  • [3] Estimation and mutual information
    Duncan, Tyrone E.
    Pasik-Duncan, Bozenna
    [J]. PROCEEDINGS OF THE 46TH IEEE CONFERENCE ON DECISION AND CONTROL, VOLS 1-14, 2007, : 2319 - 2322
  • [4] THE LAMBDA-DIVERGENCE AND THE LAMBDA-MUTUAL INFORMATION - ESTIMATION IN THE STRATIFIED SAMPLING
    MORALES, D
    PARDO, L
    SALICRU, M
    MENENDEZ, ML
    [J]. JOURNAL OF COMPUTATIONAL AND APPLIED MATHEMATICS, 1993, 47 (01) : 1 - 10
  • [5] Estimation of mutual information using copula density function
    Zeng, X.
    Durrani, T. S.
    [J]. ELECTRONICS LETTERS, 2011, 47 (08) : 493 - 494
  • [6] SCALABLE MUTUAL INFORMATION ESTIMATION USING DEPENDENCE GRAPHS
    Noshad, Morteza
    Zeng, Yu
    Hero, Alfred O., III
    [J]. 2019 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2019, : 2962 - 2966
  • [7] ESTIMATION OF MUTUAL INFORMATION USING KERNEL DENSITY ESTIMATORS
    MOON, YI
    RAJAGOPALAN, B
    LALL, U
    [J]. PHYSICAL REVIEW E, 1995, 52 (03): : 2318 - 2321
  • [8] Neural Estimators for Conditional Mutual Information Using Nearest Neighbors Sampling
    Molavipour, Sina
    Bassi, German
    Skoglund, Mikael
    [J]. IEEE TRANSACTIONS ON SIGNAL PROCESSING, 2021, 69 : 766 - 780
  • [9] Ensemble Estimation of Mutual Information
    Moon, Kevin R.
    Sricharan, Kumar
    Hero, Alfred O., III
    [J]. 2017 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2017,
  • [10] Improved Mutual Information Estimation
    Mroueh, Youssef
    Melnyk, Igor
    Dognin, Pierre
    Ross, Jarret
    Sercu, Tom
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 9009 - 9017