Online Sampling from Log-Concave Distributions

被引:0
|
作者
Lee, Holden [1 ]
Mangoubi, Oren [2 ]
Vishnoi, Nisheeth K. [3 ]
机构
[1] Duke Univ, Durham, NC 27706 USA
[2] Worcester Polytech Inst, Worcester, MA 01609 USA
[3] Yale Univ, New Haven, CT 06520 USA
基金
瑞士国家科学基金会;
关键词
BINARY;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Given a sequence of convex functions integral(0), integral(1),...,integral(T) we study the problem of sampling from the Gibbs distribution pi(t) proportional to e(-) Sigma(t)(k=0) f(k) for each epoch t in an online manner. Interest in this problem derives from applications in machine learning, Bayesian statistics, and optimization where, rather than obtaining all the observations at once, one constantly acquires new data, and must continuously update the distribution. Our main result is an algorithm that generates roughly independent samples from pi(t) for every epoch t and, under mild assumptions, makes polylog(T) gradient evaluations per epoch. All previous results imply a bound on the number of gradient or function evaluations which is at least linear in T. Motivated by real-world applications, we assume that functions are smooth, their associated distributions have a bounded second moment, and their minimizer drifts in a bounded manner, but do not assume they are strongly convex. In particular, our assumptions hold for online Bayesian logistic regression, when the data satisfy natural regularity properties, giving a sampling algorithm with updates that are poly-logarithmic in T. In simulations, our algorithm achieves accuracy comparable to an algorithm specialized to logistic regression. Key to our algorithm is a novel stochastic gradient Langevin dynamics Markov chain with a carefully designed variance reduction step and constant batch size. Technically, lack of strong convexity is a significant barrier to analysis and, here, our main contribution is a martingale exit time argument that shows our Markov chain remains in a ball of radius roughly poly-logarithmic in T for enough time to reach within epsilon of pi(t).
引用
收藏
页数:12
相关论文
共 50 条
  • [21] CONCENTRATION OF THE INFORMATION IN DATA WITH LOG-CONCAVE DISTRIBUTIONS
    Bobkov, Sergey
    Madiman, Mokshay
    ANNALS OF PROBABILITY, 2011, 39 (04): : 1528 - 1543
  • [22] Concentration inequalities for ultra log-concave distributions
    Aravinda, Heshan
    Marsiglietti, Arnaud
    Melbourne, James
    STUDIA MATHEMATICA, 2022, 265 (01) : 111 - 120
  • [23] A UNIVERSAL GENERATOR FOR DISCRETE LOG-CONCAVE DISTRIBUTIONS
    HORMANN, W
    COMPUTING, 1994, 52 (01) : 89 - 96
  • [24] Query Lower Bounds for Log-concave Sampling
    Chewi, Sinho
    Pont, Jaume De Dios
    Li, Jerry
    Lu, Chen
    Narayanan, Shyam
    JOURNAL OF THE ACM, 2024, 71 (04)
  • [25] The Randomized Midpoint Method for Log-Concave Sampling
    Shen, Ruoqi
    Lee, Yin Tat
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [26] Implicit langevin algorithms for sampling from log-concave densities
    Hodgkinson, Liam
    Salomone, Robert
    Roosta, Fred
    Journal of Machine Learning Research, 2021, 22
  • [27] Implicit Langevin Algorithms for Sampling From Log-concave Densities
    Hodgkinson, Liam
    Salomone, Robert
    Roosta, Fred
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [28] A SIMPLE GENERATOR FOR DISCRETE LOG-CONCAVE DISTRIBUTIONS
    DEVROYE, L
    COMPUTING, 1987, 39 (01) : 87 - 91
  • [29] Query lower bounds for log-concave sampling
    Chewi, Sinho
    Pont, Jaume de Dios
    Li, Jerry
    Lu, Chen
    Narayanan, Shyam
    2023 IEEE 64TH ANNUAL SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, FOCS, 2023, : 2139 - 2148
  • [30] Sampling from a Log-Concave Distribution with Projected Langevin Monte Carlo
    Sébastien Bubeck
    Ronen Eldan
    Joseph Lehec
    Discrete & Computational Geometry, 2018, 59 : 757 - 783