Oracle lower bounds for stochastic gradient sampling algorithms

被引:4
|
作者
Chatterji, Niladri S. [1 ]
Bartlett, Peter L. [2 ,3 ]
Long, Philip M. [3 ]
机构
[1] Stanford Univ, Dept Comp Sci, 353 Jane Stanford Way, Stanford, CA 94305 USA
[2] Univ Calif Berkeley, 367 Evans Hall 3860, Berkeley, CA 94720 USA
[3] Google, 1600 Amphitheatre Pkwy, Mountain View, CA 94043 USA
关键词
Sampling lower bounds; information theoretic lower bounds; Markov chain Monte Carlo; stochastic gradient Monte Carlo; HIT-AND-RUN; CONVERGENCE; COMPLEXITY; HASTINGS; VOLUME; RATES;
D O I
10.3150/21-BEJ1377
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
We consider the problem of sampling from a strongly log-concave density in R-d, and prove an information theoretic lower bound on the number of stochastic gradient queries of the log density needed. Several popular sampling algorithms (including many Markov chain Monte Carlo methods) operate by using stochastic gradients of the log density to generate a sample; our results establish an information theoretic limit for all these algorithms. We show that for every algorithm, there exists a well-conditioned strongly log-concave target density for which the distribution of points generated by the algorithm would be at least s away from the target in total variation distance if the number of gradient queries is less than Omega (sigma(2)d/epsilon(2)), where sigma(2)d is the variance of the stochastic gradient. Our lower bound follows by combining the ideas of Le Cam deficiency routinely used in the comparison of statistical experiments along with standard information theoretic tools used in lower bounding Bayes risk functions. To the best of our knowledge our results provide the first nontrivial dimension-dependent lower bound for this problem.
引用
收藏
页码:1074 / 1092
页数:19
相关论文
共 50 条
  • [1] LOWER BOUNDS FOR SAMPLING ALGORITHMS FOR ESTIMATING THE AVERAGE
    CANETTI, R
    EVEN, G
    GOLDREICH, O
    [J]. INFORMATION PROCESSING LETTERS, 1995, 53 (01) : 17 - 25
  • [2] LOWER BOUNDS ON THE WORST-CASE COMPLEXITY OF SOME ORACLE ALGORITHMS
    HAUSMANN, D
    KORTE, B
    [J]. DISCRETE MATHEMATICS, 1978, 24 (03) : 261 - 276
  • [3] Graph Oracle Models, Lower Bounds, and Gaps for Parallel Stochastic Optimization
    Woodworth, Blake
    Wang, Jialei
    Smith, Adam
    McMahan, Brendan
    Srebro, Nathan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [4] Information-Theoretic Lower Bounds on the Oracle Complexity of Stochastic Convex Optimization
    Agarwal, Alekh
    Bartlett, Peter L.
    Ravikumar, Pradeep
    Wainwright, Martin J.
    [J]. IEEE TRANSACTIONS ON INFORMATION THEORY, 2012, 58 (05) : 3235 - 3249
  • [5] SUBMODULAR APPROXIMATION: SAMPLING-BASED ALGORITHMS AND LOWER BOUNDS
    Svitkina, Zoya
    Fleischer, Lisa
    [J]. SIAM JOURNAL ON COMPUTING, 2011, 40 (06) : 1715 - 1737
  • [6] Lower Bounds for Noisy Wireless Networks using Sampling Algorithms
    Dutta, Chinmoy
    Radhakrishnan, Jaikumar
    [J]. PROCEEDINGS OF THE 49TH ANNUAL IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, 2008, : 394 - 402
  • [7] Submodular Approximation: Sampling-Based Algorithms and Lower Bounds
    Svitkina, Zoya
    Fleischer, Lisa
    [J]. PROCEEDINGS OF THE 49TH ANNUAL IEEE SYMPOSIUM ON FOUNDATIONS OF COMPUTER SCIENCE, 2008, : 697 - 706
  • [8] Sampling bounds for stochastic optimization
    Charikar, M
    Chekuri, C
    Pál, M
    [J]. APPROXIMATION, RANDOMIZATION AND COMBINATORIAL OPTIMIZATION: ALGORITHMS AND TECHNIQUES, 2005, 3624 : 257 - 269
  • [9] Lower bounds and stochastic optimization algorithms for uniform designs with three or four levels
    Fang, KT
    Maringer, D
    Tang, Y
    Winker, P
    [J]. MATHEMATICS OF COMPUTATION, 2006, 75 (254) : 859 - 878
  • [10] Information-theoretic lower bounds for zero-order stochastic gradient estimation
    Alabdulkareem, Abdulrahman
    Honorio, Jean
    [J]. 2021 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2021, : 2316 - 2321