On the convergence of the Monte Carlo maximum likelihood method for latent variable models

被引:9
|
作者
Cappé, O
Douc, R
Moulines, E
Robert, C
机构
[1] ENST, Dept TSI, F-75634 Paris 13, France
[2] Univ Paris 09, F-75775 Paris, France
关键词
maximum likelihood estimation; Monte Carlo maximum likelihood; simulated likelihood ratio; stochastic approximation; stochastic optimization;
D O I
10.1111/1467-9469.00309
中图分类号
O21 [概率论与数理统计]; C8 [统计学];
学科分类号
020208 ; 070103 ; 0714 ;
摘要
While much used in practice, latent variable models raise challenging estimation problems due to the intractability of their likelihood. Monte Carlo maximum likelihood (MCML), as proposed by Geyer & Thompson (1992), is a simulation-based approach to maximum likelihood approximation applicable to general latent variable models. MCML can be described as an importance sampling method in which the likelihood ratio is approximated by Monte Carlo averages of importance ratios simulated from the complete data model corresponding to an arbitrary value phi of the unknown parameter. This paper studies the asymptotic (in the number of observations) performance of the MCML method in the case of latent variable models with independent observations. This is in contrast with previous works on the same topic which only considered conditional convergence to the maximum likelihood estimator, for a fixed set of observations. A first important result is that when phi is fixed, the MCML method can only be consistent if the number of simulations grows exponentially fast with the number of observations. If on the other hand, phi is obtained from a consistent sequence of estimates of the unknown parameter, then the requirements on the number of simulations are shown to be much weaker.
引用
收藏
页码:615 / 635
页数:21
相关论文
共 50 条