Adaptive Natural Gradient Learning Algorithms for Unnormalized Statistical Models

被引:1
|
作者
Karakida, Ryo [1 ]
Okada, Masato [1 ,2 ]
Amari, Shun-ichi [2 ]
机构
[1] Univ Tokyo, 5-1-5 Kashiwanoha, Kashiwa, Chiba 2778561, Japan
[2] RIKEN Brain Sci Inst, 2-1 Hirosawa, Wako, Saitama 3510198, Japan
关键词
Natural gradient; Score matching; Ratio matching; Unnormalized statistical model; Multi-layer neural network;
D O I
10.1007/978-3-319-44778-0_50
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The natural gradient is a powerful method to improve the transient dynamics of learning by utilizing the geometric structure of the parameter space. Many natural gradient methods have been developed for maximum likelihood learning, which is based on Kullback-Leibler (KL) divergence and its Fisher metric. However, they require the computation of the normalization constant and are not applicable to statistical models with an analytically intractable normalization constant. In this study, we extend the natural gradient framework to divergences for the unnormalized statistical models: score matching and ratio matching. In addition, we derive novel adaptive natural gradient algorithms that do not require computationally demanding inversion of the metric and show their effectiveness in some numerical experiments. In particular, experimental results in a multi-layer neural network model demonstrate that the proposed method can escape from the plateau phenomena much faster than the conventional stochastic gradient descent method.
引用
收藏
页码:427 / 434
页数:8
相关论文
共 50 条