Energetic Natural Gradient Descent

被引:0
|
作者
Thomas, Philip S. [1 ]
da Silva, Bruno Castro [2 ]
Dann, Christoph [3 ]
Brunskill, Emma [3 ]
机构
[1] Univ Massachusetts, Amherst, MA 01003 USA
[2] Univ Fed Rio Grande do Sul, Porto Alegre, RS, Brazil
[3] Carnegie Mellon Univ, Pittsburgh, PA 15213 USA
基金
美国国家科学基金会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a new class of algorithms for minimizing or maximizing functions of parametric probabilistic models. These new algorithms are natural gradient algorithms that leverage more information than prior methods by using a new metric tensor in place of the commonly used Fisher information matrix. This new metric tensor is derived by computing directions of steepest ascent where the distance between distributions is measured using an approximation of energy distance (as opposed to Kullback-Leibler divergence, which produces the Fisher information matrix), and so we refer to our new ascent direction as the energetic natural gradient.
引用
收藏
页数:9
相关论文
共 50 条
  • [31] Nonlinear System Identification Using Neural Networks Trained with Natural Gradient Descent
    Mohamed Ibnkahla
    [J]. EURASIP Journal on Advances in Signal Processing, 2003
  • [32] Nonlinear system identification using neural networks trained with natural gradient descent
    [J]. Ibnkahla, M. (mohamed.ibnkahla@ece.queensu.ca), 1600, Hindawi Publishing Corporation (2003):
  • [33] Gradient learning in a classification setting by gradient descent
    Cai, Jia
    Wang, Hongyan
    Zhou, Ding-Xuan
    [J]. JOURNAL OF APPROXIMATION THEORY, 2009, 161 (02) : 674 - 692
  • [34] Learning to Learn Gradient Aggregation by Gradient Descent
    Ji, Jinlong
    Chen, Xuhui
    Wang, Qianlong
    Yu, Lixing
    Li, Pan
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2614 - 2620
  • [35] Stein Variational Gradient Descent as Gradient Flow
    Liu, Qiang
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [36] Stein Variational Gradient Descent Without Gradient
    Han, Jun
    Liu, Qiang
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [37] Fast yet Simple Natural-Gradient Descent for Variational Inference in Complex Models
    Khan, Mohammad Emtiyaz
    Nielsen, Didrik
    [J]. PROCEEDINGS OF 2018 INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY AND ITS APPLICATIONS (ISITA2018), 2018, : 31 - 35
  • [38] A Neural Network MLSE Receiver Based on Natural Gradient Descent: Application to Satellite Communications
    Mohamed Ibnkahla
    Jun Yuan
    [J]. EURASIP Journal on Advances in Signal Processing, 2004
  • [39] Accelerating Extreme Search of Multidimensional Functions Based on Natural Gradient Descent with Dirichlet Distributions
    Abdulkadirov, Ruslan
    Lyakhov, Pavel
    Nagornov, Nikolay
    [J]. MATHEMATICS, 2022, 10 (19)
  • [40] A neural network MLSE receiver based on natural gradient descent: Application to satellite communications
    Ibnkahla, M
    Yuan, J
    [J]. EURASIP JOURNAL ON APPLIED SIGNAL PROCESSING, 2004, 2004 (16) : 2580 - 2591