Sinkhorn Natural Gradient for Generative Models

被引:0
|
作者
Shen, Zebang [1 ]
Wang, Zhenfu [2 ]
Ribeiro, Alejandro [1 ]
Hassani, Hamed [1 ]
机构
[1] Univ Penn, Dept Elect & Syst Engn, Philadelphia, PA 19104 USA
[2] Univ Penn, Dept Math, Philadelphia, PA 19104 USA
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We consider the problem of minimizing a functional over a parametric family of probability measures, where the parameterization is characterized via a push-forward structure. An important application of this problem is in training generative adversarial networks. In this regard, we propose a novel Sinkhorn Natural Gradient (SiNG) algorithm which acts as a steepest descent method on the probability space endowed with the Sinkhorn divergence. We show that the Sinkhorn information matrix (SIM), a key component of SiNG, has an explicit expression and can be evaluated accurately in complexity that scales logarithmically with respect to the desired accuracy. This is in sharp contrast to existing natural gradient methods that can only be carried out approximately. Moreover, in practical applications when only Monte-Carlo type integration is available, we design an empirical estimator for SIM and provide the stability analysis. In our experiments, we quantitatively compare SiNG with state-of-the-art SGD-type solvers on generative tasks to demonstrate its efficiency and efficacy of our method.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Learning Generative Models with Sinkhorn Divergences
    Genevay, Aude
    Peyre, Gabriel
    Cuturi, Marco
    [J]. INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 84, 2018, 84
  • [2] A Gradient Descent Perspective on Sinkhorn
    Flavien Léger
    [J]. Applied Mathematics & Optimization, 2021, 84 : 1843 - 1855
  • [3] A Gradient Descent Perspective on Sinkhorn
    Leger, Flavien
    [J]. APPLIED MATHEMATICS AND OPTIMIZATION, 2021, 84 (02): : 1843 - 1855
  • [4] Don't Generate Me: Training Differentially Private Generative Models with Sinkhorn Divergence
    Cao, Tianshi
    Bie, Alex
    Vandat, Arash
    Fidler, Sanja
    Kreis, Karsten
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [5] Sinkhorn Barycenter via Functional Gradient Descent
    Shen, Zebang
    Wang, Zhenfu
    Ribeiro, Alejandro
    Hassani, Hamed
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [6] Learning generative models of natural images
    Wu, JM
    Lin, ZH
    [J]. NEURAL NETWORKS, 2002, 15 (03) : 337 - 347
  • [7] Composite Functional Gradient Learning of Generative Adversarial Models
    Johnson, Rie
    Zhang, Tong
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [8] Contextual grammars as generative models of natural languages
    Marcus, S
    Martin-Vide, C
    Paun, G
    [J]. COMPUTATIONAL LINGUISTICS, 1998, 24 (02) : 245 - 274
  • [9] Structured Generative Models of Natural Source Code
    Maddison, Chris J.
    Tarlow, Daniel
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 32 (CYCLE 2), 2014, 32 : 649 - 657
  • [10] A Framework of Composite Functional Gradient Methods for Generative Adversarial Models
    Johnson, Rie
    Zhang, Tong
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2021, 43 (01) : 17 - 32