Learning the Stein Discrepancy for Training and Evaluating Energy-Based Models without Sampling

被引:0
|
作者
Grathwohl, Will [1 ,2 ]
Wang, Kuan-Chieh [1 ,2 ]
Jacobsen, Jorn-Henrik [1 ,2 ]
Duvenaud, David [1 ,2 ]
Zemel, Richard [1 ,2 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Vector Inst, Toronto, ON, Canada
关键词
NETWORKS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We present a new method for evaluating and training unnormalized density models. Our approach only requires access to the gradient of the unnormalized model's log-density. We estimate the Stein discrepancy between the data density p(x) and the model density q(x) defined by a vector function of the data. We parameterize this function with a neural network and fit its parameters to maximize the discrepancy. This yields a novel goodness-of-fit test which outperforms existing methods on high dimensional data. Furthermore, optimizing q(x) to minimize this discrepancy produces a novel method for training unnormalized models which scales more gracefully than existing methods. The ability to both learn and compare models is a unique feature of the proposed method.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] The physics of energy-based models
    Patrick Huembeli
    Juan Miguel Arrazola
    Nathan Killoran
    Masoud Mohseni
    Peter Wittek
    Quantum Machine Intelligence, 2022, 4
  • [22] A New Method for Evaluating Liquefaction by Energy-Based Pore Water Pressure Models
    Zhang, Jianlei
    Cheng, Qiangong
    Fan, Haozhen
    Dai, Mengjie
    Li, Yan
    Wu, Jiujiang
    Wang, Yufeng
    COATINGS, 2025, 15 (01):
  • [23] On the Anatomy of MCMC-Based Maximum Likelihood Learning of Energy-Based Models
    Nijkamp, Erik
    Hill, Mitch
    Han, Tian
    Zhu, Song-Chun
    Wu, Ying Nian
    THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 5272 - 5280
  • [24] Energy-Based Models in Earth Observation: From Generation to Semisupervised Learning
    Castillo-Navarro, Javiera
    Le Saux, Bertrand
    Boulch, Alexandre
    Lefevre, Sebastien
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2022, 60
  • [25] Efficient training of energy-based models via spin-glass control
    Pozas-Kerstjens, Alejandro
    Munoz-Gil, Gorka
    Pinol, Eloy
    Garcia-March, Miguel Angel
    Acin, Antonio
    Lewenstein, Maciej
    Grzybowski, Przemyslaw R.
    MACHINE LEARNING-SCIENCE AND TECHNOLOGY, 2021, 2 (02):
  • [26] Surprise Minimizing Multi-Agent Learning with Energy-based Models
    Suri, Karush
    Shi, Xiao Qi
    Plataniotis, Konstantinos
    Lawryshyn, Yuri
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [27] Explaining the effects of non-convergent MCMC in the training of Energy-Based Models
    Agoritsas, Elisabeth
    Catania, Giovanni
    Decelle, Aurelien
    Seoane, Beatriz
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 202, 2023, 202 : 322 - 336
  • [28] EBM-WGF: Training energy-based models with Wasserstein gradient flow
    Wan, Ben
    Geng, Cong
    Zheng, Tianyi
    Wang, Jia
    Neural Networks, 2025, 187
  • [29] Learning With Kernel Smoothing Models and Low-Discrepancy Sampling
    Cervellera, Cristiano
    Maccio, Danilo
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2013, 24 (03) : 504 - 509
  • [30] Deep Energy-Based NARX Models
    Hendriks, Johannes N.
    Gustafsson, Fredrik K.
    Ribeiro, Antonio H.
    Wills, Adrian G.
    Schon, Thomas B.
    IFAC PAPERSONLINE, 2021, 54 (07): : 505 - 510