Learning the Stein Discrepancy for Training and Evaluating Energy-Based Models without Sampling

被引:0
|
作者
Grathwohl, Will [1 ,2 ]
Wang, Kuan-Chieh [1 ,2 ]
Jacobsen, Jorn-Henrik [1 ,2 ]
Duvenaud, David [1 ,2 ]
Zemel, Richard [1 ,2 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Vector Inst, Toronto, ON, Canada
关键词
NETWORKS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We present a new method for evaluating and training unnormalized density models. Our approach only requires access to the gradient of the unnormalized model's log-density. We estimate the Stein discrepancy between the data density p(x) and the model density q(x) defined by a vector function of the data. We parameterize this function with a neural network and fit its parameters to maximize the discrepancy. This yields a novel goodness-of-fit test which outperforms existing methods on high dimensional data. Furthermore, optimizing q(x) to minimize this discrepancy produces a novel method for training unnormalized models which scales more gracefully than existing methods. The ability to both learn and compare models is a unique feature of the proposed method.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Energy-based models for environmental biotechnology
    Rodriguez, Jorge
    Lema, Juan M.
    Kleerebezem, Robbert
    TRENDS IN BIOTECHNOLOGY, 2008, 26 (07) : 366 - 374
  • [32] Energy-Based Models of P Systems
    Mauri, Giancarlo
    Leporati, Alberto
    Zandron, Claudio
    MEMBRANE COMPUTING, 2010, 5957 : 104 - 124
  • [33] Residual Energy-Based Models for Text
    Bakhtin, Anton
    Deng, Yuntian
    Gross, Sam
    Ott, Myle
    Ranzato, Marc'Aurelio
    Szlam, Arthur
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22
  • [34] Residual energy-based models for text
    Bakhtin, Anton
    Deng, Yuntian
    Gross, Sam
    Ott, Myle
    Ranzato, Marc'Aurelio
    Szlam, Arthur
    Journal of Machine Learning Research, 2021, 22
  • [35] Energy-Based Reranking: Improving Neural Machine Translation Using Energy-Based Models
    Bhattacharyya, Sumanta
    Rooshenas, Amirmohammad
    Naskar, Subhajit
    Sun, Simeng
    Iyyer, Mohit
    McCallum, Andrew
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING (ACL-IJCNLP 2021), VOL 1, 2021, : 4528 - 4537
  • [36] Learning Equivariant Energy Based Models with Equivariant Stein Variational Gradient Descent
    Jaini, Priyank
    Holdijk, Lars
    Welling, Max
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [37] Energy-based dissolution simulation using SPH sampling
    Jiang, Min
    Southern, Richard
    Zhang, Jiang J.
    COMPUTER ANIMATION AND VIRTUAL WORLDS, 2018, 29 (02)
  • [38] Exploring Energy-based Language Models with Different Architectures and Training Methods for Speech Recognition
    Liu, Hong
    Lv, Zhaobiao
    Ou, Zhijian
    Zhao, Wenbo
    Xiao, Qing
    INTERSPEECH 2023, 2023, : 476 - 480
  • [39] Learning Energy-Based Models for 3D Human Pose Estimation
    Zhu, Xianglu
    Zhang, Zhang
    Wang, Wei
    Wang, Zilei
    Wang, Liang
    2024 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN 2024, 2024,
  • [40] Towards Robust Models of Code via Energy-Based Learning on Auxiliary Datasets
    Bui, Nghi D. Q.
    Yu, Yijun
    PROCEEDINGS OF THE 37TH IEEE/ACM INTERNATIONAL CONFERENCE ON AUTOMATED SOFTWARE ENGINEERING, ASE 2022, 2022,