Learning the Stein Discrepancy for Training and Evaluating Energy-Based Models without Sampling

被引:0
|
作者
Grathwohl, Will [1 ,2 ]
Wang, Kuan-Chieh [1 ,2 ]
Jacobsen, Jorn-Henrik [1 ,2 ]
Duvenaud, David [1 ,2 ]
Zemel, Richard [1 ,2 ]
机构
[1] Univ Toronto, Toronto, ON, Canada
[2] Vector Inst, Toronto, ON, Canada
关键词
NETWORKS;
D O I
暂无
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
We present a new method for evaluating and training unnormalized density models. Our approach only requires access to the gradient of the unnormalized model's log-density. We estimate the Stein discrepancy between the data density p(x) and the model density q(x) defined by a vector function of the data. We parameterize this function with a neural network and fit its parameters to maximize the discrepancy. This yields a novel goodness-of-fit test which outperforms existing methods on high dimensional data. Furthermore, optimizing q(x) to minimize this discrepancy produces a novel method for training unnormalized models which scales more gracefully than existing methods. The ability to both learn and compare models is a unique feature of the proposed method.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Joint Energy-based Model Training for Better Calibrated Natural Language Understanding Models
    He, Tianxing
    McCann, Bryan
    Xiong, Caiming
    Hosseini-Asl, Ehsan
    16TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (EACL 2021), 2021, : 1754 - 1761
  • [42] Model-Based Planning with Energy-Based Models
    Du, Yilun
    Lin, Toru
    Mordatch, Igor
    CONFERENCE ON ROBOT LEARNING, VOL 100, 2019, 100
  • [43] AN EMPIRICAL COMPARISON OF JOINT-TRAINING AND PRE-TRAINING FOR DOMAIN-AGNOSTIC SEMI-SUPERVISED LEARNING VIA ENERGY-BASED MODELS
    Song, Yunfu
    Zheng, Huahuan
    Ou, Zhijian
    2021 IEEE 31ST INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2021,
  • [44] Energy-based models for sparse overcomplete representations
    Teh, YW
    Welling, M
    Osindero, S
    Hinton, GE
    JOURNAL OF MACHINE LEARNING RESEARCH, 2004, 4 (7-8) : 1235 - 1260
  • [45] Towards understanding retrosynthesis by energy-based models
    Sun, Ruoxi
    Dai, Hanjun
    Li, Li
    Kearnes, Steven
    Dai, Bo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [46] Rate -Distortion via Energy-Based Models
    Li, Qing
    Kim, Yongjune
    Guyot, Cyril
    2023 DATA COMPRESSION CONFERENCE, DCC, 2023, : 351 - 351
  • [47] Implicit Generation and Modeling with Energy-Based Models
    Du, Yilun
    Mordatch, Igor
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [48] LOCAL AND NONLOCAL ENERGY-BASED COUPLING MODELS
    Acosta, Gabriel
    Bersetche, Francisco
    Rossi, Julio D.
    SIAM JOURNAL ON MATHEMATICAL ANALYSIS, 2022, 54 (06) : 6288 - 6322
  • [49] Energy-Based Survival Models for Predictive Maintenance
    Holmer, Olov
    Frisk, Erik
    Krysander, Mattias
    IFAC PAPERSONLINE, 2023, 56 (02): : 10862 - 10867
  • [50] RNA pseudoknot prediction in energy-based models
    Lyngso, RB
    Pedersen, CNS
    JOURNAL OF COMPUTATIONAL BIOLOGY, 2000, 7 (3-4) : 409 - 427