Optimal Neural Network Approximation of Wasserstein Gradient Direction via Convex Optimization\ast

被引:0
|
作者
Wang, Yifei [1 ]
Chen, Peng [2 ]
Pilanci, Mert [2 ]
Li, Wuchen [3 ]
机构
[1] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[2] Georgia Inst Technol, Coll Comp, Sch Computat Sci & Engn, Atlanta, GA 30332 USA
[3] Univ South Carolina, Dept Math, Columbia, SC 29208 USA
来源
关键词
Key words. Bayesian inference; convex optimization; neural network; semipositive definite program; AUGMENTED LAGRANGIAN METHOD; INVERSE PROBLEMS; EQUATIONS;
D O I
10.1137/23M1573173
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
The calculation of the direction of the Wasserstein gradient is vital for addressing problems related to posterior sampling and scientific computing. To approximate the Wasserstein gradient using finite samples, it is necessary to solve a variation problem. Our study focuses on the variation problem within the framework of two-layer networks with squared ReLU activations. We present a semidefinite program (SDP) relaxation as a solution, which can be viewed as an approximation of the Wasserstein gradient for a broader range of functions, including two-layer networks. By solving the convex SDP, we achieve the best approximation of the Wasserstein gradient direction in this function class. We also provide conditions to ensure the relaxation is tight. Additionally, we propose methods for practical implementation, such as subsampling and dimension reduction. The effectiveness and efficiency of our proposed method are demonstrated through numerical experiments, including Bayesian inference with PDE constraints and parameter estimation in COVID-19 modeling.
引用
收藏
页码:978 / 999
页数:22
相关论文
共 50 条
  • [21] A Neural Network for Constrained Fuzzy Convex Optimization Problems
    Liu, Na
    Zhang, Han
    Qin, Sitian
    2019 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI 2019), 2019, : 1007 - 1012
  • [22] A TRANSLATION ROTATION INVARIANT NEURAL NETWORK TRAINED VIA CONJUGATE-GRADIENT OPTIMIZATION
    REED, KA
    HELFERTY, JJ
    IEEE INTERNATIONAL CONFERENCE ON SYSTEMS ENGINEERING ///, 1989, : 161 - 164
  • [23] Solving Convex Multi-Objective Optimization Problems via a Capable Neural Network Scheme
    Jahangiri, Mohammadreza
    Nazemi, Alireza
    INTERNATIONAL JOURNAL OF COMPUTATIONAL INTELLIGENCE AND APPLICATIONS, 2024, 23 (04)
  • [24] Neural Network for Change Direction Prediction in Dynamic Optimization
    Liu, Xiao-Fang
    Zhan, Zhi-Hui
    Zhang, Jun
    IEEE ACCESS, 2018, 6 : 72649 - 72662
  • [25] Designing a neural network via gradient flow
    Del Buono, N
    COMPUTATIONAL INTELLIGENCE AND APPLICATIONS, 2002, : 69 - 74
  • [26] Multi-Objective Optimization via Wasserstein-Fisher-Rao Gradient Flow
    Ren, Yinuo
    Xiao, Tesi
    Gangwani, Tanmay
    Rangi, Anshuka
    Rahmanian, Holakou
    Ying, Lexing
    Sanyal, Subhajit
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 238, 2024, 238
  • [27] PARALLEL AND DISTRIBUTED TRAINING OF NEURAL NETWORKS VIA SUCCESSIVE CONVEX APPROXIMATION
    Di Lorenzo, Paolo
    Scardapane, Simone
    2016 IEEE 26TH INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2016,
  • [28] Approximation of feasible sets in energy system applications via convex optimization
    Sari, Andrija T.
    Stankovic, Aleksandar M.
    2006 IEEE/PES POWER SYSTEMS CONFERENCE AND EXPOSITION. VOLS 1-5, 2006, : 1619 - +
  • [29] Wasserstein generative adversarial network with gradient penalty and convolutional neural network based motor imagery EEG classification
    Xiong, Hui
    Li, Jiahe
    Liu, Jinzhen
    Song, Jinlong
    Han, Yuqing
    JOURNAL OF NEURAL ENGINEERING, 2024, 21 (04)
  • [30] Optimal, worst case filter design via convex optimization
    Sun, KP
    Packard, A
    CONTROL OF UNCERTAIN SYSTEMS: MODELLING, APPROXIMATION, AND DESIGN, 2006, 329 : 293 - 315