Optimal Neural Network Approximation of Wasserstein Gradient Direction via Convex Optimization\ast

被引:0
|
作者
Wang, Yifei [1 ]
Chen, Peng [2 ]
Pilanci, Mert [2 ]
Li, Wuchen [3 ]
机构
[1] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[2] Georgia Inst Technol, Coll Comp, Sch Computat Sci & Engn, Atlanta, GA 30332 USA
[3] Univ South Carolina, Dept Math, Columbia, SC 29208 USA
来源
关键词
Key words. Bayesian inference; convex optimization; neural network; semipositive definite program; AUGMENTED LAGRANGIAN METHOD; INVERSE PROBLEMS; EQUATIONS;
D O I
10.1137/23M1573173
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
The calculation of the direction of the Wasserstein gradient is vital for addressing problems related to posterior sampling and scientific computing. To approximate the Wasserstein gradient using finite samples, it is necessary to solve a variation problem. Our study focuses on the variation problem within the framework of two-layer networks with squared ReLU activations. We present a semidefinite program (SDP) relaxation as a solution, which can be viewed as an approximation of the Wasserstein gradient for a broader range of functions, including two-layer networks. By solving the convex SDP, we achieve the best approximation of the Wasserstein gradient direction in this function class. We also provide conditions to ensure the relaxation is tight. Additionally, we propose methods for practical implementation, such as subsampling and dimension reduction. The effectiveness and efficiency of our proposed method are demonstrated through numerical experiments, including Bayesian inference with PDE constraints and parameter estimation in COVID-19 modeling.
引用
收藏
页码:978 / 999
页数:22
相关论文
共 50 条
  • [1] A neural network for convex optimization
    Krasopoulos, Panagiotis T.
    Maratos, Nicholas G.
    2006 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOLS 1-11, PROCEEDINGS, 2006, : 747 - +
  • [2] Adversarial Attack Against Convolutional Neural Network via Gradient Approximation
    Wang, Zehao
    Li, Xiaoran
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT VI, ICIC 2024, 2024, 14867 : 221 - 232
  • [3] Optimal Energy Shaping via Neural Approximators\ast
    Massaroli, Stefano
    Poli, Michael
    Califano, Federico
    Park, Jinkyoo
    Yamashita, Atsushi
    Asama, Hajime
    SIAM JOURNAL ON APPLIED DYNAMICAL SYSTEMS, 2022, 21 (03): : 2126 - 2147
  • [4] GEOMETRIC DUALITY RESULTS AND APPROXIMATION ALGORITHMS FOR CONVEX VECTOR OPTIMIZATION PROBLEMS\ast
    Ararat, Cagin
    Tekgul, Imay
    Ulus, Firdevs
    SIAM JOURNAL ON OPTIMIZATION, 2023, 33 (01) : 116 - 146
  • [5] The GroupMax Neural Network Approximation of Convex Functions
    Warin, Xavier
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 11608 - 11612
  • [6] Safety Filter Design for Neural Network Systems via Convex Optimization
    Chen, Shaoru
    Chee, Kong Yao
    Matni, Nikolai
    Hsieh, M. Ani
    Pappas, George J.
    2023 62ND IEEE CONFERENCE ON DECISION AND CONTROL, CDC, 2023, : 6356 - 6363
  • [7] Scalable Computations of Wasserstein Barycenter via Input Convex Neural Networks
    Fan, Jiaojiao
    Taghvaei, Amirhossein
    Chen, Yongxin
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 139, 2021, 139
  • [8] Comparative study on gradient and Hessian estimation by Kriging and neural network approximation for optimization
    Sakata, S.
    Ashida, F.
    Zako, M.
    Advances in Computational Methods in Sciences and Engineering 2005, Vols 4 A & 4 B, 2005, 4A-4B : 509 - 512
  • [9] Global optimization via neural network approximation of inverse coordinate mappings
    Koshur V.D.
    Pushkaryov K.V.
    Optical Memory and Neural Networks, 2011, 20 (3) : 181 - 193
  • [10] A smooth gradient approximation neural network for general constrained nonsmooth nonconvex optimization problems
    Liu, Na
    Jia, Wenwen
    Qin, Sitian
    NEURAL NETWORKS, 2025, 184