Optimal Neural Network Approximation of Wasserstein Gradient Direction via Convex Optimization\ast

被引:0
|
作者
Wang, Yifei [1 ]
Chen, Peng [2 ]
Pilanci, Mert [2 ]
Li, Wuchen [3 ]
机构
[1] Stanford Univ, Dept Elect Engn, Stanford, CA 94305 USA
[2] Georgia Inst Technol, Coll Comp, Sch Computat Sci & Engn, Atlanta, GA 30332 USA
[3] Univ South Carolina, Dept Math, Columbia, SC 29208 USA
来源
关键词
Key words. Bayesian inference; convex optimization; neural network; semipositive definite program; AUGMENTED LAGRANGIAN METHOD; INVERSE PROBLEMS; EQUATIONS;
D O I
10.1137/23M1573173
中图分类号
O29 [应用数学];
学科分类号
070104 ;
摘要
The calculation of the direction of the Wasserstein gradient is vital for addressing problems related to posterior sampling and scientific computing. To approximate the Wasserstein gradient using finite samples, it is necessary to solve a variation problem. Our study focuses on the variation problem within the framework of two-layer networks with squared ReLU activations. We present a semidefinite program (SDP) relaxation as a solution, which can be viewed as an approximation of the Wasserstein gradient for a broader range of functions, including two-layer networks. By solving the convex SDP, we achieve the best approximation of the Wasserstein gradient direction in this function class. We also provide conditions to ensure the relaxation is tight. Additionally, we propose methods for practical implementation, such as subsampling and dimension reduction. The effectiveness and efficiency of our proposed method are demonstrated through numerical experiments, including Bayesian inference with PDE constraints and parameter estimation in COVID-19 modeling.
引用
收藏
页码:978 / 999
页数:22
相关论文
共 50 条
  • [31] Optimal Location of Cellular Base Station via Convex Optimization
    Kalantari, Elham
    Loyka, Sergey
    Yanikomeroglu, Halim
    Yongacoglu, Abbas
    2020 IEEE INTERNATIONAL BLACK SEA CONFERENCE ON COMMUNICATIONS AND NETWORKING (BLACKSEACOM), 2020,
  • [32] Optimal, worst case filter design via convex optimization
    Sun, KP
    Packard, A
    42ND IEEE CONFERENCE ON DECISION AND CONTROL, VOLS 1-6, PROCEEDINGS, 2003, : 1380 - 1385
  • [33] Neural approximation of Wasserstein distance via a universal architecture for symmetric and factorwise group invariant functions
    Chen, Samantha
    Wang, Yusu
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 36 (NEURIPS 2023), 2023,
  • [34] Convex optimization via finite-time projected gradient flows
    Chen, Fei
    Ren, Wei
    2018 IEEE CONFERENCE ON DECISION AND CONTROL (CDC), 2018, : 4072 - 4077
  • [35] A FAST DUAL GRADIENT METHOD FOR SEPARABLE CONVEX OPTIMIZATION VIA SMOOTHING
    Li, Jueyou
    Wu, Zhiyou
    Wu, Changzhi
    Long, Qiang
    Wang, Xiangyu
    Lee, Jae-Myung
    Jung, Kwang-Hyo
    PACIFIC JOURNAL OF OPTIMIZATION, 2016, 12 (02): : 289 - +
  • [36] Gradient-Based Algorithms for Convex Discrete Optimization via Simulation
    Zhang, Haixiang
    Zheng, Zeyu
    Lavaei, Javad
    OPERATIONS RESEARCH, 2023, 71 (05) : 1815 - 1834
  • [37] Manifold Neural Network With Non-Gradient Optimization
    Zhang, Rui
    Jiao, Ziheng
    Zhang, Hongyuan
    Li, Xuelong
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (03) : 3986 - 3993
  • [38] Neural network for nonsmooth pseudoconvex optimization with general convex constraints
    Bian, Wei
    Ma, Litao
    Qin, Sitian
    Xue, Xiaoping
    NEURAL NETWORKS, 2018, 101 : 1 - 14
  • [39] Quasi-Lagrangian Neural Network for Convex Quadratic Optimization
    Costantini, Giovanni
    Perfetti, Renzo
    Todisco, Massimiliano
    IEEE TRANSACTIONS ON NEURAL NETWORKS, 2008, 19 (10): : 1804 - 1809
  • [40] A delayed neural network method for solving convex optimization problems
    Yang, Yongqing
    Cao, Jinde
    INTERNATIONAL JOURNAL OF NEURAL SYSTEMS, 2006, 16 (04) : 295 - 303