Improving weight clipping in Wasserstein GANs

被引:1
|
作者
Massart, Estelle [1 ]
机构
[1] Catholic Univ Louvain, ICTEAM, Ave Georges Lemaitre 4,L4-05-01, B-1348 Louvain La Neuve, Belgium
关键词
D O I
10.1109/ICPR56361.2022.9956056
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Weight clipping is a well-known strategy to keep the Lipschitz constant of the critic under control, in Wasserstein GAN training. After each training iteration, all parameters of the critic are clipped to a given box, impacting the progress made by the optimizer. In this work, we propose a new strategy for weight clipping in Wasserstein GANs. Instead of directly clipping the parameters, we first obtain an equivalent model that is closer to the clipping box, and only then clip the parameters. Our motivation is to decrease the impact of the clipping strategy on the objective, at each iteration. This equivalent model is obtained by following invariant curves in the critic loss landscape, whose existence is a consequence of the positive homogeneity of common activations: resealing the input and output signals to each activation by inverse factors preserves the loss. We provide preliminary experiments showing that the proposed strategy speeds up training on Wasserstein GANs with simple feedforward architectures.
引用
收藏
页码:2286 / 2292
页数:7
相关论文
共 50 条
  • [1] Wasserstein Divergence for GANs
    Wu, Jiqing
    Huang, Zhiwu
    Thoma, Janine
    Acharya, Dinesh
    Van Gool, Luc
    COMPUTER VISION - ECCV 2018, PT V, 2018, 11209 : 673 - 688
  • [2] Wasserstein Proximal of GANs
    Lin, Alex Tong
    Li, Wuchen
    Osher, Stanley
    Montufar, Guido
    GEOMETRIC SCIENCE OF INFORMATION (GSI 2021), 2021, 12829 : 524 - 533
  • [3] Quantum Wasserstein GANs
    Chakrabarti, Shouvanik
    Huang, Yiming
    Li, Tongyang
    Feizi, Soheil
    Wu, Xiaodi
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [4] Improved Training of Wasserstein GANs
    Ishaan Gulrajani
    Ahmed, Faruk
    Arjovsky, Martin
    Dumoulin, Vincent
    Courville, Aaron
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [5] RELAXED WASSERSTEIN WITH APPLICATIONS TO GANS
    Guo, Xin
    Hong, Johnny
    Lin, Tianyi
    Yang, Nan
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 3325 - 3329
  • [6] A Wasserstein perspective of Vanilla GANs
    Kunkel, Lea
    Trabs, Mathias
    NEURAL NETWORKS, 2025, 181
  • [7] Some theoretical insights into wasserstein gans
    Biau, Gerard
    Sangnier, Maxime
    Tanielian, Ugo
    Journal of Machine Learning Research, 2021, 22
  • [8] Some Theoretical Insights into Wasserstein GANs
    Biau, Gerard
    Sangnier, Maxime
    Tanielian, Ugo
    JOURNAL OF MACHINE LEARNING RESEARCH, 2021, 22 : 1 - 45
  • [9] Towards Generalized Implementation of Wasserstein Distance in GANs
    Xu, Minkai
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10514 - 10522
  • [10] Visual Feature Attribution using Wasserstein GANs
    Baumgartner, Christian F.
    Koch, Lisa M.
    Tezcan, Kerem Can
    Ang, Jia Xi
    Konukoglu, Ender
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 8309 - 8319