PolicyGAN: Training generative adversarial networks using policy gradient

被引:0
|
作者
Paria, Biswajit [1 ]
Lahiri, Avisek [2 ]
Biswas, Prabir Kumar [2 ]
机构
[1] IIT, Dept CSE, Kharagpur, W Bengal, India
[2] IIT, Dept E&ECE, Kharagpur, W Bengal, India
关键词
Generative Adversarial Networks; Reinforcement Learning; Policy Gradient; Inception Score; Adversarial Learning;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents PolicyGAN, a policy gradient paradigm for training Generative Adversarial Networks that views the generator as an image generation neural agent which is rewarded by another neural agent, termed as the discriminator. Rewards are higher for samples near the original data manifold. In PolicyGAN, only reward signal from the output of the discriminator is used for updating the generator network using policy gradient. This obviates the need for gradient signal to flow through the discriminator for training the generator; an intrinsic property of original GAN formulation. Given the inherent difficulty of training adversarial models, and low convergence speed of policy gradient, training GANs using policy gradient is a non-trivial problem and requires deep study. Currently GANs have used only differentiable discriminators for training. Policy-GAN germinates the possibility of using a wide variety of non-differentiable discriminator networks for training GANs, something which was not possible with the original GAN framework. Another advantage of using policy gradient is that now the generator need not produce deterministic samples, but can generate a probability distribution from which samples can be taken. PolicyGAN thus paves the path to use a variety of probabilistic models.
引用
收藏
页码:151 / 156
页数:6
相关论文
共 50 条
  • [1] Training Generative Adversarial Networks with Adaptive Composite Gradient
    Qi, Huiqing
    Li, Fang
    Tan, Shengli
    Zhang, Xiangyun
    [J]. DATA INTELLIGENCE, 2024, 6 (01) : 120 - 157
  • [2] Gradient Normalization for Generative Adversarial Networks
    Wu, Yi-Lun
    Shuai, Hong-Han
    Tam, Zhi-Rui
    Chiu, Hong-Yu
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 6353 - 6362
  • [3] Exploring generative adversarial networks and adversarial training
    Sajeeda A.
    Hossain B.M.M.
    [J]. Int. J. Cogn. Comp. Eng., (78-89): : 78 - 89
  • [4] TRAINING GENERATIVE ADVERSARIAL NETWORKS WITH WEIGHTS
    Pantazis, Yannis
    Paul, Dipjyoti
    Fasoulakis, Michail
    Stylianou, Yannis
    [J]. 2019 27TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO), 2019,
  • [5] Improved Training of Generative Adversarial Networks Using Decision Forests
    Zuo, Yan
    Avraham, Gil
    Drummond, Tom
    [J]. 2021 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION WACV 2021, 2021, : 3491 - 3500
  • [6] Improved Training of Generative Adversarial Networks using Representative Features
    Bang, Duhyeon
    Shim, Hyunjung
    [J]. INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 80, 2018, 80
  • [7] Generative Adversarial Network with Policy Gradient for Text Summarization
    Rekabdar, Banafsheh
    Mousas, Christos
    Gupta, Bidyut
    [J]. 2019 13TH IEEE INTERNATIONAL CONFERENCE ON SEMANTIC COMPUTING (ICSC), 2019, : 204 - 207
  • [8] SeqGAN: Sequence Generative Adversarial Nets with Policy Gradient
    Yu, Lantao
    Zhang, Weinan
    Wang, Jun
    Yu, Yong
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 2852 - 2858
  • [9] Training generative adversarial networks by auxiliary adversarial example regulator
    Gan, Yan
    Ye, Mao
    Liu, Dan
    Liu, Yiguang
    [J]. APPLIED SOFT COMPUTING, 2023, 136
  • [10] Multiobjective coevolutionary training of Generative Adversarial Networks
    Ripa, Guillermo
    Mautone, Agustin
    Vidal, Andres
    Nesmachnow, Sergio
    Toutouh, Jamal
    [J]. PROCEEDINGS OF THE 2023 GENETIC AND EVOLUTIONARY COMPUTATION CONFERENCE COMPANION, GECCO 2023 COMPANION, 2023, : 319 - 322