PolicyGAN: Training generative adversarial networks using policy gradient

被引:0
|
作者
Paria, Biswajit [1 ]
Lahiri, Avisek [2 ]
Biswas, Prabir Kumar [2 ]
机构
[1] IIT, Dept CSE, Kharagpur, W Bengal, India
[2] IIT, Dept E&ECE, Kharagpur, W Bengal, India
关键词
Generative Adversarial Networks; Reinforcement Learning; Policy Gradient; Inception Score; Adversarial Learning;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper presents PolicyGAN, a policy gradient paradigm for training Generative Adversarial Networks that views the generator as an image generation neural agent which is rewarded by another neural agent, termed as the discriminator. Rewards are higher for samples near the original data manifold. In PolicyGAN, only reward signal from the output of the discriminator is used for updating the generator network using policy gradient. This obviates the need for gradient signal to flow through the discriminator for training the generator; an intrinsic property of original GAN formulation. Given the inherent difficulty of training adversarial models, and low convergence speed of policy gradient, training GANs using policy gradient is a non-trivial problem and requires deep study. Currently GANs have used only differentiable discriminators for training. Policy-GAN germinates the possibility of using a wide variety of non-differentiable discriminator networks for training GANs, something which was not possible with the original GAN framework. Another advantage of using policy gradient is that now the generator need not produce deterministic samples, but can generate a probability distribution from which samples can be taken. PolicyGAN thus paves the path to use a variety of probabilistic models.
引用
收藏
页码:151 / 156
页数:6
相关论文
共 50 条
  • [31] ECG Generation With Sequence Generative Adversarial Nets Optimized by Policy Gradient
    Ye, Fei
    Zhu, Fei
    Fu, Yuchen
    Shen, Bairong
    [J]. IEEE ACCESS, 2019, 7 : 159369 - 159378
  • [32] Generative Adversarial Inverse Reinforcement Learning With Deep Deterministic Policy Gradient
    Zhan, Ming
    Fan, Jingjing
    Guo, Jianying
    [J]. IEEE ACCESS, 2023, 11 : 87732 - 87746
  • [33] Handwriting Profiling Using Generative Adversarial Networks
    Ghosh, Arna
    Bhattacharya, Biswarup
    Chowdhury, Somnath Basu Roy
    [J]. THIRTY-FIRST AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2017, : 4927 - 4928
  • [34] Phylogenetic inference using generative adversarial networks
    Smith, Megan L.
    Hahn, Matthew W.
    [J]. BIOINFORMATICS, 2023, 39 (09)
  • [35] Training data independent image registration using generative adversarial networks and domain adaptation
    Mahapatra, Dwarikanath
    Ge, Zongyuan
    [J]. PATTERN RECOGNITION, 2020, 100
  • [36] Enhanced cell segmentation with limited training datasets using cycle generative adversarial networks
    Zargari, Abolfazl
    Topacio, Benjamin R.
    Mashhadi, Najmeh
    Shariati, S. Ali
    [J]. ISCIENCE, 2024, 27 (05)
  • [37] Image Inpainting Using Generative Adversarial Networks
    Luo H.-L.
    Ao Y.
    Yuan P.
    [J]. Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2020, 48 (10): : 1891 - 1898
  • [38] VOICE IMPERSONATION USING GENERATIVE ADVERSARIAL NETWORKS
    Gao, Yang
    Singh, Rita
    Raj, Bhiksha
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 2506 - 2510
  • [39] Training generative adversarial networks for optical property mapping using synthetic image data
    Osman, A.
    Crowley, J.
    Gordon, G. S. D.
    [J]. BIOMEDICAL OPTICS EXPRESS, 2022, 13 (10) : 5171 - 5186
  • [40] Generative Adversarial Networks Using Adaptive Convolution
    Nguyen, Nhat M.
    Ray, Nilanjan
    [J]. 2019 16TH CONFERENCE ON COMPUTER AND ROBOT VISION (CRV 2019), 2019, : 129 - 134