UVCGAN: UNet Vision Transformer cycle-consistent GAN for unpaired image-to-image translation

被引:76
|
作者
Torbunov, Dmitrii [1 ]
Huang, Yi [1 ]
Yu, Haiwang [1 ]
Huang, Jin [1 ]
Yoo, Shinjae [1 ]
Lin, Meifeng [1 ]
Viren, Brett [1 ]
Ren, Yihui [1 ]
机构
[1] Brookhaven Natl Lab, Upton, NY 11973 USA
关键词
D O I
10.1109/WACV56688.2023.00077
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unpaired image-to-image translation has broad applications in art, design, and scientific simulations. One early breakthrough was CycleGAN that emphasizes one-to-one mappings between two unpaired image domains via generative-adversarial networks (GAN) coupled with the cycle-consistency constraint, while more recent works promote one-to-many mapping to boost diversity of the translated images. Motivated by scientific simulation and one-to-one needs, this work revisits the classic CycleGAN framework and boosts its performance to outperform more contemporary models without relaxing the cycle-consistency constraint. To achieve this, we equip the generator with a Vision Transformer (ViT) and employ necessary training and regularization techniques. Compared to previous best-performing models, our model performs better and retains a strong correlation between the original and translated image. An accompanying ablation study shows that both the gradient penalty and self-supervised pre-training are crucial to the improvement. To promote reproducibility and open science, the source code, hyperparameter configurations, and pre-trained model are available at https: //github.com/LS4GAN/uvcgan.
引用
收藏
页码:702 / 712
页数:11
相关论文
共 50 条
  • [31] UMGAN: Underwater Image Enhancement Network for Unpaired Image-to-Image Translation
    Sun, Boyang
    Mei, Yupeng
    Yan, Ni
    Chen, Yingyi
    JOURNAL OF MARINE SCIENCE AND ENGINEERING, 2023, 11 (02)
  • [32] IMAGE DATA AUGMENTATION WITH UNPAIRED IMAGE-TO-IMAGE CAMERA MODEL TRANSLATION
    Foo, Chi Fa
    Winkler, Stefan
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 3246 - 3250
  • [33] Thermal Image Calibration and Correction using Unpaired Cycle-Consistent Adversarial Networks
    Rajoli, Hossein
    Afshin, Pouya
    Afghah, Fatemeh
    FIFTY-SEVENTH ASILOMAR CONFERENCE ON SIGNALS, SYSTEMS & COMPUTERS, IEEECONF, 2023, : 1425 - 1429
  • [34] Holographic reconstruction enhancement via unpaired image-to-image translation
    Scherrer, Romane
    Selmaoui-folcher, Nazha
    Quiniou, Thomas
    Jauffrais, Thierry
    Lemonnier, Hugues
    Bonnet, Sophie
    APPLIED OPTICS, 2022, 61 (33) : 9807 - 9816
  • [35] UNPAIRED IMAGE-TO-IMAGE TRANSLATION FROM SHARED DEEP SPACE
    Wu, Xuehui
    Shao, Jie
    Gao, Lianli
    Shen, Heng Tao
    2018 25TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2018, : 2127 - 2131
  • [36] Augmented Cyclic Consistency Regularization for Unpaired Image-to-Image Translation
    Ohkawa, Takehiko
    Inoue, Naoto
    Kataoka, Hirokatsu
    Inoue, Nakamasa
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 362 - 369
  • [37] Homomorphic Latent Space Interpolation for Unpaired Image-To-Image Translation
    Chen, Ying-Cong
    Xu, Xiaogang
    Tian, Zhuotao
    Jia, Jiaya
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2403 - 2411
  • [38] Exploring Negatives in Contrastive Learning for Unpaired Image-to-Image Translation
    Lin, Yupei
    Zhang, Sen
    Chen, Tianshui
    Lu, Yongyi
    Li, Guangping
    Shi, Yukai
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 1186 - 1194
  • [39] UNPAIRED IMAGE-TO-IMAGE SHAPE TRANSLATION ACROSS FASHION DATA
    Wang, Kaili
    Ma, Liqian
    Oramas, Jose M.
    Van Gool, Luc
    Tuytelaars, Tinne
    2020 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2020, : 206 - 210
  • [40] Rethinking the Paradigm of Content Constraints in Unpaired Image-to-Image Translation
    Cai, Xiuding
    Zhu, Yaoyao
    Miao, Dong
    Fu, Linjie
    Yao, Yu
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 2, 2024, : 891 - 899