Contrastive learning for unsupervised image-to-image translation

被引:2
|
作者
Lee, Hanbit [1 ]
Seol, Jinseok [2 ]
Lee, Sang-goo [2 ]
Park, Jaehui [3 ]
Shim, Junho [4 ]
机构
[1] SK Telecom, AIX Ctr, Seongnam, South Korea
[2] Seoul Natl Univ, Dept Comp Sci & Engn, Seoul, South Korea
[3] Univ Seoul, Dept Stat, Seoul, South Korea
[4] Sookmyung Womens Univ, Dept Comp Sci, Seoul, South Korea
关键词
Image-to-image translation; Generative adversarial networks; Contrastive learning; Self-supervised learning; Style transfer;
D O I
10.1016/j.asoc.2023.111170
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Image -to-image translation (I2I) aims to learn a mapping function to transform images into different styles or domains while preserving their key structures. Typically, I2I models require manually defined image domains as a training set to learn the visual differences among the image domains and achieve the ability to translate images across them. However, constructing such multi-domain datasets on a large scale requires expensive data collection and annotation processes. Moreover, if the target domain changes or is expanded, a new dataset should be collected, and the model should be retrained. To address these challenges, this article presents a novel unsupervised I2I method that does not require manually defined image domains. The proposed method automatically learns the visual similarity between individual samples and leverages the learned similarity function to transfer a specific style or appearance across images. Therefore, the developed method does not rely on cost-intensive manual domains or unstable clustering results, leading to improved translation accuracy at minimal cost. For quantitative evaluation, we implemented a state -of -the -art I2I models and performed image transformation on the same input image using the baselines and our method. The image quality was then assessed using two quantitative metrics: Frechet inception distance (FID) and translation accuracy. The proposed method exhibited significant improvements in image quality and translation accuracy compared with the latest unsupervised I2I methods. Specifically, the developed technique achieved a 25% and 19% improvement over the best-performing unsupervised baseline in terms of FID and translation accuracy, respectively. Furthermore, this approach demonstrated performance nearly comparable to those of supervised learning-based methods trained using manually collected and constructed domains.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Dual Contrastive Learning for Unsupervised Image-to-Image Translation
    Han, Junlin
    Shoeiby, Mehrdad
    Petersson, Lars
    Armin, Mohammad Ali
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 746 - 755
  • [2] Truly Unsupervised Image-to-Image Translation with Contrastive Representation Learning
    Hong, Zhiwei
    Feng, Jianxing
    Jiang, Tao
    [J]. COMPUTER VISION - ACCV 2022, PT III, 2023, 13843 : 239 - 255
  • [3] DualGAN: Unsupervised Dual Learning for Image-to-Image Translation
    Yi, Zili
    Zhang, Hao
    Tan, Ping
    Gong, Minglun
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2017, : 2868 - 2876
  • [4] Improving Learning time in Unsupervised Image-to-Image Translation
    Min, Tae-Hong
    Kim, Do-Yun
    Choi, Young-June
    [J]. 2019 1ST INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE IN INFORMATION AND COMMUNICATION (ICAIIC 2019), 2019, : 455 - 458
  • [5] Exploring Negatives in Contrastive Learning for Unpaired Image-to-Image Translation
    Lin, Yupei
    Zhang, Sen
    Chen, Tianshui
    Lu, Yongyi
    Li, Guangping
    Shi, Yukai
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 1186 - 1194
  • [6] Multimodal Unsupervised Image-to-Image Translation
    Huang, Xun
    Liu, Ming-Yu
    Belongie, Serge
    Kautz, Jan
    [J]. COMPUTER VISION - ECCV 2018, PT III, 2018, 11207 : 179 - 196
  • [7] Unsupervised Image-to-Image Translation Networks
    Liu, Ming-Yu
    Breuel, Thomas
    Kautz, Jan
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 30 (NIPS 2017), 2017, 30
  • [8] Unsupervised Image-to-Image Translation: A Review
    Hoyez, Henri
    Schockaert, Cedric
    Rambach, Jason
    Mirbach, Bruno
    Stricker, Didier
    [J]. SENSORS, 2022, 22 (21)
  • [9] Perceptual Contrastive Generative Adversarial Network based on image warping for unsupervised image-to-image translation
    Huang, Lin-Chieh
    Tsai, Hung-Hsu
    [J]. NEURAL NETWORKS, 2023, 166 : 313 - 325
  • [10] Multi-feature contrastive learning for unpaired image-to-image translation
    Yao Gou
    Min Li
    Yu Song
    Yujie He
    Litao Wang
    [J]. Complex & Intelligent Systems, 2023, 9 : 4111 - 4122