ARDA-UNIT recurrent dense self-attention block with adaptive feature fusion for unpaired (unsupervised) image-to-image translation

被引:0
|
作者
Ghombavani, Farzane Maghsoudi [1 ]
Fadaeieslam, Mohammad Javad [1 ]
Yaghmaee, Farzin [1 ]
机构
[1] Semnan Univ, Dept Elect & Comp Engn, Semnan, Iran
关键词
computer vision; image processing;
D O I
10.1049/ipr2.12894
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One of the most challenging topics in artificial intelligence is image-to-image translation, the purpose of which is generating images close to those in the target domain while preserving the important features of the images in the source domain. In this direction, various types of generative adversarial networks have been developed. ARDA-UNIT, presented in this paper, seeks to meet the main challenges of these networks, that is, producing a high-quality image in a reasonable amount of time, and transferring content between two images with different structures. The proposed recurrent dense self-attention block, applied in ARDA-UNIT's generator latent space, simultaneously increases its generating capability and decreases the training parameters. ARDA-UNIT has a feature extraction module which feeds both the generator and the discriminator. This module uses a new adaptive feature fusion method which combines multi-scale features in such a way that the characteristics of each scale are preserved. The module also uses a pre-trained CNN that reduces the training parameters. Moreover, a feature similarity loss is introduced that guides the model to change the structure of the source domain in accordance with that in the target domain. Experiments performed on different datasets using FID, KID and IS evaluation criteria have shown that the model reduces computational loads, transfers structures well, and achieves better qualities.
引用
收藏
页码:3746 / 3758
页数:13
相关论文
共 21 条
  • [1] Unsupervised Image-to-Image Translation with Self-Attention Networks
    Kang, Taewon
    Lee, Kwang Hee
    [J]. 2020 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING (BIGCOMP 2020), 2020, : 102 - 108
  • [2] AFF-UNIT: Adaptive feature fusion for unsupervised image-to-image translation
    Li, Yuqiang
    Meng, Haochen
    Lin, Hong
    Liu, Chun
    [J]. IET IMAGE PROCESSING, 2021, 15 (13) : 3172 - 3188
  • [3] Unsupervised image-to-image translation by semantics consistency and self-attention
    Zhang Zhibin
    Xue Wanli
    Fu Guokai
    [J]. OPTOELECTRONICS LETTERS, 2022, 18 (03) : 175 - 180
  • [4] Unsupervised image-to-image translation by semantics consistency and self-attention
    Zhibin Zhang
    Wanli Xue
    Guokai Fu
    [J]. Optoelectronics Letters, 2022, 18 : 175 - 180
  • [5] Unsupervised image-to-image translation by semantics consistency and self-attention
    ZHANG Zhibin
    XUE Wanli
    FU Guokai
    [J]. Optoelectronics Letters, 2022, 18 (03) : 175 - 180
  • [6] Self-attention StarGAN for Multi-domain Image-to-Image Translation
    He, Ziliang
    Yang, Zhenguo
    Mao, Xudong
    Lv, Jianming
    Li, Qing
    Liu, Wenyin
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2019: IMAGE PROCESSING, PT III, 2019, 11729 : 537 - 549
  • [7] Improving Disentanglement-Based Image-to-Image Translation with Feature Joint Block Fusion
    Zhang, Zhejian
    Zhang, Rui
    Wang, Qiu-Feng
    Huang, Kaizhu
    [J]. ADVANCES IN BRAIN INSPIRED COGNITIVE SYSTEMS, 2020, 11691 : 540 - 549
  • [8] Unsupervised Infrared and Visible Image Fusion with Pixel Self-attention
    Cui, Saijia
    Zhou, Zhiqiang
    Li, Linhao
    Fei, Erfang
    [J]. PROCEEDINGS OF THE 33RD CHINESE CONTROL AND DECISION CONFERENCE (CCDC 2021), 2021, : 437 - 441
  • [9] SAFE: Unsupervised image feature extraction using self-attention based feature extraction network
    Choi, Yeoung Je
    Lee, Gyeong Taek
    Kim, Chang Ouk
    [J]. EXPERT SYSTEMS, 2024, 41 (08)
  • [10] Wide Self-attention Mechanism Fusion Dense Residual Network Image Dehazing
    Wu, Kaijun
    Ding, Yuan
    [J]. Hunan Daxue Xuebao/Journal of Hunan University Natural Sciences, 2023, 50 (08): : 13 - 22