Diagonal Attention and Style-based GAN for Content-Style Disentanglement in Image Generation and Translation

被引:5
|
作者
Kwon, Gihyun [1 ]
Ye, Jong Chul [1 ,2 ]
机构
[1] Korea Adv Inst Sci & Technol, Dept Bio & Brain Engn, Daejeon, South Korea
[2] Korea Adv Inst Sci & Technol, Grad Sch AI, Daejeon, South Korea
基金
新加坡国家研究基金会;
关键词
D O I
10.1109/ICCV48922.2021.01372
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One of the important research topics in image generative models is to disentangle the spatial contents and styles for their separate control. Although StyleGAN can generate content feature vectors from random noises, the resulting spatial content control is primarily intended for minor spatial variations, and the disentanglement of global content and styles is by no means complete. Inspired by a mathematical understanding of normalization and attention, here we present a novel hierarchical adaptive Diagonal spatial ATtention (DAT) layers to separately manipulate the spatial contents from styles in a hierarchical manner. Using DAT and AdaIN, our method enables coarse-to-fine level disentanglement of spatial contents and styles. In addition, our generator can be easily integrated into the GAN inversion framework so that the content and style of translated images from multi-domain image translation tasks can be flexibly controlled. By using various datasets, we confirm that the proposed method not only outperforms the existing models in disentanglement scores, but also provides more flexible control over spatial features in the generated images.
引用
收藏
页码:13960 / 13969
页数:10
相关论文
共 50 条
  • [1] Unsupervised underwater image enhancement via content-style representation disentanglement
    Zhu, Pengli
    Liu, Yancheng
    Wen, Yuanquan
    Xu, Minyi
    Fu, Xianping
    Liu, Siyuan
    [J]. ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 126
  • [2] GOYA: Leveraging Generative Art for Content-Style Disentanglement
    Wu, Yankun
    Nakashima, Yuta
    Garcia, Noa
    [J]. JOURNAL OF IMAGING, 2024, 10 (07)
  • [3] SFGAN: Style-Former GAN for Disentanglement in Image Generation
    [J]. 2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [4] Density-Aware Haze Image Synthesis by Self-Supervised Content-Style Disentanglement
    Zhang, Chi
    Lin, Zihang
    Xu, Liheng
    Li, Zongliang
    Tang, Wei
    Liu, Yuehu
    Meng, Gaofeng
    Wang, Le
    Li, Li
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (07) : 4552 - 4572
  • [5] Reproducibility Companion Paper: Stable Diffusion for Content-Style Disentanglement in Art Analysis
    Wu, Yankun
    Nakashima, Yuta
    Garcia, Noa
    Li, Sheng
    Zeng, Zhaoyang
    [J]. PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 1228 - 1231
  • [6] Remote Sensing Image Translation via Style-Based Recalibration Module and Improved Style Discriminator
    Zhang, Tiange
    Gao, Feng
    Dong, Junyu
    Du, Qian
    [J]. IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [7] Not Only Generative Art: Stable Diffusion for Content-Style Disentanglement in Art Analysis
    Wu, Yankun
    Nakashima, Yuta
    Garcia, Noa
    [J]. PROCEEDINGS OF THE 2023 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2023, 2023, : 199 - 208
  • [8] On the generation of realistic synthetic petrographic datasets using a style-based GAN
    Ivan Ferreira
    Luis Ochoa
    Ardiansyah Koeshidayatullah
    [J]. Scientific Reports, 12
  • [9] On the generation of realistic synthetic petrographic datasets using a style-based GAN
    Ferreira, Ivan
    Ochoa, Luis
    Koeshidayatullah, Ardiansyah
    [J]. SCIENTIFIC REPORTS, 2022, 12 (01)
  • [10] Image-to-image Translation via Hierarchical Style Disentanglement
    Li, Xinyang
    Zhang, Shengchuan
    Hu, Jie
    Cao, Liujuan
    Hong, Xiaopeng
    Mao, Xudong
    Huang, Feiyue
    Wu, Yongjian
    Ji, Rongrong
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 8635 - 8644