Towards Robust Gan-Generated Image Detection: A Multi-View Completion Representation

被引:0
|
作者
Liu, Chi [1 ]
Zhu, Tianqing [1 ]
Shen, Sheng [2 ]
Zhou, Wanlei [3 ]
机构
[1] Univ Technol Sydney, Sch Comp Sci, Ultimo, Australia
[2] Univ Sydney, Sch Elect & Informat Engn, Camperdown, Australia
[3] City Univ Macau, Macau, Peoples R China
基金
澳大利亚研究理事会;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
GAN-generated image detection now becomes the first line of defense against the malicious uses of machine-synthesized image manipulations such as deepfakes. Although some existing detectors work well in detecting clean, known GAN samples, their success is largely attributable to overfitting unstable features such as frequency artifacts, which will cause failures when facing unknown GANs or perturbation attacks. To overcome the issue, we propose a robust detection framework based on a novel multi-view image completion representation. The framework first learns various view-to-image tasks to model the diverse distributions of genuine images. Frequency-irrelevant features can be represented from the distributional discrepancies characterized by the completion models, which are stable, generalized, and robust for detecting unknown fake patterns. Then, a multi-view classification is devised with elaborated intra- and inter-view learning strategies to enhance view-specific feature representation and cross-view feature aggregation, respectively. We evaluated the generalization ability of our framework across six popular GANs at different resolutions and its robustness against a broad range of perturbation attacks. The results confirm our method's improved effectiveness, generalization, and robustness over various baselines.
引用
收藏
页码:464 / 472
页数:9
相关论文
共 50 条
  • [31] Transferability of CNN models for GAN-generated face detection
    Aieprasert, Thanapat
    Mahdlang, Yada
    Pansiri, Chadaya
    Sae-Bae, Napa
    Khomkham, Banphatree
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (33) : 79815 - 79831
  • [32] Multi-view Low Rank Representation for Multi-Source Traffic Data Completion
    Rong Du
    Shudong Chen
    International Journal of Intelligent Transportation Systems Research, 2019, 17 : 200 - 211
  • [33] Multi-view representation learning for multi-view action recognition
    Hao, Tong
    Wu, Dan
    Wang, Qian
    Sun, Jin-Sheng
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2017, 48 : 453 - 460
  • [34] An Architecture for the detection of GAN-generated Flood Images with Localization Capabilities
    Wang, Jun
    Alamayreh, Omran
    Tondi, Benedetta
    Barni, Mauro
    2022 IEEE 14TH IMAGE, VIDEO, AND MULTIDIMENSIONAL SIGNAL PROCESSING WORKSHOP (IVMSP), 2022,
  • [35] CrossScore: Towards Multi-View Image Evaluation and Scoring
    Wang, Zirui
    Bian, Wenjing
    Prisacariu, Victor Adrian
    COMPUTER VISION - ECCV 2024, PT IX, 2025, 15067 : 492 - 510
  • [36] Manifold Regularized Multi-View Subspace Clustering for Image Representation
    Wang, Lei
    Li, Danping
    He, Tiancheng
    Xue, Zhong
    2016 23RD INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2016, : 283 - 288
  • [37] Image classification base on PCA of multi-view deep representation
    Sun, Yaoqi
    Li, Liang
    Zheng, Liang
    Hu, Ji
    Li, Wenchao
    Jiang, Yatong
    Yan, Chenggang
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2019, 62 : 253 - 258
  • [38] Image automatic annotation via multi-view deep representation
    Yang, Yang
    Zhang, Wensheng
    Xie, Yuan
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2015, 33 : 368 - 377
  • [39] A Competition of Shape and Texture Bias by Multi-view Image Representation
    Kong, Lingwei
    Wang, Jianzong
    Huang, Zhangcheng
    Xiao, Jing
    PATTERN RECOGNITION AND COMPUTER VISION, PT IV, 2021, 13022 : 140 - 151
  • [40] Multimodal Transformer With Multi-View Visual Representation for Image Captioning
    Yu, Jun
    Li, Jing
    Yu, Zhou
    Huang, Qingming
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (12) : 4467 - 4480