Attack as the Best Defense: Nullifying Image-to-image Translation GANs via Limit-aware Adversarial Attack

被引:6
|
作者
Yeh, Chin-Yuan [1 ,3 ]
Chen, Hsi-Wen [1 ]
Shuai, Hong-Han [2 ]
Yang, De-Nian [3 ]
Chen, Ming-Syan [1 ]
机构
[1] Natl Taiwan Univ, Taipei, Taiwan
[2] Natl Yangming Jiaotong Univ, Taipei, Taiwan
[3] Acad Sinica, Taipei, Taiwan
关键词
D O I
10.1109/ICCV48922.2021.01588
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Due to the great success of image-to-image (Img2Img) translation GANs, many applications with ethics issues arise, e.g., DeepFake and DeepNude, presenting a challenging problem to prevent the misuse of these techniques. In this work, we tackle the problem by a new adversarial attack scheme, namely the Nullifying Attack, which cancels the image translation process and proposes a corresponding framework, the Limit-Aware Self-Guiding Gradient Sliding Attack (LaS-GSA) under a black-box setting. In other words, by processing the image with the proposed LaS-GSA before publishing, any image translation functions can be nullified, which prevents the images from malicious manipulations. First, we introduce the limit-aware RGF and the gradient sliding mechanism to estimate the gradient that adheres to the adversarial limit, i.e., the pixel value limitations of the adversarial example. We theoretically prove that our model is able to avoid the error caused by the projection in both the direction and the length. Then, an effective self-guiding prior is extracted solely from the threat model and the target image to efficiently leverage the prior information and guide the gradient estimation process. Extensive experiments demonstrate that LaS-GSA requires fewer queries to nullify the image translation process with higher success rates than 4 state-of-the-art methods.
引用
收藏
页码:16168 / 16177
页数:10
相关论文
共 34 条
  • [1] Generative Reversible Data Hiding by Image-to-Image Translation via GANs
    Zhang, Zhuo
    Fu, Guangyuan
    Di, Fuqiang
    Li, Changlong
    Liu, Jia
    SECURITY AND COMMUNICATION NETWORKS, 2019, 2019
  • [2] Adversarial Attack Defense Based on the Deep Image Prior Network
    Sutanto, Richard Evan
    Lee, Sukho
    INFORMATION SCIENCE AND APPLICATIONS, 2020, 621 : 519 - 526
  • [3] Output-correlated adversarial attack for image translation network
    Liu, Peiyuan
    Sun, Lei
    Mao, XiuQing
    JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (02)
  • [4] Artistic image adversarial attack via style perturbation
    Zhang, Haiyan
    Wang, Quan
    Feng, Guorui
    MULTIMEDIA SYSTEMS, 2023, 29 (06) : 3745 - 3755
  • [5] Universal Physical Adversarial Attack via Background Image
    Xu, Yidan
    Wang, Juan
    Li, Yuanzhang
    Wang, Yajie
    Xu, Zixuan
    Wang, Dianxin
    APPLIED CRYPTOGRAPHY AND NETWORK SECURITY WORKSHOPS, ACNS 2022, 2022, 13285 : 3 - 14
  • [6] Artistic image adversarial attack via style perturbation
    Haiyan Zhang
    Quan Wang
    Guorui Feng
    Multimedia Systems, 2023, 29 (6) : 3745 - 3755
  • [7] Search the Steganographic Policy for Image via Adversarial Attack
    Li L.
    Fan M.
    Hao J.
    Dianzi Keji Daxue Xuebao/Journal of the University of Electronic Science and Technology of China, 2022, 51 (02): : 259 - 263
  • [8] FPGA Adaptive Neural Network Quantization for Adversarial Image Attack Defense
    Lu, Yufeng
    Shi, Xiaokang
    Jiang, Jianan
    Deng, Hanhui
    Wang, Yanwen
    Lu, Jiwu
    Wu, Di
    IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2024, 20 (12) : 14017 - 14028
  • [9] Survey on Adversarial Attack and Defense for Medical Image Analysis: Methods and Challenges
    Dong, Junhao
    Chen, Junxi
    Xie, Xiaohua
    Lai, Jianhuang
    Chen, Hao
    ACM COMPUTING SURVEYS, 2025, 57 (03)
  • [10] RetouchUAA: Unconstrained Adversarial Attack via Realistic Image Retouching
    Xie, Mengda
    He, Yiling
    Qin, Zhan
    Fang, Meie
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (03) : 2586 - 2602