CTNet: hybrid architecture based on CNN and transformer for image inpainting detection

被引:3
|
作者
Xiao, Fengjun [1 ]
Zhang, Zhuxi [2 ]
Yao, Ye [2 ]
机构
[1] Hangzhou Dianzi Univ, Zhejiang Informatizat Dev Inst, Xiasha Higher Educ Zone, Hangzhou 310018, Zhejiang, Peoples R China
[2] Hangzhou Dianzi Univ, Sch Cyberspace, Xiasha Higher Educ Zone, Hangzhou 310018, Zhejiang, Peoples R China
基金
中国国家自然科学基金;
关键词
Image inpainting detection; Deep neural network; Hybrid CNN-Transformer encoder; High-pass filter; DIFFUSION; NETWORK;
D O I
10.1007/s00530-023-01184-w
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Digital image inpainting technology has increasingly gained popularity as a result of the development of image processing and machine vision. However, digital image inpainting can be used not only to repair damaged photographs, but also to remove specific people or distort the semantic content of images. To address the issue of image inpainting forgeries, a hybrid CNN-Transformer Network (CTNet), which is composed of the hybrid CNN-Transformer encoder, the feature enhancement module, and the decoder module, is proposed for image inpainting detection and localization. Different from existing inpainting detection methods that rely on hand-crafted attention mechanisms, the hybrid CNN-Transformer encoder employs CNN as a feature extractor to build feature maps tokenized as the input patches of the Transformer encoder. The hybrid structure exploits the innate global self-attention mechanisms of Transformer and can effectively capture the long-term dependency of the image. Since inpainting traces mainly exist in the high-frequency components of digital images, the feature enhancement module performs feature extraction in the frequency domain. The decoder regularizes the upsampling process of the predicted masks with the assistance of high-frequency features. We investigate the generalization capacity of our CTNet on datasets generated by ten commonly used inpainting methods. The experimental results show that the proposed model can detect a variety of unknown inpainting operations after being trained on the datasets generated by a single inpainting method.
引用
收藏
页码:3819 / 3832
页数:14
相关论文
共 50 条
  • [41] HCformer: Hybrid CNN-Transformer for LDCT Image Denoising
    Jinli Yuan
    Feng Zhou
    Zhitao Guo
    Xiaozeng Li
    Hengyong Yu
    Journal of Digital Imaging, 2023, 36 (5) : 2290 - 2305
  • [42] TransInpaint: Transformer-based Image Inpainting with Context Adaptation
    Shamsolmoali, Pourya
    Zareapoor, Masoumeh
    Granger, Eric
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS, ICCVW, 2023, : 849 - 858
  • [43] Transformer with Convolution for Irregular Image Inpainting
    Xie, Donglin
    Wang, Lingfeng
    2022 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, COMPUTER VISION AND MACHINE LEARNING (ICICML), 2022, : 35 - 38
  • [44] Continuously Masked Transformer for Image Inpainting
    Ko, Keunsoo
    Kim, Chang-Su
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 13123 - 13132
  • [45] Transformer architecture based on mutual attention for image-anomaly detection
    Mengting ZHANG
    Xiuxia TIAN
    虚拟现实与智能硬件(中英文), 2023, 5 (01) : 57 - 67
  • [46] Architecture Search for Image Inpainting
    Li, Yaoman
    King, Irwin
    ADVANCES IN NEURAL NETWORKS - ISNN 2019, PT I, 2019, 11554 : 106 - 115
  • [47] Land Cover Classification of UAV Remote Sensing Based on Transformer-CNN Hybrid Architecture
    Lu, Tingyu
    Wan, Luhe
    Qi, Shaoqun
    Gao, Meixiang
    SENSORS, 2023, 23 (11)
  • [48] FOTCA: hybrid transformer-CNN architecture using AFNO for accurate plant leaf disease image recognition
    Hu, Bo
    Jiang, Wenqian
    Zeng, Juan
    Cheng, Chen
    He, Laichang
    FRONTIERS IN PLANT SCIENCE, 2023, 14
  • [49] TransU2-Net: A Hybrid Transformer Architecture for Image Splicing Forgery Detection
    Yan, Caiping
    Li, Shuyuan
    Li, Hong
    IEEE ACCESS, 2023, 11 : 33313 - 33323
  • [50] A novel hybrid transformer-CNN architecture for environmental microorganism classification
    Shao, Ran
    Bi, Xiao-Jun
    Chen, Zheng
    PLOS ONE, 2022, 17 (11):