Deep Image Inpainting With Enhanced Normalization and Contextual Attention

被引:13
|
作者
Liu, Jialu [1 ]
Gong, Maoguo [1 ]
Tang, Zedong [1 ]
Qin, A. K. [2 ]
Li, Hao [1 ]
Jiang, Fenlong [1 ]
机构
[1] Xidian Univ, Key Lab Intelligent Percept & Image Understanding, Minist Educ, Sch Elect Engn, Xian 710071, Shaanxi, Peoples R China
[2] Swinburne Univ Technol, Dept Comp Technol, Hawthorn, Vic 3122, Australia
基金
中国国家自然科学基金;
关键词
Task analysis; Convolution; Correlation; Context modeling; Adaptation models; Semantics; Feature extraction; Image inpainting; normalization; mean and variance shifts; contextual attention; NETWORK;
D O I
10.1109/TCSVT.2022.3175171
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Deep learning-based image inpainting has been widely studied, leading to great success. However, many methods adopt convolution and normalization operations, which will bring up some issues to affect the performance. The vanilla normalization cannot distinguish the pixels in corrupted regions from the other valid pixels, resulting in the mean and variance shifts. In addition, the limited receptive field of convolution makes it unable to capture long-range valid information directly. In order to tackle these challenges, we propose a novel deep generative model for image inpainting with two key modules, namely, the channel and spatially adaptive batch normalization (CSA-BN) module, and the selective latent-space-mapping-based contextual attention (SLSM-CA) layer. We replace the vanilla normalization with the CSA-BN module. By channel and spatially adaptive denormalization, the CSA-BN module can mitigate the spatial mean and variance shifts in each channel in a targeted way. In addition, we also integrate the SLSM-CA layer into our model to capture the long-range correlations explicitly. By introducing dual-branch attention and a feature selection module, the SLSM-CA layer can selectively utilize the multi-scale background information to improve prediction quality. What's more, it introduces the latent spaces to achieve the low-rank approximations of attention matrices and to reduce computational costs. Extensive quantitative and qualitative evaluations demonstrate the superiority of the proposed method compared with state-of-the-art methods.
引用
收藏
页码:6599 / 6614
页数:16
相关论文
共 50 条
  • [31] Contextual-Based Image Inpainting: Infer, Match, and Translate
    Song, Yuhang
    Yang, Chao
    Lin, Zhe
    Liu, Xiaofeng
    Huang, Qin
    Li, Hao
    Kuo, C-C Jay
    [J]. COMPUTER VISION - ECCV 2018, PT II, 2018, 11206 : 3 - 18
  • [32] Position and Channel Attention for Image Inpainting by Semantic Structure
    Qiu, Jingjun
    Gao, Yan
    [J]. 2020 IEEE 32ND INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE (ICTAI), 2020, : 1290 - 1295
  • [33] Multi-scale attention network for image inpainting
    Qin, Jia
    Bai, Huihui
    Zhao, Yao
    [J]. COMPUTER VISION AND IMAGE UNDERSTANDING, 2021, 204
  • [34] Research on Image Inpainting Algorithms Based on Attention Guidance
    Shen, Yankun
    Su, Yaya
    Wang, Lin
    Jia, Dongli
    [J]. JOURNAL OF ADVANCED COMPUTATIONAL INTELLIGENCE AND INTELLIGENT INFORMATICS, 2023, 27 (02) : 190 - 197
  • [35] PCNet: partial convolution attention mechanism for image inpainting
    Yan S.
    Zhang X.
    [J]. International Journal of Computers and Applications, 2022, 44 (08) : 738 - 745
  • [36] One-Stage Image Inpainting with Hybrid Attention
    Zhao, Lulu
    Shen, Ling
    Hong, Richang
    [J]. MULTIMEDIA MODELING (MMM 2022), PT I, 2022, 13141 : 505 - 517
  • [37] Image Inpainting With Learnable Edge-Attention Maps
    Sun, Liujie
    Zhang, Qinghan
    Wang, Wenju
    Zhang, Mingxi
    [J]. IEEE ACCESS, 2021, 9 : 3816 - 3827
  • [38] Sparse self-attention transformer for image inpainting
    Huang, Wenli
    Deng, Ye
    Hui, Siqi
    Wu, Yang
    Zhou, Sanping
    Wang, Jinjun
    [J]. PATTERN RECOGNITION, 2024, 145
  • [39] Deep Image Inpainting to Support Endoscopic Procedures
    Menegatti, Danilo
    Betello, Filippo
    Delli Priscoli, Francesco
    Giuseppi, Alessandro
    [J]. 2023 31ST MEDITERRANEAN CONFERENCE ON CONTROL AND AUTOMATION, MED, 2023, : 507 - 512
  • [40] A transformer–CNN for deep image inpainting forensics
    Xinshan Zhu
    Junyan Lu
    Honghao Ren
    Hongquan Wang
    Biao Sun
    [J]. The Visual Computer, 2023, 39 : 4721 - 4735