Image super-resolution reconstruction using Swin Transformer with efficient channel attention networks

被引:0
|
作者
Sun, Zhenxi [1 ,2 ]
Zhang, Jin [1 ,2 ,3 ]
Chen, Ziyi [1 ,2 ]
Hong, Lu [1 ,2 ]
Zhang, Rui [1 ,2 ]
Li, Weishi [1 ,2 ,3 ]
Xia, Haojie [1 ,2 ,3 ]
机构
[1] Hefei Univ Technol, Sch Instrument Sci & Optoelect Engn, Hefei 230009, Peoples R China
[2] Anhui Prov Key Lab Measuring Theory & Precis Instr, Hefei 230009, Peoples R China
[3] Minist Educ, Engn Res Ctr Safety Crit Ind Measurement & Control, Hefei 230009, Peoples R China
基金
国家重点研发计划;
关键词
Image super-resolution; Swin Transformer; Efficient channel attention; Multi-attention fusion;
D O I
10.1016/j.engappai.2024.108859
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Image super-resolution reconstruction (SR) is an important ill-posed problem in low-level vision, which aims to reconstruct high-resolution images from low-resolution images. Although current state-of-the-art methods exhibit impressive performance, their recovery of image detail information and edge information is still unsatisfactory. To address this problem, this paper proposes a shifted window Transformer (Swin Transformer) with an efficient channel attention network (S-ECAN), which combines the attention based on convolutional neural networks and the self-attention of the Swin Transformer to combine the advantages of both and focuses on learning high-frequency features of images. In addition, to solve the problem of Convolutional Neural Network (CNN) based channel attention consumes a large number of parameters to achieve good performance, this paper proposes the Efficient Channel Attention Block (ECAB), which only involves a handful of parameters while bringing clear performance gain. Extensive experimental validation shows that the proposed model can recover more high-frequency details and texture information. The model is validated on Set5, Set14, B100, Urban100, and Manga109 datasets, where it outperforms the state-of-the-art methods by 0.03-0.13 dB, 0.04- 0.09 dB, 0.01-0.06 dB, 0.13-0.20 dB, and 0.06-0.17 dB respectively in terms of objective metrics. Ultimately, the substantial performance gains and enhanced visual results over prior arts validate the effectiveness and competitiveness of our proposed approach, which achieves an improved performance-complexity trade-off.
引用
下载
收藏
页数:10
相关论文
共 50 条
  • [31] Edge-Aware Attention Transformer for Image Super-Resolution
    Wang, Haoqian
    Xing, Zhongyang
    Xu, Zhongjie
    Cheng, Xiangai
    Li, Teng
    IEEE Signal Processing Letters, 2024, 31 : 2905 - 2909
  • [32] Parallel attention recursive generalization transformer for image super-resolution
    Jing Wang
    Yuanyuan Hao
    Hongxing Bai
    Lingyu Yan
    Scientific Reports, 15 (1)
  • [33] Super-Resolution Image Reconstruction Based on Residual Channel Attention and Multilevel Feature Fusion
    Xi Zhihong
    Yuan Kunpeng
    LASER & OPTOELECTRONICS PROGRESS, 2020, 57 (04)
  • [34] An Iris Image Super-Resolution Model Based on Swin Transformer and Generative Adversarial Network
    Lu, Hexin
    Zhu, Xiaodong
    Cui, Jingwei
    Jiang, Haifeng
    ALGORITHMS, 2024, 17 (03)
  • [35] Hybrid recursive super-resolution image reconstruction using neural networks
    Zhang, Di
    Miao, Xuelan
    He, Jiazhong
    ICNC 2007: Third International Conference on Natural Computation, Vol 1, Proceedings, 2007, : 192 - 196
  • [36] Image denoising using channel attention residual enhanced Swin Transformer
    Dai, Qiang
    Cheng, Xi
    Zhang, Li
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (07) : 19041 - 19059
  • [37] Image denoising using channel attention residual enhanced Swin Transformer
    Qiang Dai
    Xi Cheng
    Li Zhang
    Multimedia Tools and Applications, 2024, 83 : 19041 - 19059
  • [38] Efficient image magnification and applications to super-resolution reconstruction
    Shao, Wenze
    Wei, Zhihui
    IEEE ICMA 2006: PROCEEDING OF THE 2006 IEEE INTERNATIONAL CONFERENCE ON MECHATRONICS AND AUTOMATION, VOLS 1-3, PROCEEDINGS, 2006, : 2372 - +
  • [39] HCT: image super-resolution restoration using hierarchical convolution transformer networks
    Ying Guo
    Chang Tian
    Han Wang
    Jie Liu
    Chong Di
    Keqing Ning
    Pattern Analysis and Applications, 2025, 28 (2)
  • [40] N-Gram in Swin Transformers for Efficient Lightweight Image Super-Resolution
    Choi, Haram
    Lee, Jeongmin
    Yang, Jihoon
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 2071 - 2081