An Amalgamated CNN-Transformer Network for Lightweight Image Super-Resolution

被引:0
|
作者
Fang, Jinsheng [1 ,2 ]
Lin, Hanjiang [1 ,2 ]
Zeng, Kun [3 ]
机构
[1] School of Computer Science and Engineering Minnan Normal University, Zhangzhou,363000, China
[2] Key Laboratory of Data Science and Intelligence Application Fujian Province University, Fujian, Zhangzhou,363000, China
[3] School of Computer and Control engineering Fujian Provincial Key Laboratory of Information Processing and Intelligent Control Minjiang University, Fuzhou,350108, China
来源
Journal of Network Intelligence | 2024年 / 9卷 / 03期
关键词
HTTP;
D O I
暂无
中图分类号
学科分类号
摘要
Recently, Transformer-based methods for single image super-resolution (SISR) have achieved better performance advantages than the methods based on convolutional neural network (CNN). Exploiting self-attention mechanism to model global context definitely improves the SR results. However, the neglect of local information will bring inevitable reduction of the network performance. In this work, we propose an Amalgamated CNN-Transformer network for lightweight SR, namely ACTSR. Specifically, an amalgamated CNN-Transformer block (ACTB) is developed to extract the useful information of both local and global features. By employing stacked ACTBs, our ACTSR extracts more informative deep features beneficially for super-resolution reconstruction to improve network performance while keeps lightweight and flexible enough. Extensive experiments on commonly used benchmark datasets validate our ACTSR outperforms the advanced competitors. Our codes are available at: https://github.com/ginsengf/ACTSR. © 2024, Taiwan Ubiquitous Information CO LTD. All rights reserved.
引用
收藏
页码:1376 / 1387
相关论文
共 50 条
  • [31] Lightweight image super-resolution network using involution
    Liang, Jiu
    Zhang, Yu
    Xue, Jiangbo
    Hu, Yanda
    MACHINE VISION AND APPLICATIONS, 2022, 33 (05)
  • [32] Lightweight Image Super-Resolution with ConvNeXt Residual Network
    Yong Zhang
    Haomou Bai
    Yaxing Bing
    Xiao Liang
    Neural Processing Letters, 2023, 55 : 9545 - 9561
  • [33] An efficient lightweight network for single image super-resolution*
    Tang, Yinggan
    Zhang, Xiang
    Zhang, Xuguang
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2023, 93
  • [34] Lightweight subpixel sampling network for image super-resolution
    Hongfei Zeng
    Qiang Wu
    Jin Zhang
    Haojie Xia
    The Visual Computer, 2024, 40 : 3781 - 3793
  • [35] Lightweight Parallel Feedback Network for Image Super-Resolution
    Beibei Wang
    Changjun Liu
    Binyu Yan
    Xiaomin Yang
    Neural Processing Letters, 2023, 55 : 3225 - 3243
  • [36] A sparse lightweight attention network for image super-resolution
    Hongao Zhang
    Jinsheng Fang
    Siyu Hu
    Kun Zeng
    The Visual Computer, 2024, 40 (2) : 1261 - 1272
  • [37] Lightweight Parallel Feedback Network for Image Super-Resolution
    Wang, Beibei
    Liu, Changjun
    Yan, Binyu
    Yang, Xiaomin
    NEURAL PROCESSING LETTERS, 2023, 55 (03) : 3225 - 3243
  • [38] A sparse lightweight attention network for image super-resolution
    Zhang, Hongao
    Fang, Jinsheng
    Hu, Siyu
    Zeng, Kun
    VISUAL COMPUTER, 2024, 40 (02): : 1261 - 1272
  • [39] Learning Texture Transformer Network for Image Super-Resolution
    Yang, Fuzhi
    Yang, Huan
    Fu, Jianlong
    Lu, Hongtao
    Guo, Baining
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 5790 - 5799
  • [40] Single Image Super-Resolution Using Lightweight CNN with Maxout Units
    Choi, Jae-Seok
    Kim, Munchurl
    COMPUTER VISION - ACCV 2018, PT VI, 2019, 11366 : 471 - 487