Dual Attention with the Self-Attention Alignment for Efficient Video Super-resolution

被引:0
|
作者
Yuezhong Chu
Yunan Qiao
Heng Liu
Jungong Han
机构
[1] Anhui University of Technology,School of Computer Science and Technology
[2] Aberystwyth University,Department of Computer Science
来源
Cognitive Computation | 2022年 / 14卷
关键词
Video super-resolution; Dual attention; Self-attention alignment; FLOPS;
D O I
暂无
中图分类号
学科分类号
摘要
By selectively enhancing the features extracted from convolution networks, the attention mechanism has shown its effectiveness for low-level visual tasks, especially for image super-resolution (SR). However, due to the spatiotemporal continuity of video sequences, simply applying image attention to a video does not seem to obtain good SR results. At present, there is still a lack of suitable attention structure to achieve efficient video SR. In this work, building upon the dual attention, i.e., position attention and channel attention, we proposed deep dual attention, underpinned by self-attention alignment (DASAA), for video SR. Specifically, we start by constructing a dual attention module (DAM) to strengthen the acquired spatiotemporal features and adopt a self-attention structure with the morphological mask to achieve attention alignment. Then, on top of the attention features, we utilize the up-sampling operation to reconstruct the super-resolved video images and introduce the LSTM (long short-time memory) network to guarantee the coherent consistency of the generated video frames both temporally and spatially. Experimental results and comparisons on the actual Youku-VESR dataset and the typical benchmark dataset-Vimeo-90 k demonstrate that our proposed approach achieves the best video SR effect while taking the least amount of computation. Specifically, in the Youku-VESR dataset, our proposed approach achieves a test PSNR of 35.290db and a SSIM of 0.939, respectively. In the Vimeo-90 k dataset, the PSNR/SSIM indexes of our approach are 32.878db and 0.774. Moreover, the FLOPS (float-point operations per second) of our approach is as low as 6.39G. The proposed DASAA method surpasses all video SR algorithms in the comparison. It is also revealed that there is no linear relationship between positional attention and channel attention. It suggests that our DASAA with LSTM coherent consistency architecture may have great potential for many low-level vision video applications.
引用
收藏
页码:1140 / 1151
页数:11
相关论文
共 50 条
  • [1] Dual Attention with the Self-Attention Alignment for Efficient Video Super-resolution
    Chu, Yuezhong
    Qiao, Yunan
    Liu, Heng
    Han, Jungong
    COGNITIVE COMPUTATION, 2022, 14 (03) : 1140 - 1151
  • [2] SELF-ATTENTION FOR AUDIO SUPER-RESOLUTION
    Rakotonirina, Nathanael Carraz
    2021 IEEE 31ST INTERNATIONAL WORKSHOP ON MACHINE LEARNING FOR SIGNAL PROCESSING (MLSP), 2021,
  • [3] Dual Self-Attention Swin Transformer for Hyperspectral Image Super-Resolution
    Long, Yaqian
    Wang, Xun
    Xu, Meng
    Zhang, Shuyu
    Jiang, Shuguo
    Jia, Sen
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [4] Self-attention learning network for face super-resolution
    Zeng, Kangli
    Wang, Zhongyuan
    Lu, Tao
    Chen, Jianyu
    Wang, Jiaming
    Xiong, Zixiang
    NEURAL NETWORKS, 2023, 160 : 164 - 174
  • [5] Multi-feature self-attention super-resolution network
    Yang, Aiping
    Wei, Zihao
    Wang, Jinbin
    Cao, Jiale
    Ji, Zhong
    Pang, Yanwei
    VISUAL COMPUTER, 2024, 40 (05): : 3473 - 3486
  • [6] Multi-feature self-attention super-resolution network
    Aiping Yang
    Zihao Wei
    Jinbin Wang
    Jiale Cao
    Zhong Ji
    Yanwei Pang
    The Visual Computer, 2024, 40 : 3473 - 3486
  • [7] SRFormer: Permuted Self-Attention for Single Image Super-Resolution
    Zhou, Yupeng
    Li, Zhen
    Guo, Chun-Le
    Bai, Song
    Cheng, Ming-Ming
    Hou, Qibin
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 12734 - 12745
  • [8] Efficient Dual Attention Transformer for Image Super-Resolution
    Park, Soobin
    Jeong, Yuna
    Choi, Yong Suk
    39TH ANNUAL ACM SYMPOSIUM ON APPLIED COMPUTING, SAC 2024, 2024, : 963 - 970
  • [9] Image super-resolution reconstruction based on self-attention GAN
    Wang X.-S.
    Chao J.
    Cheng Y.-H.
    Kongzhi yu Juece/Control and Decision, 2021, 36 (06): : 1324 - 1332
  • [10] Learning a Deep Dual Attention Network for Video Super-Resolution
    Li, Feng
    Bai, Huihui
    Zhao, Yao
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 (29) : 4474 - 4488