Dual Attention with the Self-Attention Alignment for Efficient Video Super-resolution

被引:0
|
作者
Yuezhong Chu
Yunan Qiao
Heng Liu
Jungong Han
机构
[1] Anhui University of Technology,School of Computer Science and Technology
[2] Aberystwyth University,Department of Computer Science
来源
Cognitive Computation | 2022年 / 14卷
关键词
Video super-resolution; Dual attention; Self-attention alignment; FLOPS;
D O I
暂无
中图分类号
学科分类号
摘要
By selectively enhancing the features extracted from convolution networks, the attention mechanism has shown its effectiveness for low-level visual tasks, especially for image super-resolution (SR). However, due to the spatiotemporal continuity of video sequences, simply applying image attention to a video does not seem to obtain good SR results. At present, there is still a lack of suitable attention structure to achieve efficient video SR. In this work, building upon the dual attention, i.e., position attention and channel attention, we proposed deep dual attention, underpinned by self-attention alignment (DASAA), for video SR. Specifically, we start by constructing a dual attention module (DAM) to strengthen the acquired spatiotemporal features and adopt a self-attention structure with the morphological mask to achieve attention alignment. Then, on top of the attention features, we utilize the up-sampling operation to reconstruct the super-resolved video images and introduce the LSTM (long short-time memory) network to guarantee the coherent consistency of the generated video frames both temporally and spatially. Experimental results and comparisons on the actual Youku-VESR dataset and the typical benchmark dataset-Vimeo-90 k demonstrate that our proposed approach achieves the best video SR effect while taking the least amount of computation. Specifically, in the Youku-VESR dataset, our proposed approach achieves a test PSNR of 35.290db and a SSIM of 0.939, respectively. In the Vimeo-90 k dataset, the PSNR/SSIM indexes of our approach are 32.878db and 0.774. Moreover, the FLOPS (float-point operations per second) of our approach is as low as 6.39G. The proposed DASAA method surpasses all video SR algorithms in the comparison. It is also revealed that there is no linear relationship between positional attention and channel attention. It suggests that our DASAA with LSTM coherent consistency architecture may have great potential for many low-level vision video applications.
引用
收藏
页码:1140 / 1151
页数:11
相关论文
共 50 条
  • [21] FNSAM: Image super-resolution using a feedback network with self-attention mechanism
    Huang, Yu
    Wang, Wenqian
    Li, Min
    TECHNOLOGY AND HEALTH CARE, 2023, 31 : S383 - S395
  • [22] Super-resolution of Pneumocystis carinii pneumonia CT via self-attention GAN
    Xie, Hongqiang
    Zhang, Tongtong
    Song, Weiwei
    Wang, Shoujun
    Zhu, Hongchang
    Zhang, Rumin
    Zhang, Weiping
    Yu, Yong
    Zhao, Yan
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2021, 212
  • [23] Learning Attention from Attention: Efficient Self-Refinement Transformer for Face Super-Resolution
    Li, Guanxin
    Shi, Jingang
    Zong, Yuan
    Wang, Fei
    Wang, Tian
    Gong, Yihong
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 1035 - 1043
  • [24] Image super-resolution reconstruction network based on expectation maximization self-attention residual
    Huang S.
    Hu H.
    Yang Y.
    Wan W.
    Wu Z.
    Beijing Hangkong Hangtian Daxue Xuebao/Journal of Beijing University of Aeronautics and Astronautics, 2024, 50 (02): : 388 - 397
  • [25] Wavelet Attention Embedding Networks for Video Super-Resolution
    Choi, Young-Ju
    Lee, Young-Woon
    Kim, Byung-Gyu
    2020 25TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2021, : 7314 - 7320
  • [26] RDNet: Lightweight Residual and Detail self-attention Network for infrared image super-resolution
    Chen, Feiyang
    Huang, Detian
    Lin, Mingxin
    Song, Jiaxun
    Huang, Xiaoqian
    INFRARED PHYSICS & TECHNOLOGY, 2024, 141
  • [27] Self-attention negative feedback network for real-time image super-resolution
    Liu, Xiangbin
    Chen, Shuqi
    Song, Liping
    Wozniak, Marcin
    Liu, Shuai
    JOURNAL OF KING SAUD UNIVERSITY-COMPUTER AND INFORMATION SCIENCES, 2022, 34 (08) : 6179 - 6186
  • [28] Degradation-Aware Self-Attention Based Transformer for Blind Image Super-Resolution
    Liu, Qingguo
    Gao, Pan
    Han, Kang
    Liu, Ningzhong
    Xiang, Wei
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 7516 - 7528
  • [29] Optimal Deep Multi-Route Self-Attention for Single Image Super-Resolution
    Ngambenjavichaikul, Nisawan
    Chen, Sovann
    Aramvith, Supavadee
    PROCEEDINGS OF 2022 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2022, : 1181 - 1186
  • [30] Single-Image Super-Resolution based on a Self-Attention Deep Neural Network
    Jiang, Linfu
    Zhong, Minzhi
    Qiu, Fangchi
    2020 13TH INTERNATIONAL CONGRESS ON IMAGE AND SIGNAL PROCESSING, BIOMEDICAL ENGINEERING AND INFORMATICS (CISP-BMEI 2020), 2020, : 387 - 391