Learning a spatial-temporal symmetry network for video super-resolution

被引:0
|
作者
Wang, Xiaohang [1 ,2 ]
Liu, Mingliang [1 ,2 ]
Wei, Pengying [1 ,2 ]
机构
[1] Heilongjiang Univ, Dept Automat, Harbin 150080, Heilongjiang, Peoples R China
[2] Heilongjiang Univ, Key Lab Informat Fus Estimat & Detect, Harbin 150080, Heilongjiang, Peoples R China
关键词
Video super-resolution; Motion estimation; Spatial-temporal symmetry; Convolutional neural network; CONVOLUTION;
D O I
10.1007/s10489-022-03603-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The video super-resolution (VSR) method is designed to estimate and restore high-resolution (HR) sequences from low-resolution (LR) input. For the past few years, many VSR methods with machine learning have been proposed that combine both the convolutional neural network (CNN) and motion compensation. Most mainstream approaches are based on optical flow or deformation convolution, and both need accurate estimates for motion compensation. However, most previous methods have not been able to fully utilize the spatial-temporal symmetrical information from input sequences. Moreover, much computation is consumed by aligning every neighbouring frame to the reference frame separately. Furthermore, many methods reconstruct HR results on only a single scale, which limits the reconstruction accuracy of the network and its performance in complex scenes. In this study, we propose a spatial-temporal symmetry network (STSN) to solve the above deficiencies. STSN includes four parts: prefusion, alignment, postfusion and reconstruction. First, a two-stage fusion strategy is applied to reduce the computation consumption of the network. Furthermore, ConvGRU is utilized in the prefusion module, the redundant features between neighbouring frames are eliminated, and several neighbouring frames are fused and condensed into two parts. To generate accurate offset maps, we present a spatial-temporal symmetry attention block (STSAB). This component exploits the symmetry of spatial-temporal combined spatial attention. In the reconstruction module, we propose an SR multiscale residual block (SR-MSRB) to enhance reconstruction performance. Abundant experimental results that test several datasets show that our method possesses better effects and efficiency in both quantitative and qualitative measurement indices compared with state-of-the-art methods.
引用
收藏
页码:3530 / 3544
页数:15
相关论文
共 50 条
  • [31] Spatial-Temporal Super-Resolution of Satellite Imagery via Conditional Pixel Synthesis
    He, Yutong
    Wang, Dingjie
    Lai, Nicholas
    Zhang, William
    Meng, Chenlin
    Burke, Marshall
    Lobell, David B.
    Ermon, Stefano
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021,
  • [32] Residual Invertible Spatio-Temporal Network for Video Super-Resolution
    Zhu, Xiaobin
    Li, Zhuangzi
    Zhang, Xiao-Yu
    Li, Changsheng
    Liu, Yaqi
    Xue, Ziyu
    [J]. THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 5981 - 5988
  • [33] Video super-resolution based on a spatio-temporal matching network
    Zhu, Xiaobin
    Li, Zhuangzi
    Lou, Jungang
    Shen, Qing
    [J]. PATTERN RECOGNITION, 2021, 110
  • [34] Video Super-Resolution via a Spatio-Temporal Alignment Network
    Wen, Weilei
    Ren, Wenqi
    Shi, Yinghuan
    Nie, Yunfeng
    Zhang, Jingang
    Cao, Xiaochun
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 1761 - 1773
  • [35] Fast Spatio-Temporal Residual Network for Video Super-Resolution
    Li, Sheng
    He, Fengxiang
    Du, Bo
    Zhang, Lefei
    Xu, Yonghao
    Tao, Dacheng
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 10514 - 10523
  • [36] Grouped Spatio-Temporal Alignment Network for Video Super-Resolution
    Lu, Mingxuan
    Zhang, Peng
    [J]. IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 2193 - 2197
  • [37] Temporal Consistency Learning of Inter-Frames for Video Super-Resolution
    Liu, Meiqin
    Jin, Shuo
    Yao, Chao
    Lin, Chunyu
    Zhao, Yao
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (04) : 1507 - 1520
  • [38] An Efficient and Lightweight Structure for Spatial-Temporal Feature Extraction in Video Super Resolution
    He, Xiaonan
    Xia, Yukun
    Qiao, Yuansong
    Lee, Brian
    Ye, Yuhang
    [J]. ADVANCES IN COMPUTER GRAPHICS, CGI 2023, PT I, 2024, 14495 : 362 - 374
  • [39] Learning a spatial-temporal texture transformer network for video inpainting
    Ma, Pengsen
    Xue, Tao
    [J]. FRONTIERS IN NEUROROBOTICS, 2022, 16
  • [40] Spatial and temporal super-resolution for fluorescence microscopy by a recurrent neural network
    Li, Jinyang
    Tong, Geng
    Pan, Yining
    Yu, Yiting
    [J]. OPTICS EXPRESS, 2021, 29 (10) : 15747 - 15763