Learning a spatial-temporal symmetry network for video super-resolution

被引:0
|
作者
Wang, Xiaohang [1 ,2 ]
Liu, Mingliang [1 ,2 ]
Wei, Pengying [1 ,2 ]
机构
[1] Heilongjiang Univ, Dept Automat, Harbin 150080, Heilongjiang, Peoples R China
[2] Heilongjiang Univ, Key Lab Informat Fus Estimat & Detect, Harbin 150080, Heilongjiang, Peoples R China
关键词
Video super-resolution; Motion estimation; Spatial-temporal symmetry; Convolutional neural network; CONVOLUTION;
D O I
10.1007/s10489-022-03603-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The video super-resolution (VSR) method is designed to estimate and restore high-resolution (HR) sequences from low-resolution (LR) input. For the past few years, many VSR methods with machine learning have been proposed that combine both the convolutional neural network (CNN) and motion compensation. Most mainstream approaches are based on optical flow or deformation convolution, and both need accurate estimates for motion compensation. However, most previous methods have not been able to fully utilize the spatial-temporal symmetrical information from input sequences. Moreover, much computation is consumed by aligning every neighbouring frame to the reference frame separately. Furthermore, many methods reconstruct HR results on only a single scale, which limits the reconstruction accuracy of the network and its performance in complex scenes. In this study, we propose a spatial-temporal symmetry network (STSN) to solve the above deficiencies. STSN includes four parts: prefusion, alignment, postfusion and reconstruction. First, a two-stage fusion strategy is applied to reduce the computation consumption of the network. Furthermore, ConvGRU is utilized in the prefusion module, the redundant features between neighbouring frames are eliminated, and several neighbouring frames are fused and condensed into two parts. To generate accurate offset maps, we present a spatial-temporal symmetry attention block (STSAB). This component exploits the symmetry of spatial-temporal combined spatial attention. In the reconstruction module, we propose an SR multiscale residual block (SR-MSRB) to enhance reconstruction performance. Abundant experimental results that test several datasets show that our method possesses better effects and efficiency in both quantitative and qualitative measurement indices compared with state-of-the-art methods.
引用
收藏
页码:3530 / 3544
页数:15
相关论文
共 50 条
  • [21] Spatial-Temporal Video Enhancement using Super-Resolution from a Multi-Camera System
    Quevedo, E.
    de la Cruz, J.
    Callico, G. M.
    Tobajas, F.
    Sarmiento, R.
    [J]. 2014 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS (ICCE), 2014, : 538 - 539
  • [22] Spatial-Temporal Motion Compensation Based Video Super Resolution
    An, Yaozu
    Lu, Yao
    Yan, Ziye
    [J]. COMPUTER VISION - ACCV 2010, PT II, 2011, 6493 : 282 - 292
  • [23] Spatial and temporal information as camera parameters for super-resolution video
    Tarvainen, Jussi
    Nuutinen, Mikko
    Oittinen, Pirkko
    [J]. 2012 IEEE INTERNATIONAL SYMPOSIUM ON MULTIMEDIA (ISM), 2012, : 302 - 305
  • [24] A video Coding System with Spatial-temporal down-/up-sampling and Super-Resolution Reconstruction
    He, Xiaohai
    Fu, Wenjie
    Lin, Hongwei
    Li, Xueqing
    Peng, Xin
    [J]. 2016 8TH INTERNATIONAL CONFERENCE ON COMPUTATIONAL INTELLIGENCE AND COMMUNICATION NETWORKS (CICN), 2016, : 236 - 239
  • [25] Attention-guided video super-resolution with recurrent multi-scale spatial-temporal transformer
    Sun, Wei
    Kong, Xianguang
    Zhang, Yanning
    [J]. COMPLEX & INTELLIGENT SYSTEMS, 2023, 9 (04) : 3989 - 4002
  • [26] Towards High-Quality and Efficient Video Super-Resolution via Spatial-Temporal Data Overfitting
    Li, Gen
    Ji, Jie
    Qin, Minghai
    Niu, Wei
    Ren, Bin
    Afghah, Fatemeh
    Guo, Linke
    Ma, Xiaolong
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 10259 - 10269
  • [27] Explore spatial-temporal relations: transient super-resolution with PMD sensors
    Han, Chaosheng
    Lin, Xing
    Lin, Jingyu
    Yan, Chenggang
    Dai, Qionghai
    [J]. OPTOELECTRONIC IMAGING AND MULTIMEDIA TECHNOLOGY III, 2014, 9273
  • [28] Spatio-Temporal Fusion Network for Video Super-Resolution
    Li, Huabin
    Zhang, Pingjian
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [29] Learning Temporal Dynamics for Video Super-Resolution: A Deep Learning Approach
    Liu, Ding
    Wang, Zhaowen
    Fan, Yuchen
    Liu, Xianming
    Wang, Zhangyang
    Chang, Shiyu
    Wang, Xinchao
    Huang, Thomas S.
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (07) : 3432 - 3445
  • [30] Learning a Deep Dual Attention Network for Video Super-Resolution
    Li, Feng
    Bai, Huihui
    Zhao, Yao
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 4474 - 4488