Temporal Consistency Learning of Inter-Frames for Video Super-Resolution

被引:20
|
作者
Liu, Meiqin [1 ,2 ]
Jin, Shuo [1 ,2 ]
Yao, Chao [3 ]
Lin, Chunyu [1 ,2 ]
Zhao, Yao [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Inst Informat Sci, Beijing 100044, Peoples R China
[2] Beijing Jiaotong Univ, Beijing Key Lab Adv Informat Sci & Network Technol, Beijing 100044, Peoples R China
[3] Univ Sci & Technol Beijing, Sch Comp & Commun Engn, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
Superresolution; Circuit stability; Optical flow; Image restoration; Image reconstruction; Degradation; Convolution; Bidirectional motion estimation; temporal consistency; self-alignment; video super-resolution; FUSION NETWORK;
D O I
10.1109/TCSVT.2022.3214538
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Video super-resolution (VSR) is a task that aims to reconstruct high-resolution (HR) frames from the low-resolution (LR) reference frame and multiple neighboring frames. The vital operation is to utilize the relative misaligned frames for the current frame reconstruction and preserve the consistency of the results. Existing methods generally explore information propagation and frame alignment to improve the performance of VSR. However, few studies focus on the temporal consistency of inter-frames. In this paper, we propose a Temporal Consistency learning Network (TCNet) for VSR in an end-to-end manner, to enhance the consistency of the reconstructed videos. A spatio-temporal stability module is designed to learn the self-alignment from inter-frames. Especially, the correlative matching is employed to exploit the spatial dependency from each frame to maintain structural stability. Moreover, a self-attention mechanism is utilized to learn the temporal correspondence to implement an adaptive warping operation for temporal consistency among multi-frames. Besides, a hybrid recurrent architecture is designed to leverage short-term and long-term information. We further present a progressive fusion module to perform a multistage fusion of spatio-temporal features. And the final reconstructed frames are refined by these fused features. Objective and subjective results of various experiments demonstrate that TCNet has superior performance on different benchmark datasets, compared to several state-of-the-art methods.
引用
收藏
页码:1507 / 1520
页数:14
相关论文
共 50 条
  • [31] Video super-resolution reconstruction based on correlation learning and spatio-temporal nonlocal similarity
    Liang, Meiyu
    Du, Junping
    Li, Linghui
    MULTIMEDIA TOOLS AND APPLICATIONS, 2016, 75 (17) : 10241 - 10269
  • [32] A Framework for Super-Resolution of Scalable Video via Sparse Reconstruction of Residual Frames
    Moghaddam, Mohammad Hossein
    Azizipour, Mohammad Javad
    Vahidian, Saeed
    Smida, Besma
    MILCOM 2017 - 2017 IEEE MILITARY COMMUNICATIONS CONFERENCE (MILCOM), 2017, : 164 - 168
  • [33] Video Super-Resolution Using Codebooks Derived From Key-Frames
    Hung, Edson Mintsu
    de Queiroz, Ricardo L.
    Brandi, Fernanda
    de Oliveira, Karen Franca
    Mukherjee, Debargha
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2012, 22 (09) : 1321 - 1331
  • [34] Omniscient Video Super-Resolution
    Yi, Peng
    Wang, Zhongyuan
    Jiang, Kui
    Jiang, Junjun
    Lu, Tao
    Tian, Xin
    Ma, Jiayi
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 4409 - 4418
  • [35] Learning Super-Resolution Reconstruction for High Temporal Resolution Spike Stream
    Xiang, Xijie
    Zhu, Lin
    Li, Jianing
    Wang, Yixuan
    Huang, Tiejun
    Tian, Yonghong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (01) : 16 - 29
  • [36] Learning for Unconstrained Space-Time Video Super-Resolution
    Shi, Zhihao
    Liu, Xiaohong
    Li, Chengqi
    Dai, Linhui
    Chen, Jun
    Davidson, Timothy N.
    Zhao, Jiying
    IEEE TRANSACTIONS ON BROADCASTING, 2022, 68 (02) : 345 - 358
  • [37] Video super-resolution based on deep learning: a comprehensive survey
    Liu, Hongying
    Ruan, Zhubo
    Zhao, Peng
    Dong, Chao
    Shang, Fanhua
    Liu, Yuanyuan
    Yang, Linlin
    Timofte, Radu
    ARTIFICIAL INTELLIGENCE REVIEW, 2022, 55 (08) : 5981 - 6035
  • [38] Learning Trajectory-Aware Transformer for Video Super-Resolution
    Liu, Chengxu
    Yang, Huan
    Fu, Jianlong
    Qian, Xueming
    arXiv, 2022,
  • [39] Learning Via Decision Trees Approach for Video Super-Resolution
    Zhang, Yu-Zhu
    Siu, Wan-Chi
    Liu, Zhi-Song
    Law, Ngai-Fong
    PROCEEDINGS 2017 INTERNATIONAL CONFERENCE ON COMPUTATIONAL SCIENCE AND COMPUTATIONAL INTELLIGENCE (CSCI), 2017, : 558 - 562
  • [40] Video super-resolution based on deep learning: a comprehensive survey
    Hongying Liu
    Zhubo Ruan
    Peng Zhao
    Chao Dong
    Fanhua Shang
    Yuanyuan Liu
    Linlin Yang
    Radu Timofte
    Artificial Intelligence Review, 2022, 55 : 5981 - 6035