Temporal Consistency Learning of Inter-Frames for Video Super-Resolution

被引:20
|
作者
Liu, Meiqin [1 ,2 ]
Jin, Shuo [1 ,2 ]
Yao, Chao [3 ]
Lin, Chunyu [1 ,2 ]
Zhao, Yao [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Inst Informat Sci, Beijing 100044, Peoples R China
[2] Beijing Jiaotong Univ, Beijing Key Lab Adv Informat Sci & Network Technol, Beijing 100044, Peoples R China
[3] Univ Sci & Technol Beijing, Sch Comp & Commun Engn, Beijing 100083, Peoples R China
基金
中国国家自然科学基金;
关键词
Superresolution; Circuit stability; Optical flow; Image restoration; Image reconstruction; Degradation; Convolution; Bidirectional motion estimation; temporal consistency; self-alignment; video super-resolution; FUSION NETWORK;
D O I
10.1109/TCSVT.2022.3214538
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Video super-resolution (VSR) is a task that aims to reconstruct high-resolution (HR) frames from the low-resolution (LR) reference frame and multiple neighboring frames. The vital operation is to utilize the relative misaligned frames for the current frame reconstruction and preserve the consistency of the results. Existing methods generally explore information propagation and frame alignment to improve the performance of VSR. However, few studies focus on the temporal consistency of inter-frames. In this paper, we propose a Temporal Consistency learning Network (TCNet) for VSR in an end-to-end manner, to enhance the consistency of the reconstructed videos. A spatio-temporal stability module is designed to learn the self-alignment from inter-frames. Especially, the correlative matching is employed to exploit the spatial dependency from each frame to maintain structural stability. Moreover, a self-attention mechanism is utilized to learn the temporal correspondence to implement an adaptive warping operation for temporal consistency among multi-frames. Besides, a hybrid recurrent architecture is designed to leverage short-term and long-term information. We further present a progressive fusion module to perform a multistage fusion of spatio-temporal features. And the final reconstructed frames are refined by these fused features. Objective and subjective results of various experiments demonstrate that TCNet has superior performance on different benchmark datasets, compared to several state-of-the-art methods.
引用
收藏
页码:1507 / 1520
页数:14
相关论文
共 50 条
  • [41] Learning Trajectory-Aware Transformer for Video Super-Resolution
    Liu, Chengxu
    Yang, Huan
    Fu, Jianlong
    Qian, Xueming
    Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2022, 2022-June : 5677 - 5686
  • [42] Video super-resolution based on a spatio-temporal matching network
    Zhu, Xiaobin
    Li, Zhuangzi
    Lou, Jungang
    Shen, Qing
    PATTERN RECOGNITION, 2021, 110
  • [43] Residual Invertible Spatio-Temporal Network for Video Super-Resolution
    Zhu, Xiaobin
    Li, Zhuangzi
    Zhang, Xiao-Yu
    Li, Changsheng
    Liu, Yaqi
    Xue, Ziyu
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 5981 - 5988
  • [44] AIM 2019 Challenge on Video Temporal Super-Resolution: Methods and Results
    Nah, Seungjun
    Son, Sanghyun
    Timofte, Radu
    Lee, Kyoung Mu
    Li Siyao
    Pan, Ze
    Xu, Xiangyu
    Sun, Wenxiu
    Choi, Myungsub
    Kim, Heewon
    Han, Bohyung
    Xu, Ning
    Park, Bumjun
    Yu, Songhyun
    Kim, Sangmin
    Jeong, Jechang
    Shen, Wang
    Bao, Wenbo
    Zhai, Guangtao
    Chen, Li
    Gao, Zhiyong
    Chen, Guannan
    Lu, Yunhua
    Duan, Ran
    Liu, Tong
    Zhang, Lijie
    Park, Woonsung
    Kim, Munchurl
    Pisha, George
    Naor, Eyal
    Aloni, Lior
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 3388 - 3398
  • [45] Conditional Neural Video Coding with Spatial-Temporal Super-Resolution
    Wang, Henan
    Pan, Xiaohan
    Feng, Runsen
    Guo, Zongyu
    Chen, Zhibo
    2024 DATA COMPRESSION CONFERENCE, DCC, 2024, : 591 - 591
  • [46] VIDEO SUPER-RESOLUTION VIA EVENT-DRIVEN TEMPORAL ALIGNMENT
    Kai, Dachun
    Zhang, Yueyi
    Sun, Xiaoyan
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 2950 - 2954
  • [47] Learning Trajectory-Aware Transformer for Video Super-Resolution
    Liu, Chengxu
    Yang, Huan
    Fu, Jianlong
    Qian, Xueming
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 5677 - 5686
  • [48] Video Super-Resolution via a Spatio-Temporal Alignment Network
    Wen, Weilei
    Ren, Wenqi
    Shi, Yinghuan
    Nie, Yunfeng
    Zhang, Jingang
    Cao, Xiaochun
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 1761 - 1773
  • [49] CTVSR: Collaborative Spatial-Temporal Transformer for Video Super-Resolution
    Tang, Jun
    Lu, Chenyan
    Liu, Zhengxue
    Li, Jiale
    Dai, Hang
    Ding, Yong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (06) : 5018 - 5032
  • [50] Learning a Deep Dual Attention Network for Video Super-Resolution
    Li, Feng
    Bai, Huihui
    Zhao, Yao
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 (29) : 4474 - 4488