A Simple Baseline for Video Restoration with Grouped Spatial-temporal Shift

被引:8
|
作者
Li, Dasong [1 ]
Shi, Xiaoyu [1 ]
Zhang, Yi [1 ]
Cheung, Ka Chun [2 ]
See, Simon [2 ]
Wang, Xiaogang [1 ,4 ]
Qin, Hongwei [3 ]
Li, Hongsheng [1 ,4 ]
机构
[1] CUHK MMLab, Hong Kong, Peoples R China
[2] NVIDIA AI Technol Ctr, Hong Kong, Peoples R China
[3] SenseTime Res, Hong Kong, Peoples R China
[4] CPII InnoHK, Hong Kong, Peoples R China
基金
国家重点研发计划;
关键词
D O I
10.1109/CVPR52729.2023.00947
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Video restoration, which aims to restore clear frames from degraded videos, has numerous important applications. The key to video restoration depends on utilizing inter-frame information. However, existing deep learning methods often rely on complicated network architectures, such as optical flow estimation, deformable convolution, and cross-frame self-attention layers, resulting in high computational costs. In this study, we propose a simple yet effective framework for video restoration. Our approach is based on grouped spatial-temporal shift, which is a lightweight and straightforward technique that can implicitly capture inter-frame correspondences for multi-frame aggregation. By introducing grouped spatial shift, we attain expansive effective receptive fields. Combined with basic 2D convolution, this simple framework can effectively aggregate inter-frame information. Extensive experiments demonstrate that our framework outperforms the previous state-of-the-art method, while using less than a quarter of its computational cost, on both video deblurring and video denoising tasks. These results indicate the potential for our approach to significantly reduce computational overhead while maintaining high-quality results. Code is avaliable at https://github.com/dasongli1/Shift-Net.
引用
收藏
页码:9822 / 9832
页数:11
相关论文
共 50 条
  • [1] ShiftFormer: Spatial-Temporal Shift Operation in Video Transformer
    Yang, Beiying
    Zhu, Guibo
    Ge, Guojing
    Luo, Jinzhao
    Wang, Jinqiao
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1895 - 1900
  • [2] Grouped Spatial-Temporal Aggregation for Efficient Action Recognition
    Luo, Chenxu
    Yuille, Alan
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 5511 - 5520
  • [3] Video Description with Spatial-Temporal Attention
    Tu, Yunbin
    Zhang, Xishan
    Liu, Bingtao
    Yan, Chenggang
    [J]. PROCEEDINGS OF THE 2017 ACM MULTIMEDIA CONFERENCE (MM'17), 2017, : 1014 - 1022
  • [4] Spatial-Temporal Identity: A Simple yet Effective Baseline for Multivariate Time Series Forecasting
    Shao, Zezhi
    Zhang, Zhao
    Wang, Fei
    Wei, Wei
    Xu, Yongjun
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON INFORMATION AND KNOWLEDGE MANAGEMENT, CIKM 2022, 2022, : 4454 - 4458
  • [5] Optimum Video Subset and Spatial-Temporal Video Retrieval
    Wang M.-Z.
    Liu X.-J.
    Sun K.-X.
    Wang Z.-R.
    [J]. Jisuanji Xuebao/Chinese Journal of Computers, 2019, 42 (09): : 2004 - 2023
  • [6] A SIMPLE SPATIAL-TEMPORAL MODEL OF RAINFALL
    COX, DR
    ISHAM, V
    [J]. PROCEEDINGS OF THE ROYAL SOCIETY OF LONDON SERIES A-MATHEMATICAL PHYSICAL AND ENGINEERING SCIENCES, 1988, 415 (1849): : 317 - 328
  • [7] A SIMPLE DESCRIPTION OF SPATIAL-TEMPORAL PROCESSES
    POSA, D
    [J]. COMPUTATIONAL STATISTICS & DATA ANALYSIS, 1993, 15 (04) : 425 - 437
  • [8] Spatial-temporal decorrelation for image/video coding
    Wang, Miaohui
    Ngan, King Ngi
    Xu, Long
    [J]. 2012 PICTURE CODING SYMPOSIUM (PCS), 2012, : 201 - 204
  • [9] SPATIAL-TEMPORAL ATTENTION ANALYSIS FOR HOME VIDEO
    Qiu, Xuekan
    Jiang, Shuqiang
    Liu, Huiying
    Huang, Qingming
    Cao, Longbing
    [J]. 2008 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, VOLS 1-4, 2008, : 1517 - +
  • [10] Video summarization by spatial-temporal graph optimization
    Lu, S
    Lyu, MR
    King, I
    [J]. 2004 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS, VOL 2, PROCEEDINGS, 2004, : 197 - 200