Space-Time Distillation for Video Super-Resolution

被引:29
|
作者
Xiao, Zeyu [1 ]
Fu, Xueyang [1 ]
Huang, Jie [1 ]
Cheng, Zhen [1 ]
Xiong, Zhiwei [1 ]
机构
[1] Univ Sci & Technol China, Hefei, Anhui, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
D O I
10.1109/CVPR46437.2021.00215
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Compact video super-resolution (VSR) networks can be easily deployed on resource-limited devices, e.g., smartphones and wearable devices, but have considerable performance gaps compared with complicated VSR networks that require a large amount of computing resources. In this paper, we aim to improve the performance of compact VSR networks without changing their original architectures, through a knowledge distillation approach that transfers knowledge from a complicated VSR network to a compact one. Specifically, we propose a space-time distillation (STD) scheme to exploit both spatial and temporal knowledge in the VSR task. For space distillation, we extract spatial attention maps that hint the high-frequency video content from both networks, which are further used for transferring spatial modeling capabilities. For time distillation, we narrow the performance gap between compact models and complicated models by distilling the feature similarity of the temporal memory cells, which are encoded from the sequence of feature maps generated in the training clips using ConvLSTM. During the training process, STD can be easily incorporated into any network without changing the original network architecture. Experimental results on standard benchmarks demonstrate that, in resource-constrained situations, the proposed method notably improves the performance of existing VSR networks without increasing the inference time.
引用
收藏
页码:2113 / 2122
页数:10
相关论文
共 50 条
  • [1] Learning for Unconstrained Space-Time Video Super-Resolution
    Shi, Zhihao
    Liu, Xiaohong
    Li, Chengqi
    Dai, Linhui
    Chen, Jun
    Davidson, Timothy N.
    Zhao, Jiying
    [J]. IEEE TRANSACTIONS ON BROADCASTING, 2022, 68 (02) : 345 - 358
  • [2] Space-Time Super-Resolution from a Single Video
    Shahar, Oded
    Faktor, Alon
    Irani, Michal
    [J]. 2011 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2011,
  • [3] Space-time super-resolution
    Shechtman, E
    Caspi, Y
    Irani, M
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2005, 27 (04) : 531 - 545
  • [4] Space-Time Video Super-Resolution Using Temporal Profiles
    Xiao, Zeyu
    Xiong, Zhiwei
    Fu, Xueyang
    Liu, Dong
    Zha, Zheng-Jun
    [J]. MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 664 - 672
  • [5] Temporal Modulation Network for Controllable Space-Time Video Super-Resolution
    Xu, Gang
    Xu, Jun
    Li, Zhen
    Wang, Liang
    Sun, Xing
    Cheng, Ming-Ming
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 6384 - 6393
  • [6] STDAN: Deformable Attention Network for Space-Time Video Super-Resolution
    Wang, Hai
    Xiang, Xiaoyu
    Tian, Yapeng
    Yang, Wenming
    Liao, Qingmin
    [J]. IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (08) : 10606 - 10616
  • [7] Learned Low Bitrate Video Compression with Space-Time Super-Resolution
    Yang, Jiayu
    Yang, Chunhui
    Xiong, Fei
    Wang, Feng
    Wang, Ronggang
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2022, 2022, : 1785 - 1789
  • [8] Space-Time Video Super-Resolution 3D Transformer
    Zheng, Minyan
    Luo, Jianping
    [J]. MULTIMEDIA MODELING, MMM 2023, PT II, 2023, 13834 : 374 - 385
  • [9] STDAN: Deformable Attention Network for Space-Time Video Super-Resolution
    Wang, Hai
    Xiang, Xiaoyu
    Tian, Yapeng
    Yang, Wenming
    Liao, Qingmin
    [J]. arXiv, 2022,
  • [10] Space-Time Video Super-Resolution Based on Long-Term Time Dependence
    Huang, Xinyi
    Xue, Tong
    Hu, Yiwen
    Lan, Hui
    Sun, Jinshan
    [J]. ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VII, 2023, 14260 : 514 - 526