Textural Detail Preservation Network for Video Frame Interpolation

被引:0
|
作者
Yoon, Kihwan [1 ,2 ]
Huh, Jingang [1 ]
Kim, Yong Han [2 ]
Kim, Sungjei [1 ]
Jeong, Jinwoo [1 ]
机构
[1] Korea Elect Technol Inst KETI, Seongnam Si 13488, Gyeonggi Do, South Korea
[2] Univ Seoul, Sch Elect & Comp Engn, Seoul 02504, South Korea
关键词
Video frame interpolation; textural detail preservation; perceptual loss; synthesis network; ENHANCEMENT;
D O I
10.1109/ACCESS.2023.3294964
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
The subjective image quality of the Video Frame Interpolation (VFI) result depends on whether image features such as edges, textures and blobs are preserved. With the development of deep learning, various algorithms have been proposed and the objective results of VFI have significantly improved. Moreover, perceptual loss has been used in a method that enhances subjective quality by preserving the features of the image, and as a result, the subjective quality is improved. Despite the quality enhancements achieved in VFI, no analysis has been performed to preserve specific features in the interpolated frames. Therefore, we conducted an analysis to preserve textural detail, such as film grain noise, which can represent the texture of an image, and weak textures, such as droplets or particles. Based on our analysis, we identify the importance of synthesis networks in textural detail preservation and propose an enhanced synthesis network, the Textural Detail Preservation Network (TDPNet). Furthermore, based on our analysis, we propose a Perceptual Training Method (PTM) to address the issue of degraded Peak Signal-to-Noise Ratio (PSNR) when simply applying perceptual loss and to preserve more textural detail. We also propose a Multi-scale Resolution Training Method (MRTM) to address the issue of poor performance when testing datasets with a resolution different from that of the training dataset. The experimental results of the proposed network was outperformed in LPIPS and DISTS on the Vimeo90K, HD, SNU-FILM and UVG datasets compared with the state-of-the-art VFI algorithms, and the subjective results were also outperformed. Furthermore, applying PTM improved PSNR results by an average of 0.293dB compared to simply applying perceptual loss.
引用
收藏
页码:71994 / 72006
页数:13
相关论文
共 50 条
  • [31] MFNet:Real-Time Motion Focus Network for Video Frame Interpolation
    Zhu, Guosong
    Qin, Zhen
    Ding, Yi
    Liu, Yao
    Qin, Zhiguang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 3251 - 3262
  • [32] MVFI-Net: Motion-Aware Video Frame Interpolation Network
    Lin, Xuhu
    Zhao, Lili
    Liu, Xi
    Chen, Jianwen
    COMPUTER VISION - ACCV 2022, PT III, 2023, 13843 : 340 - 356
  • [33] UPR-Net: A Unified Pyramid Recurrent Network for Video Frame Interpolation
    Jin, Xin
    Wu, Longhai
    Chen, Jie
    Chen, Youxin
    Koo, Jayoon
    Hahm, Cheul-Hee
    Chen, Zhao-Min
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025, 133 (01) : 16 - 30
  • [34] Deep Video Frame Interpolation Using Cyclic Frame Generation
    Liu, Yu-Lun
    Liao, Yi-Tung
    Lin, Yen-Yu
    Chuang, Yung-Yu
    THIRTY-THIRD AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FIRST INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / NINTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2019, : 8794 - 8802
  • [35] Video Object Segmentation-aware Video Frame Interpolation
    Yoo, Jun-Sang
    Lee, Hongjae
    Jung, Seung-Won
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 12288 - 12299
  • [36] Forward Warping-Based Video Frame Interpolation Using a Motion Selective Network
    Heo, Jeonghwan
    Jeong, Jechang
    ELECTRONICS, 2022, 11 (16)
  • [37] A comprehensive survey on video frame interpolation techniques
    Parihar, Anil Singh
    Varshney, Disha
    Pandya, Kshitija
    Aggarwal, Ashray
    VISUAL COMPUTER, 2022, 38 (01): : 295 - 319
  • [38] Video Frame Interpolation via Adaptive Convolution
    Niklaus, Simon
    Mai, Long
    Liu, Feng
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 2270 - 2279
  • [39] Hybrid Warping Fusion for Video Frame Interpolation
    Li, Yu
    Zhu, Ye
    Li, Ruoteng
    Wang, Xintao
    Luo, Yue
    Shan, Ying
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2022, 130 (12) : 2980 - 2993
  • [40] Phase-Based Frame Interpolation for Video
    Meyer, Simone
    Wang, Oliver
    Zimmer, Henning
    Grosse, Max
    Sorkine-Hornung, Alexander
    2015 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2015, : 1410 - 1418