Editable Free-Viewpoint Video using a Layered Neural Representation

被引:62
|
作者
Zhang, Jiakai [1 ,2 ]
Liu, Xinhang [1 ]
Ye, Xinyi [1 ]
Zhao, Fuqiang [1 ]
Zhang, Yanshun [3 ]
Wu, Minye [1 ]
Zhang, Yingliang [3 ]
Xu, Lan [1 ]
Yu, Jingyi [1 ]
机构
[1] ShanghaiTech Univ, Shanghai, Peoples R China
[2] Stereye Intelligent Technol Co Ltd, Shanghai, Peoples R China
[3] DGene Digital Technol Co Ltd, Pudong, Peoples R China
来源
ACM TRANSACTIONS ON GRAPHICS | 2021年 / 40卷 / 04期
关键词
free-viewpoint video; novel view syntheis; neural rendering; visual editing; neural representation; dynamic scene modeling; CAPTURE;
D O I
10.1145/3450626.3459756
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Generating free-viewpoint videos is critical for immersive VR/AR experience, but recent neural advances still lack the editing ability to manipulate the visual perception for large dynamic scenes. To fill this gap, in this paper, we propose the first approach for editable free-viewpoint video generation for large-scale view-dependent dynamic scenes using only 16 cameras. The core of our approach is a new layered neural representation, where each dynamic entity, including the environment itself, is formulated into a spatiotemporal coherent neural layered radiance representation called ST-NeRF. Such a layered representation supports manipulations of the dynamic scene while still supporting a wide free viewing experience. In our ST-NeRF, we represent the dynamic entity/layer as a continuous function, which achieves the disentanglement of location, deformation as well as the appearance of the dynamic entity in a continuous and self-supervised manner. We propose a scene parsing 4D label map tracking to disentangle the spatial information explicitly and a continuous deform module to disentangle the temporal motion implicitly. An object-aware volume rendering scheme is further introduced for the re-assembling of all the neural layers. We adopt a novel layered loss and motion-aware ray sampling strategy to enable efficient training for a large dynamic scene with multiple performers, Our framework further enables a variety of editing functions, i.e., manipulating the scale and location, duplicating or retiming individual neural layers to create numerous visual effects while preserving high realism. Extensive experiments demonstrate the effectiveness of our approach to achieve high-quality, photo-realistic, and editable free-viewpoint video generation for dynamic scenes.
引用
收藏
页数:18
相关论文
共 50 条
  • [1] Efficient Neural Radiance Fields for Interactive Free-viewpoint Video
    Lin, Haotong
    Peng, Sida
    Xu, Zhen
    Yan, Yunzhi
    Shuai, Qing
    Bao, Hujun
    Zhou, Xiaowei
    [J]. PROCEEDINGS SIGGRAPH ASIA 2022, 2022,
  • [2] Interactive free-viewpoint video generation
    Wang, Yanru
    Huang, Zhihao
    Zhu, Hao
    Li, Wei
    Cao, Xun
    Yang, Ruigang
    [J]. Virtual Reality and Intelligent Hardware, 2020, 2 (03): : 247 - 260
  • [3] Reflectance Estimation for Free-viewpoint Video
    Ash, George
    Tomori, Juraj
    Pelton, Mike
    Dupont, Charles
    [J]. SIGGRAPH '21: ACM SIGGRAPH 2021 POSTERS, 2021,
  • [4] Free-viewpoint video of human actors
    Carranza, J
    Theobalt, C
    Magnor, MA
    Seidel, HP
    [J]. ACM TRANSACTIONS ON GRAPHICS, 2003, 22 (03): : 569 - 577
  • [5] Unconstrained free-viewpoint video coding
    Lamboray, E
    Würmlin, S
    Waschbüsch, M
    Gross, M
    Pfister, H
    [J]. ICIP: 2004 INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, VOLS 1- 5, 2004, : 3261 - 3264
  • [6] Rendering of Free-viewpoint Video on the Cloud
    Mallia, Mark
    Debono, Carl James
    [J]. 17TH IEEE INTERNATIONAL CONFERENCE ON SMART TECHNOLOGIES - IEEE EUROCON 2017 CONFERENCE PROCEEDINGS, 2017, : 9 - 14
  • [7] Space-time Neural Irradiance Fields for Free-Viewpoint Video
    Xian, Wenqi
    Huang, Jia-Bin
    Kopf, Johannes
    Kim, Changil
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 9416 - 9426
  • [8] FreeCast: Graceful Free-Viewpoint Video Delivery
    Fujihashi, Takuya
    Koike-Akino, Toshiaki
    Watanabe, Takashi
    Orlik, Philip V.
    [J]. IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (04) : 1000 - 1010
  • [9] OPTIMIZATION OF DISTRIBUTED FREE-VIEWPOINT VIDEO SYNTHESIS
    Huszak, Arpad
    [J]. 2014 3DTV-CONFERENCE: THE TRUE VISION - CAPTURE, TRANSMISSION AND DISPLAY OF 3D VIDEO (3DTV-CON), 2014,
  • [10] Intrinsic Textures for Relightable Free-Viewpoint Video
    Imber, James
    Guillemaut, Jean-Yves
    Hilton, Adrian
    [J]. COMPUTER VISION - ECCV 2014, PT II, 2014, 8690 : 392 - 407