GaussianAvatar: Human avatar Gaussian splatting from monocular videos☆

被引:1
|
作者
Lin, Haian
Zhan, Yinwei [1 ]
机构
[1] Guangdong Univ Technol, Guangzhou Univ Town,West Rd, Guangzhou 510006, Peoples R China
来源
COMPUTERS & GRAPHICS-UK | 2025年 / 126卷
关键词
Neural radiance field; 3D Gaussian; Human reconstruction;
D O I
10.1016/j.cag.2024.104155
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Many application fields including virtual reality and movie production demand reconstructing high-quality digital human avatars from monocular videos and real-time rendering. However, existing neural radiance field (NeRF)-based methods are costly to train and render. In this paper, we propose GaussianAvatar, a novel framework that extends 3D Gaussian to dynamic human scenes, enabling fast training and real-time rendering. The human 3D Gaussian in canonical space is initialized and transformed to posed space using Linear Blend Skinning (LBS), based on pose parameters, to learn the fine details of the human body at a very small computational cost. We design a pose parameter refinement module and a LBS weight optimization module to increase the accuracy of the pose parameter detection in the real dataset and introduce multi-resolution hash coding to accelerate the training speed. Experimental results demonstrate that our method outperforms existing methods in terms of training time, rendering speed, and reconstruction quality.
引用
收藏
页数:9
相关论文
共 50 条
  • [21] Unsupervised Learning of Monocular Depth from Videos
    Gao Haosheng
    Teng Wang
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 3945 - 3950
  • [22] Toward Realistic 3D Avatar Generation with Dynamic 3D Gaussian Splatting for AR/VR Communication
    Song, Hail
    2024 IEEE CONFERENCE ON VIRTUAL REALITY AND 3D USER INTERFACES ABSTRACTS AND WORKSHOPS, VRW 2024, 2024, : 1124 - 1125
  • [23] An Optimization Based Framework for Human Pose Estimation in Monocular Videos
    Agarwal, Priyanshu
    Kumar, Suren
    Ryde, Julian
    Corso, Jason J.
    Krovi, Venkat N.
    ADVANCES IN VISUAL COMPUTING, ISVC 2012, PT I, 2012, 7431 : 575 - 586
  • [24] PPR: Physically Plausible Reconstruction from Monocular Videos
    Yang, Gengshan
    Yang, Shuo
    Zhang, John Z.
    Manchester, Zachary
    Ramanan, Deva
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 3891 - 3901
  • [25] Neural Head Avatars from Monocular RGB Videos
    Grassal, Philip-William
    Prinzler, Malte
    Leistner, Titus
    Rother, Carsten
    Niessner, Matthias
    Thies, Justus
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 18632 - 18643
  • [26] SelfRecon: Self Reconstruction Your Digital Avatar from Monocular Video
    Jiang, Boyi
    Hong, Yang
    Bao, Hujun
    Zhang, Juyong
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 5595 - 5605
  • [27] Clip Fusion with Bi-level Optimization for Human Mesh Reconstruction from Monocular Videos
    Wu, Peng
    Lu, Xiankai
    Shen, Jianbing
    Yin, Yilong
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 105 - 115
  • [28] HSR: Holistic 3D Human-Scene Reconstruction from Monocular Videos
    Xue, Lixin
    Guo, Chen
    Zheng, Chengwei
    Wang, Fangjinghua
    Jiang, Tianjian
    Ho, Hsuan-, I
    Kaufmann, Manuel
    Song, Jie
    Hilliges, Otmar
    COMPUTER VISION - ECCV 2024, PT LXXII, 2025, 15130 : 429 - 448
  • [29] Joint 3D Human Motion Capture and Physical Analysis from Monocular Videos
    Zell, Petrissa
    Wandt, Bastian
    Rosenhahn, Bodo
    2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, : 17 - 26
  • [30] Real-Time Retargeting of Human Poses from Monocular Images and Videos to the NAO Robot
    Burga O.
    Villegas J.
    Ugarte W.
    Journal of Computing Science and Engineering, 2024, 18 (01) : 47 - 56