NeuralDome: A Neural Modeling Pipeline on Multi-View Human-Object Interactions

被引:4
|
作者
Zhang, Juze [1 ,2 ,3 ,4 ]
Luo, Haimin [1 ,4 ,5 ]
Yang, Hongdi [1 ,4 ]
Xu, Xinru [1 ,4 ]
Wu, Qianyang [1 ,4 ]
Shi, Ye [1 ,4 ]
Yu, Jingyi [1 ,4 ]
Xu, Lan [1 ,4 ]
Wang, Jingya [1 ,4 ]
机构
[1] ShanghaiTech Univ, Shanghai, Peoples R China
[2] Chinese Acad Sci, Shanghai Adv Res Inst, Shanghai, Peoples R China
[3] Univ Chinese Acad Sci, Shanghai, Peoples R China
[4] Shanghai Engn Res Ctr Intelligent Vis & Imaging, Shanghai, Peoples R China
[5] LumiAni Technol, Hemel Hempstead, England
关键词
D O I
10.1109/CVPR52729.2023.00853
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Humans constantly interact with objects in daily life tasks. Capturing such processes and subsequently conducting visual inferences from a fixed viewpoint suffers from occlusions, shape and texture ambiguities, motions, etc. To mitigate the problem, it is essential to build a training dataset that captures free-viewpoint interactions. We construct a dense multi-view dome to acquire a complex human object interaction dataset, named HODome, that consists of similar to 71M frames on 10 subjects interacting with 23 objects. To process the HODome dataset, we develop NeuralDome, a layer-wise neural processing pipeline tailored for multi-view video inputs to conduct accurate tracking, geometry reconstruction and free-view rendering, for both human subjects and objects. Extensive experiments on the HODome dataset demonstrate the effectiveness of NeuralDome on a variety of inference, modeling, and rendering tasks. Both the dataset and the NeuralDome tools will be disseminated to the community for further development, which can be found at https://juzezhang.github.io/NeuralDome
引用
收藏
页码:8834 / 8845
页数:12
相关论文
共 50 条
  • [1] Explicit Modeling of Human-Object Interactions in Realistic Videos
    Prest, Alessandro
    Ferrari, Vittorio
    Schmid, Cordelia
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2013, 35 (04) : 835 - 848
  • [2] Multi-view Neural Human Rendering
    Wu, Minye
    Wang, Yuehao
    Hu, Qiang
    Yu, Jingyi
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 1679 - 1688
  • [3] NeuralHOFusion: Neural Volumetric Rendering under Human-object Interactions
    Jiang, Yuheng
    Jiang, Suyi
    Sun, Guoxing
    Su, Zhuo
    Guo, Kaiwen
    Wu, Minye
    Yu, Jingyi
    Xu, Lan
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 6145 - 6155
  • [4] Learning Human-Object Interactions by Graph Parsing Neural Networks
    Qi, Siyuan
    Wang, Wenguan
    Jia, Baoxiong
    Shen, Jianbing
    Zhu, Song-Chun
    COMPUTER VISION - ECCV 2018, PT IX, 2018, 11213 : 407 - 423
  • [5] Modeling 4D Human-Object Interactions for Event and Object Recognition
    Wei, Ping
    Zhao, Yibiao
    Zheng, Nanning
    Zhu, Song-Chun
    2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 3272 - 3279
  • [6] NodeSLAM: Neural Object Descriptors for Multi-View Shape Reconstruction
    Sucar, Edgar
    Wada, Kentaro
    Davison, Andrew
    2020 INTERNATIONAL CONFERENCE ON 3D VISION (3DV 2020), 2020, : 949 - 958
  • [7] Learning to Detect Human-Object Interactions
    Chao, Yu-Wei
    Liu, Yunfan
    Liu, Xieyang
    Zeng, Huayi
    Deng, Jia
    2018 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV 2018), 2018, : 381 - 389
  • [8] Detecting and Recognizing Human-Object Interactions
    Gkioxari, Georgia
    Girshick, Ross
    Dollar, Piotr
    He, Kaiming
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 8359 - 8367
  • [9] Detecting human-object interactions in videos by modeling the trajectory of objects and human skeleton
    Li, Qiyue
    Xie, Xuemei
    Zhang, Chen
    Zhang, Jin
    Shi, Guangming
    NEUROCOMPUTING, 2022, 509 : 234 - 243
  • [10] A self-organizing neural network architecture for learning human-object interactions
    Mici, Luiza
    Parisi, German I.
    Wermter, Stefan
    NEUROCOMPUTING, 2018, 307 : 14 - 24