Geometry-Incorporated Posing of a Full-Body Avatar From Sparse Trackers

被引:0
|
作者
Anvari, Taravat [1 ]
Park, Kyoungju [1 ]
机构
[1] Chung Ang Univ, Sch Comp Sci & Engn, Seoul, South Korea
关键词
3D human pose estimation; avatar; mixed reality; virtual reality;
D O I
10.1109/ACCESS.2023.3299323
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Accurately rendering a user's full body in a virtual environment is crucial for embodied mixed reality (MR) experiences. Conventional MR systems provide sparse trackers such as a headset and two hand-held controllers. Recent studies have intensively investigated learning methods to regress untracked joints from sparse trackers and have produced plausible poses in real time for MR applications. However, most studies have assumed that they either know the position of the root joint or constrain it, yielding stiff pelvis motions. This paper presents the first geometry-incorporated learning method to generate the position and rotation of all joints, including the root joint, from the head and hands information for a wide range of motions. We split the problem into identifying a reference frame and a pose inference with respect to the new reference frame. Our method defines an avatar frame by setting a non-joint as the origin and transforms joint data in a world coordinate system into the avatar coordinate system. Our learning builds on a propagating long short-term memory (LSTM) network exploiting prior knowledge of the kinematic chains and the previous time domain. The learned joints are transformed back to obtain the positions with respect to the world frame. In our experiments, our method achieves competitive accuracy and robustness with the state-of-the-art speed of approximately 130 fps on motion capture datasets and the wild tracking data obtained from commercial MR devices. Our experiments confirm that the proposed method is practically applicable to MR systems.
引用
收藏
页码:78858 / 78866
页数:9
相关论文
共 50 条
  • [41] Deep Full-Body HPE for Activity Recognition from RGB Frames Only
    Boualia, Sameh Neili
    Amara, Najoua Essoukri Ben
    INFORMATICS-BASEL, 2021, 8 (01):
  • [42] Construction of a space of motion labels from their mapping to full-body motion symbols
    Takano, Wataru
    Nakamura, Yoshihiko
    ADVANCED ROBOTICS, 2015, 29 (02) : 115 - 126
  • [43] Synthesis of kinematically constrained full-body motion from stochastic motion model
    Takano, Wataru
    Nakamura, Yoshihiko
    AUTONOMOUS ROBOTS, 2019, 43 (07) : 1881 - 1894
  • [44] Synthesis of kinematically constrained full-body motion from stochastic motion model
    Wataru Takano
    Yoshihiko Nakamura
    Autonomous Robots, 2019, 43 : 1881 - 1894
  • [45] Discriminative Human Full-Body Pose Estimation from Wearable Inertial Sensor Data
    Schwarz, Loren Arthur
    Mateus, Diana
    Navab, Nassir
    MODELLING THE PHYSIOLOGICAL HUMAN, 2009, 5903 : 159 - 172
  • [46] Learning full-body motions from monocular vision: Dynamic imitation in a humanoid robot
    Cole, Jeffrey B.
    Grimes, David B.
    Rao, Rajesh P. N.
    2007 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, VOLS 1-9, 2007, : 246 - +
  • [47] Person Identification Using Full-Body Motion and Anthropometric Biometrics from Kinect Videos
    Munsell, Brent C.
    Temlyakov, Andrew
    Qu, Chengzheng
    Wang, Song
    COMPUTER VISION - ECCV 2012, PT III, 2012, 7585 : 91 - 100
  • [48] Emotion Recognition From Full-Body Motion Using Multiscale Spatio-Temporal Network
    Wang, Tao
    Liu, Shuang
    He, Feng
    Dai, Weina
    Du, Minghao
    Ke, Yufeng
    Ming, Dong
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2024, 15 (03) : 898 - 912
  • [49] FROM TECHNOLOGIES TO FULL-BODY EXPERIENCE: LESSONS LEARNT ON A CROSS-FRANCE STUDY PILGRIMAGE
    Sudicky, Petr
    OVERCOMING THE CHALLENGES AND BARRIERS IN OPEN EDUCATION (DISCO 2018), 2018, : 395 - 400
  • [50] Detailed Full-Body Reconstructions of Moving People from Monocular RGB-D Sequences
    Bogo, Federica
    Black, Michael J.
    Loper, Matthew
    Romero, Javier
    2015 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2015, : 2300 - 2308