Relation-aware interaction spatio-temporal network for 3D human pose estimation

被引:1
|
作者
Zhang, Hehao [1 ]
Hu, Zhengping [1 ]
Bi, Shuai [1 ]
Di, Jirui [1 ]
Sun, Zhe [1 ]
机构
[1] Yanshan Univ, Dept Informat Sci & Engn, Qinhuangdao 066000, Hebei, Peoples R China
基金
中国国家自然科学基金;
关键词
Bi-directional interaction module; Spatial kinematics modeling block; Temporal trajectory modeling block; Video processing;
D O I
10.1016/j.dsp.2024.104764
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
3D human pose estimation is a fundamental task in analyzing human behavior, which has many practical applications. However, existing methods suffer from high time complexity and weak capability to acquire the relations at the human joint level and the spatio-temporal level. To this end, the R elation-aware I nteraction S patio-temporal Net work (RISNet) is presented to achieve a better speed-accuracy trade-off in a parallel interactive architecture. Firstly, the Spatial Kinematics Modeling Block (SKMB) is proposed to encode spatially positional correlations among human joints, thereby capturing cross-joint kinematic dependencies in each frame. Secondly, the Temporal Trajectory Modeling Block (TTMB) is employed to further process the temporal motion trajectory of individual joints at several various frame scales. Besides, the bi-directional interaction modules across branches are presented to enhance modeling abilities at the spatio-temporal level. Experiments on Human 3.6M, HumanEva-I and MPI-INF-3DHP benchmarks indicate that the RISNet gains significant improvement compared to several state-of-the-art techniques. In conclusion, the proposed approach elegantly extracts critical features of body joints in the spatio-temporal domain with fewer model parameters and lower time complexity.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] Spatio-Temporal Dynamic Interlaced Network for 3D human pose estimation in video
    Xu, Feiyi
    Wang, Jifan
    Sun, Ying
    Qi, Jin
    Dong, Zhenjiang
    Sun, Yanfei
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2025, 251
  • [2] A Graph Attention Spatio-temporal Convolutional Network for 3D Human Pose Estimation in Video
    The Biomimetic and Intelligent Robotics Lab , School of Electromechanical Engineering, Guangdong University of Technology, Guangzhou
    510006, China
    不详
    不详
    Proc IEEE Int Conf Rob Autom, 2021, (3374-3380): : 3374 - 3380
  • [3] A Graph Attention Spatio-temporal Convolutional Network for 3D Human Pose Estimation in Video
    Liu, Junfa
    Rojas, Juan
    Li, Yihui
    Liang, Zhijun
    Guan, Yisheng
    Xi, Ning
    Zhu, Haifei
    2021 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2021), 2021, : 3374 - 3380
  • [4] SPATIO-TEMPORAL ATTENTION GRAPH FOR MONOCULAR 3D HUMAN POSE ESTIMATION
    Zhang, Lijun
    Shao, Xiaohu
    Li, Zhenghao
    Zhou, Xiang-Dong
    Shi, Yu
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 1231 - 1235
  • [5] Global and Local Spatio-Temporal Encoder for 3D Human Pose Estimation
    Wang, Yong
    Kang, Hongbo
    Wu, Doudou
    Yang, Wenming
    Zhang, Longbin
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 4039 - 4049
  • [6] 3D Human Pose Estimation with Spatio-Temporal Criss-cross Attention
    Tang, Zhenhua
    Qiu, Zhaofan
    Hao, Yanbin
    Hong, Richang
    Yao, Ting
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 4790 - 4799
  • [7] A fused convolutional spatio-temporal progressive approach for 3D human pose estimation
    Zhang, Hehao
    Hu, Zhengping
    Sun, Zhe
    Zhao, Mengyao
    Bi, Shuai
    Di, Jirui
    VISUAL COMPUTER, 2024, 40 (06): : 4387 - 4399
  • [8] Spatio-temporal 3D pose estimation of objects in stereo images
    Barrois, Bjoern
    Woehler, Christian
    COMPUTER VISION SYSTEMS, PROCEEDINGS, 2008, 5008 : 507 - 516
  • [9] Skeleton-Based Spatio-Temporal U-Network for 3D Human Pose Estimation in Video
    Li, Weiwei
    Du, Rong
    Chen, Shudong
    SENSORS, 2022, 22 (07)
  • [10] Event-driven Video Deblurring via Spatio-Temporal Relation-Aware Network
    Cao, Chengzhi
    Fu, Xueyang
    Zhu, Yurui
    Shi, Gege
    Zha, Zheng-Jun
    PROCEEDINGS OF THE THIRTY-FIRST INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2022, 2022, : 799 - 805