Transformer-based rapid human pose estimation network

被引:6
|
作者
Wang, Dong [1 ]
Xie, Wenjun [2 ,3 ]
Cai, Youcheng [1 ]
Li, Xinjie [1 ]
Liu, Xiaoping [1 ]
机构
[1] Hefei Univ Technol, Sch Comp Sci & Informat Engn, Hefei 230009, Peoples R China
[2] Hefei Univ Technol, Sch Software, Hefei 230009, Peoples R China
[3] Hefei Univ Technol, Anhui Prov Key Lab Ind Safety & Emergency Technol, Hefei 230601, Peoples R China
来源
COMPUTERS & GRAPHICS-UK | 2023年 / 116卷
关键词
Transformer architecture; Human pose estimation; Inference speed; Computational cost; ACTION RECOGNITION; SKELETON;
D O I
10.1016/j.cag.2023.09.001
中图分类号
TP31 [计算机软件];
学科分类号
081202 ; 0835 ;
摘要
Most current human pose estimation methods pursue excellent performance via large models and intensive computational requirements, resulting in slower models. These methods cannot be effectively adopted for human pose estimation in real applications due to their high memory and computational costs. To achieve a trade-off between accuracy and efficiency, we propose TRPose, a Transformer-based network for human pose estimation rapidly. TRPose consists of an early convolutional stage and a later Transformer stage seamlessly. Concretely, the convolutional stage forms a Rapid Fusion Module (RFM), which efficiently acquires multi-scale features via three parallel convolution branches. The Transformer stage utilizes multi-resolution Transformers to construct a Dual scale Encoder Module (DEM), aiming at learning long-range dependencies from different scale features of the whole human skeletal keypoints. The experiments show that TRPose acquires 74.3 AP and 73.8 AP on COCO validation and testdev datasets with 170+ FPS on a GTX2080Ti, which achieves the better efficiency and effectiveness tradeoffs than most state-of-the-art methods. Our model also outperforms mainstream Transformer-based architectures on MPII dataset, yielding 89.9 PCK@0.5 score on val set without extra data. (c) 2023 Elsevier Ltd. All rights reserved.
引用
收藏
页码:317 / 326
页数:10
相关论文
共 50 条
  • [21] Pose Relation Transformer Refine Occlusions for Human Pose Estimation
    Chi, Hyung-gun
    Chi, Seunggeun
    Chan, Stanley
    Ramani, Karthik
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 6138 - 6145
  • [22] A transformer-based network for speech recognition
    Tang L.
    International Journal of Speech Technology, 2023, 26 (02) : 531 - 539
  • [23] SMGNFORMER: Fusion Mamba-graph transformer network for human pose estimation
    Li, Yi
    Wang, Zan
    Niu, Weiran
    IET COMPUTER VISION, 2025, 19 (01)
  • [24] A Lightweight Context-Aware Feature Transformer Network for Human Pose Estimation
    Ma, Yanli
    Shi, Qingxuan
    Zhang, Fan
    ELECTRONICS, 2024, 13 (04)
  • [25] Vision-Based Body Pose Estimation of Excavator Using a Transformer-Based Deep-Learning Model
    Ji, Ankang
    Fan, Hongqin
    Xue, Xiaolong
    JOURNAL OF COMPUTING IN CIVIL ENGINEERING, 2025, 39 (02)
  • [26] A Fast and Effective Transformer for Human Pose Estimation
    Wang, Dong
    Xie, Wenjun
    Cai, Youcheng
    Liu, Xiaoping
    IEEE SIGNAL PROCESSING LETTERS, 2022, 29 : 992 - 996
  • [27] Spatiotemporal Learning Transformer for Video-Based Human Pose Estimation
    Gai, Di
    Feng, Runyang
    Min, Weidong
    Yang, Xiaosong
    Su, Pengxiang
    Wang, Qi
    Han, Qing
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2023, 33 (09) : 4564 - 4576
  • [28] A TRANSFORMER-BASED NETWORK FOR UNIFYING RADIO MAP ESTIMATION AND OPTIMIZED SITE SELECTION
    Zheng, Yi
    Liao, Cunyi
    Wang, Ji
    Liu, Shouyin
    2024 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING WORKSHOPS, ICASSPW 2024, 2024, : 610 - 614
  • [29] YOLOPose V2: Understanding and improving transformer-based 6D pose estimation
    Periyasamy, Arul Selvam
    Amini, Arash
    Tsaturyan, Vladimir
    Behnke, Sven
    ROBOTICS AND AUTONOMOUS SYSTEMS, 2023, 168
  • [30] Lightweight Human Pose Estimation Network Based on HRNet
    Liang Q.
    Wu Y.
    Hunan Daxue Xuebao/Journal of Hunan University Natural Sciences, 2023, 50 (02): : 112 - 121