Efficient Pose: Efficient human pose estimation with neural architecture search

被引:3
|
作者
Wenqiang Zhang [1 ]
Jiemin Fang [2 ,1 ]
Xinggang Wang [1 ]
Wenyu Liu [1 ]
机构
[1] School of EIC,Huazhong University of Science and Technology
[2] Institute of Artificial Intelligence,Huazhong University of Science and
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP391.41 []; TP183 [人工神经网络与计算];
学科分类号
080203 ; 081104 ; 0812 ; 0835 ; 1405 ;
摘要
Human pose estimation from image and video is a key task in many multimedia applications.Previous methods achieve great performance but rarely take efficiency into consideration, which makes it difficult to implement the networks on lightweight devices. Nowadays, real-time multimedia applications call for more efficient models for better interaction.Moreover, most deep neural networks for pose estimation directly reuse networks designed for image classification as the backbone, which are not optimized for the pose estimation task. In this paper, we propose an efficient framework for human pose estimation with two parts, an efficient backbone and an efficient head.By implementing a differentiable neural architecture search method, we customize the backbone network design for pose estimation, and reduce computational cost with negligible accuracy degradation. For the efficient head, we slim the transposed convolutions and propose a spatial information correction module to promote the performance of the final prediction. In experiments, we evaluate our networks on the MPII and COCO datasets. Our smallest model requires only0.65 GFLOPs with 88.1% PCKh@0.5 on MPII and our large model needs only 2 GFLOPs while its accuracy is competitive with the state-of-the-art large model,HRNet, which takes 9.5 GFLOPs.
引用
收藏
页码:335 / 347
页数:13
相关论文
共 50 条
  • [1] EfficientPose: Efficient human pose estimation with neural architecture search
    Zhang, Wenqiang
    Fang, Jiemin
    Wang, Xinggang
    Liu, Wenyu
    [J]. COMPUTATIONAL VISUAL MEDIA, 2021, 7 (03) : 335 - 347
  • [2] EfficientPose: Efficient human pose estimation with neural architecture search
    Wenqiang Zhang
    Jiemin Fang
    Xinggang Wang
    Wenyu Liu
    [J]. Computational Visual Media, 2021, 7 : 335 - 347
  • [3] ViPNAS: Efficient Video Pose Estimation via Neural Architecture Search
    Xu, Lumin
    Guan, Yingda
    Jin, Sheng
    Liu, Wentao
    Qian, Chen
    Luo, Ping
    Ouyang, Wanli
    Wang, Xiaogang
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 16067 - 16076
  • [4] Neural Architecture Search for Joint Human Parsing and Pose Estimation
    Zeng, Dan
    Huang, Yuhang
    Bao, Qian
    Zhang, Junjie
    Su, Chi
    Liu, Wu
    [J]. 2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 11365 - 11374
  • [5] Lite Pose: Efficient Architecture Design for 2D Human Pose Estimation
    Wang, Yihan
    Li, Muyang
    Cai, Han
    Chen, Weiming
    Han, Song
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 13116 - 13126
  • [6] Pose-native Neural Architecture Search for Multi-person Pose Estimation
    Bao, Qian
    Liu, Wu
    Hong, Jun
    Duan, Lingyu
    Mei, Tao
    [J]. MM '20: PROCEEDINGS OF THE 28TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, 2020, : 592 - 600
  • [7] Efficient Human Pose Estimation in Hierarchical Context
    Zhang, Feng
    Zhu, Xiatian
    Ye, Mao
    [J]. IEEE ACCESS, 2019, 7 : 29365 - 29373
  • [8] An Efficient Method for Boosting Human Pose Estimation
    Xiang, Shicheng
    Chen, Xiao
    Zhou, Jun
    [J]. 2021 IEEE INTERNATIONAL SYMPOSIUM ON BROADBAND MULTIMEDIA SYSTEMS AND BROADCASTING (BMSB), 2021,
  • [9] Efficient High-Resolution Human Pose Estimation
    Qin, Xiaofei
    Qiu, Lingfeng
    He, Changxiang
    Zhang, Xuedian
    [J]. PRICAI 2022: TRENDS IN ARTIFICIAL INTELLIGENCE, PT III, 2022, 13631 : 383 - 396
  • [10] An efficient sparse pruning method for human pose estimation
    Wang, Mingyang
    Sun, Tianyi
    Song, Kang
    Li, Shuang
    Jiang, Jing
    Sun, Linjun
    [J]. CONNECTION SCIENCE, 2022, 34 (01) : 960 - 974