A sequence models-based real-time multi-person action recognition method with monocular vision

被引:2
|
作者
Yang, Aolei [1 ]
Lu, Wei [1 ]
Naeem, Wasif [2 ]
Chen, Ling [3 ]
Fei, Minrui [1 ]
机构
[1] Shanghai Univ, Sch Mechatron Engn & Automat, Shanghai 200444, Peoples R China
[2] Queens Univ Belfast, Sch Elect Elect Engn & Comp Sci, Belfast BT7 1NN, Antrim, North Ireland
[3] Hunan Normal Univ, Sch Engn & Design, Changsha 410081, Peoples R China
基金
上海市自然科学基金; 中国国家自然科学基金;
关键词
Action recognition; Human body skeleton; Feature construction; Sequence models; Computer vision;
D O I
10.1007/s12652-021-03399-z
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In intelligent video surveillance under complex scenes, it is vital to identify the current actions of multi-target human bodies accurately and in real time. In this paper, a real-time multi-person action recognition method with monocular vision is proposed based on sequence models. Firstly, the key points of multi-target human body skeleton in the video are extracted by using the OpenPose algorithm. Then, the human action features are constructed, including limb direction vector and the skeleton height-width ratio. The multi-target human bodies tracking is then achieved by using the tracking algorithm. Next, the tracking results are matched with the action features, and the action recognition model is constructed, which includes the spatial branch based on Deep neural networks and the temporal branch based on Bi-directional RNN and Bi-directional long short-term memory networks. After pre-training, the model can be used to recognize the human body action from action features, and a recognition stabilizer is designed to minimize false alarms. Finally, extensive evaluations on the JHMDB dataset validate the effectiveness and the superiority of the proposed approach.
引用
收藏
页码:1877 / 1887
页数:11
相关论文
共 50 条
  • [21] A Real-Time Monocular Vision-Based Obstacle Detection
    Wang, Szu-Hong
    Li, Xiang-Xuan
    [J]. 2020 6TH INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND ROBOTICS (ICCAR), 2020, : 695 - 699
  • [22] Monocular Vision Based Real-Time Exploration In Autonomous Rovers
    Mishra, Prabhakar
    Kishore, J. K.
    Shetty, Rakshith
    Malhotra, Akshay
    Kukreja, Rahul
    [J]. 2013 INTERNATIONAL CONFERENCE ON ADVANCES IN COMPUTING, COMMUNICATIONS AND INFORMATICS (ICACCI), 2013, : 42 - 46
  • [23] Real-Time Positioning and Orienting of Pallets Based on Monocular Vision
    Byun, Sungmin
    Kim, Minhwan
    [J]. 20TH IEEE INTERNATIONAL CONFERENCE ON TOOLS WITH ARTIFICIAL INTELLIGENCE, VOL 2, PROCEEDINGS, 2008, : 505 - 508
  • [24] Monocular Vision based Topological Map Generation in Real-time
    Bhowmick, Soumabha
    Deb, Alok Kanti
    Mukhopadhyay, Jayanta
    [J]. 2018 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (IEEE SSCI), 2018, : 791 - 798
  • [25] Real-Time Multi-person Motion Capture from Multi-view Video and IMUs
    Charles Malleson
    John Collomosse
    Adrian Hilton
    [J]. International Journal of Computer Vision, 2020, 128 : 1594 - 1611
  • [26] Smart surveillance system for real-time multi-person multi-camera tracking at the edge
    Gaikwad, Bipin
    Karmakar, Abhijit
    [J]. JOURNAL OF REAL-TIME IMAGE PROCESSING, 2021, 18 (06) : 1993 - 2007
  • [27] Smart surveillance system for real-time multi-person multi-camera tracking at the edge
    Bipin Gaikwad
    Abhijit Karmakar
    [J]. Journal of Real-Time Image Processing, 2021, 18 : 1993 - 2007
  • [28] A Method of Real-time Detection for Distant Moving Obstacles by Monocular Vision
    Jia, Bao-zhi
    Zhu, Ming
    [J]. SIXTH INTERNATIONAL CONFERENCE ON MACHINE VISION (ICMV 2013), 2013, 9067
  • [29] Real-Time Multi-person Motion Capture from Multi-view Video and IMUs
    Malleson, Charles
    Collomosse, John
    Hilton, Adrian
    [J]. INTERNATIONAL JOURNAL OF COMPUTER VISION, 2020, 128 (06) : 1594 - 1611
  • [30] Multi-task neural network with physical constraint for real-time multi-person 3D pose estimation from monocular camera
    Dingli Luo
    Songlin Du
    Takeshi Ikenaga
    [J]. Multimedia Tools and Applications, 2021, 80 : 27223 - 27244