Discovering Motion Primitives for Unsupervised Grouping and One-Shot Learning of Human Actions, Gestures, and Expressions

被引:86
|
作者
Yang, Yang [1 ]
Saleemi, Imran [1 ]
Shah, Mubarak [1 ]
机构
[1] Univ Cent Florida, Dept Elect Engn & Comp Sci EECS, Comp Vis Lab, Orlando, FL 32816 USA
关键词
Human actions; one-shot learning; unsupervised clustering; gestures; facial expressions; action representation; action recognition; motion primitives; motion patterns; histogram of motion primitives; motion primitives strings; Hidden Markov model; RECOGNITION;
D O I
10.1109/TPAMI.2012.253
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes a novel representation of articulated human actions and gestures and facial expressions. The main goals of the proposed approach are: 1) to enable recognition using very few examples, i.e., one or k-shot learning, and 2) meaningful organization of unlabeled datasets by unsupervised clustering. Our proposed representation is obtained by automatically discovering high-level subactions or motion primitives, by hierarchical clustering of observed optical flow in four-dimensional, spatial, and motion flow space. The completely unsupervised proposed method, in contrast to state-of-the-art representations like bag of video words, provides a meaningful representation conducive to visual interpretation and textual labeling. Each primitive action depicts an atomic subaction, like directional motion of limb or torso, and is represented by a mixture of four-dimensional Gaussian distributions. For one-shot and k-shot learning, the sequence of primitive labels discovered in a test video are labeled using KL divergence, and can then be represented as a string and matched against similar strings of training videos. The same sequence can also be collapsed into a histogram of primitives or be used to learn a Hidden Markov model to represent classes. We have performed extensive experiments on recognition by one and k-shot learning as well as unsupervised action clustering on six human actions and gesture datasets, a composite dataset, and a database of facial expressions. These experiments confirm the validity and discriminative nature of the proposed representation.
引用
收藏
页码:1635 / 1648
页数:14
相关论文
共 32 条
  • [1] Domain-Adaptive Discriminative One-Shot Learning of Gestures
    Pfister, Tomas
    Charles, James
    Zisserman, Andrew
    COMPUTER VISION - ECCV 2014, PT VI, 2014, 8694 : 814 - 829
  • [2] A Bayesian approach to unsupervised one-shot learning of object categories
    Fei-Fei, L
    Fergus, R
    Perona, P
    NINTH IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION, VOLS I AND II, PROCEEDINGS, 2003, : 1134 - 1141
  • [3] An Unsupervised Hierarchical Feature Learning Framework for One-Shot Image Recognition
    Guo, Zhenyu
    Wang, Z. Jane
    IEEE TRANSACTIONS ON MULTIMEDIA, 2013, 15 (03) : 621 - 632
  • [4] Tangential Human Motion Recognition With Micro-Doppler Signatures and One-Shot Learning
    Yang, Yang
    Zhou, Zhengkang
    Li, Beichen
    Li, Junhan
    Lang, Yue
    IEEE SENSORS JOURNAL, 2023, 23 (20) : 24812 - 24821
  • [5] One-Shot Imitation Drone Filming of Human Motion Videos
    Huang, Chong
    Dang, Yuanjie
    Chen, Peng
    Yang, Xin
    Cheng, Kwang-Ting
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (09) : 5335 - 5348
  • [6] Imputing Missing Data in One-Shot Devices Using Unsupervised Learning Approach
    So, Hon Yiu
    Ling, Man Ho
    Balakrishnan, Narayanaswamy
    MATHEMATICS, 2024, 12 (18)
  • [7] Unsupervised One-Shot Learning of Both Specific Instances and Generalised Classes with a Hippocampal Architecture
    Kowadlo, Gideon
    Ahmed, Abdelrahman
    Rawlinson, David
    AI 2020: ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 12576 : 395 - 406
  • [8] Unsupervised vs. transfer learning for multimodal one-shot matching of speech and images
    Nortje, Leanne
    Kamper, Herman
    INTERSPEECH 2020, 2020, : 2712 - 2716
  • [9] Neural Computations Mediating One-Shot Learning in the Human Brain
    Lee, Sang Wan
    O'Doherty, John P.
    Shimojo, Shinsuke
    PLOS BIOLOGY, 2015, 13 (04)
  • [10] Unsupervised Simultaneous Learning of Gestures, Actions and their Associations for Human-Robot Interaction
    Mohammad, Yasser
    Nishida, Toyoaki
    Okada, Shogo
    2009 IEEE-RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS, 2009, : 2537 - 2544