Discovering Motion Primitives for Unsupervised Grouping and One-Shot Learning of Human Actions, Gestures, and Expressions

被引:86
|
作者
Yang, Yang [1 ]
Saleemi, Imran [1 ]
Shah, Mubarak [1 ]
机构
[1] Univ Cent Florida, Dept Elect Engn & Comp Sci EECS, Comp Vis Lab, Orlando, FL 32816 USA
关键词
Human actions; one-shot learning; unsupervised clustering; gestures; facial expressions; action representation; action recognition; motion primitives; motion patterns; histogram of motion primitives; motion primitives strings; Hidden Markov model; RECOGNITION;
D O I
10.1109/TPAMI.2012.253
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes a novel representation of articulated human actions and gestures and facial expressions. The main goals of the proposed approach are: 1) to enable recognition using very few examples, i.e., one or k-shot learning, and 2) meaningful organization of unlabeled datasets by unsupervised clustering. Our proposed representation is obtained by automatically discovering high-level subactions or motion primitives, by hierarchical clustering of observed optical flow in four-dimensional, spatial, and motion flow space. The completely unsupervised proposed method, in contrast to state-of-the-art representations like bag of video words, provides a meaningful representation conducive to visual interpretation and textual labeling. Each primitive action depicts an atomic subaction, like directional motion of limb or torso, and is represented by a mixture of four-dimensional Gaussian distributions. For one-shot and k-shot learning, the sequence of primitive labels discovered in a test video are labeled using KL divergence, and can then be represented as a string and matched against similar strings of training videos. The same sequence can also be collapsed into a histogram of primitives or be used to learn a Hidden Markov model to represent classes. We have performed extensive experiments on recognition by one and k-shot learning as well as unsupervised action clustering on six human actions and gesture datasets, a composite dataset, and a database of facial expressions. These experiments confirm the validity and discriminative nature of the proposed representation.
引用
收藏
页码:1635 / 1648
页数:14
相关论文
共 32 条
  • [21] MMTS: Multimodal Teacher-Student learning for One-Shot Human Action Recognition
    Lee, Jongwhoa
    Sim, Minho
    Choi, Ho-Jin
    2023 IEEE INTERNATIONAL CONFERENCE ON BIG DATA AND SMART COMPUTING, BIGCOMP, 2023, : 235 - 242
  • [22] Audio Metric Learning by Using Siamese Autoencoders for One-Shot Human Fall Detection
    Droghini, Diego
    Squartini, Stefano
    Principi, Emanuele
    Gabrielli, Leonardo
    Piazza, Francesco
    IEEE TRANSACTIONS ON EMERGING TOPICS IN COMPUTATIONAL INTELLIGENCE, 2021, 5 (01): : 108 - 118
  • [23] Learning from Physical Human Feedback: An Object-Centric One-Shot Adaptation Method
    Shek, Alvin
    Su, Bo Ying
    Chen, Rui
    Liu, Changliu
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA 2023), 2023, : 9910 - 9916
  • [24] Human-Like Rule Learning from Images Using One-Shot Hypothesis Derivation
    Varghese, Dany
    Bauer, Roman
    Baxter-Beard, Daniel
    Muggleton, Stephen
    Tamaddoni-Nezhad, Alireza
    INDUCTIVE LOGIC PROGRAMMING (ILP 2021), 2022, 13191 : 234 - 250
  • [25] AnimateAnywhere: Context-Controllable Human Video Generation with ID-Consistent One-shot Learning
    Liu, Hengyuan
    Chen, Xiaodong
    Liu, Xinchen
    Gu, Xiaoyan
    Liu, Wu
    PROCEEDINGS OF THE 5TH INTERNATIONAL WORKSHOP ON HUMAN-CENTRIC MULTIMEDIA ANALYSIS, HCMA 2024, 2024, : 41 - 43
  • [26] Improving WiFi-based Human Activity Recognition with Adaptive Initial State via One-shot Learning
    Ding, Xue
    Jiang, Ting
    Zhong, Yi
    Wu, Sheng
    Yang, Jianfei
    Xue, Wenling
    2021 IEEE WIRELESS COMMUNICATIONS AND NETWORKING CONFERENCE (WCNC), 2021,
  • [27] PLOT: Human-Like Push-Grasping Synergy Learning in Clutter With One-Shot Target Recognition
    Cao, Xiaoge
    Lu, Tao
    Zheng, Liming
    Cai, Yinghao
    Wang, Shuo
    IEEE TRANSACTIONS ON COGNITIVE AND DEVELOPMENTAL SYSTEMS, 2024, 16 (04) : 1391 - 1404
  • [28] Cascaded one-shot deformable convolutional neural networks: Developing a deep learning model for respiratory motion estimation in ultrasound sequences
    Liu, Fei
    Liu, Dan
    Tian, Jie
    Xie, Xiaoyan
    Yang, Xin
    Wang, Kun
    MEDICAL IMAGE ANALYSIS, 2020, 65
  • [29] Affect-DML: Context-Aware One-Shot Recognition of Human Affect using Deep Metric Learning
    Peng, Kunyu
    Roitberg, Alina
    Schneider, David
    Koulakis, Marios
    Yang, Kailun
    Stiefelhagen, Rainer
    2021 16TH IEEE INTERNATIONAL CONFERENCE ON AUTOMATIC FACE AND GESTURE RECOGNITION (FG 2021), 2021,
  • [30] One-shot Learning Classification and Recognition of Gesture Expression From the Egocentric Viewpoint in Intelligent Human-computer Interaction
    Lu Z.
    Qin S.-Y.
    Li L.-W.
    Zhang D.-H.
    Qin, Shi-Yin (qsy@buaa.edu.cn), 2021, Science Press (47): : 1284 - 1301