Human Action Recognition in Videos Using Kinematic Features and Multiple Instance Learning

被引:277
|
作者
Ali, Saad [1 ]
Shah, Mubarak [2 ]
机构
[1] Carnegie Mellon Univ, Inst Robot, Pittsburgh, PA 15213 USA
[2] Univ Cent Florida, Comp Vis Lab, Sch Elect Engn & Comp Sci, Harris Corp Engn Ctr, Orlando, FL 32816 USA
关键词
Action recognition; motion; video analysis; principal component analysis; kinematic features;
D O I
10.1109/TPAMI.2008.284
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We propose a set of kinematic features that are derived from the optical flow for human action recognition in videos. The set of kinematic features includes divergence, vorticity, symmetric and antisymmetric flow fields, second and third principal invariants of flow gradient and rate of strain tensor, and third principal invariant of rate of rotation tensor. Each kinematic feature, when computed from the optical flow of a sequence of images, gives rise to a spatiotemporal pattern. It is then assumed that the representative dynamics of the optical flow are captured by these spatiotemporal patterns in the form of dominant kinematic trends or kinematic modes. These kinematic modes are computed by performing Principal Component Analysis (PCA) on the spatiotemporal volumes of the kinematic features. For classification, we propose the use of multiple instance learning (MIL) in which each action video is represented by a bag of kinematic modes. Each video is then embedded into a kinematic-mode-based feature space and the coordinates of the video in that space are used for classification using the nearest neighbor algorithm. The qualitative and quantitative results are reported on the benchmark data sets.
引用
收藏
页码:288 / 303
页数:16
相关论文
共 50 条
  • [21] Handwritten Digits Recognition Using Multiple Instance Learning
    Yuan Hanning
    Wang Peng
    2013 IEEE INTERNATIONAL CONFERENCE ON GRANULAR COMPUTING (GRC), 2013, : 408 - 411
  • [22] Smart Handheld Based Human Activity Recognition Using Multiple Instance Multiple Label Learning
    Jayita Saha
    Dip Ghosh
    Chandreyee Chowdhury
    Sanghamitra Bandyopadhyay
    Wireless Personal Communications, 2021, 117 : 923 - 943
  • [23] Smart Handheld Based Human Activity Recognition Using Multiple Instance Multiple Label Learning
    Saha, Jayita
    Ghosh, Dip
    Chowdhury, Chandreyee
    Bandyopadhyay, Sanghamitra
    WIRELESS PERSONAL COMMUNICATIONS, 2021, 117 (02) : 923 - 943
  • [24] Recognizing Human Actions From Noisy Videos via Multiple Instance Learning
    Sener, Fadime
    Samet, Nermin
    Duygulu, Pinar
    Ikizler-Cinbis, Nazli
    2013 21ST SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2013,
  • [25] Deep Learning-Based Human Action Recognition in Videos
    Li, Song
    Shi, Qian
    JOURNAL OF CIRCUITS SYSTEMS AND COMPUTERS, 2025, 34 (01)
  • [26] Human Action Recognition Based on Multiple Features and Modified Deep Learning Model
    Zhu, Shaoping
    Xiao, Yongliang
    Ma, Weimin
    INTERNATIONAL JOURNAL OF PATTERN RECOGNITION AND ARTIFICIAL INTELLIGENCE, 2020, 34 (10)
  • [27] Learning Weighted Features for Human Action Recognition
    Zhou, Wen
    Wang, Chunheng
    Xiao, Baihua
    Zhang, Zhong
    Ma, Long
    2012 21ST INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR 2012), 2012, : 1160 - 1163
  • [28] Structured Learning for Action Recognition in Videos
    Long, Yinghan
    Srinivasan, Gopalakrishnan
    Panda, Priyadarshini
    Roy, Kaushik
    IEEE JOURNAL ON EMERGING AND SELECTED TOPICS IN CIRCUITS AND SYSTEMS, 2019, 9 (03) : 475 - 484
  • [29] Multiple Instance Learning for Emotion Recognition Using Physiological Signals
    Romeo, Luca
    Cavallo, Andrea
    Pepa, Lucia
    Bianchi-Berthouze, Nadia
    Pontil, Massimiliano
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2022, 13 (01) : 389 - 407
  • [30] Human Action Recognition from Depth Videos Using Pool of Multiple Projections with Greedy Selection
    Le, Chien-Quang
    Phan, Sang
    Thanh Duc Ngo
    Le, Duy-Dinh
    Satoh, Shin'ichi
    Duc Anh Duong
    IEICE TRANSACTIONS ON INFORMATION AND SYSTEMS, 2016, E99D (08) : 2161 - 2171