Exploring probabilistic localized video representation for human action recognition

被引:0
|
作者
Yan Song
Sheng Tang
Yan-Tao Zheng
Tat-Seng Chua
Yongdong Zhang
Shouxun Lin
机构
[1] Chinese Academy of Sciences,Laboratory of Advanced Computing Research, Institute of Computing Technology
[2] Graduate University of the Chinese Academy of Sciences,School of Computing
[3] Institute for Infocomm Research,undefined
[4] A*STAR,undefined
[5] National University of Singapore,undefined
来源
关键词
Human action recognition; Probabilistic video representation; Information-theoretic video matching;
D O I
暂无
中图分类号
学科分类号
摘要
In recent years, the bag-of-words (BoW) video representations have achieved promising results in human action recognition in videos. By vector quantizing local spatial temporal (ST) features, the BoW video representation brings in simplicity and efficiency, but limitations too. First, the discretization of feature space in BoW inevitably results in ambiguity and information loss in video representation. Second, there exists no universal codebook for BoW representation. The codebook needs to be re-built when video corpus is changed. To tackle these issues, this paper explores a localized, continuous and probabilistic video representation. Specifically, the proposed representation encodes the visual and motion information of an ensemble of local ST features of a video into a distribution estimated by a generative probabilistic model. Furthermore, the probabilistic video representation naturally gives rise to an information-theoretic distance metric of videos. This makes the representation readily applicable to most discriminative classifiers, such as the nearest neighbor schemes and the kernel based classifiers. Experiments on two datasets, KTH and UCF sports, show that the proposed approach could deliver promising results.
引用
收藏
页码:663 / 685
页数:22
相关论文
共 50 条
  • [21] A robust algorithm for probabilistic human recognition from video
    Zhou, SH
    Chellappa, R
    [J]. 16TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL I, PROCEEDINGS, 2002, : 226 - 229
  • [22] Video sketch: A middle-level representation for action recognition
    Xing-Yuan Zhang
    Ya-Ping Huang
    Yang Mi
    Yan-Ting Pei
    Qi Zou
    Song Wang
    [J]. Applied Intelligence, 2021, 51 : 2589 - 2608
  • [23] Space-Time Robust Video Representation for Action Recognition
    Ballas, Nicolas
    Yang, Yi
    Lan, Zhen-zhong
    Delezoide, Betrand
    Preteux, Francoise
    Hauptmann, Alex
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 2704 - 2711
  • [24] A component-based video content representation for action recognition
    Adeli, Vida
    Fazl-Ersi, Ehsan
    Harati, Ahad
    [J]. IMAGE AND VISION COMPUTING, 2019, 90
  • [25] Video sketch: A middle-level representation for action recognition
    Zhang, Xing-Yuan
    Huang, Ya-Ping
    Mi, Yang
    Pei, Yan-Ting
    Zou, Qi
    Wang, Song
    [J]. APPLIED INTELLIGENCE, 2021, 51 (04) : 2589 - 2608
  • [26] Human Action Recognition using Sparse Representation
    Liu, Changhong
    Yang, Yang
    Chen, Yong
    [J]. 2009 IEEE INTERNATIONAL CONFERENCE ON INTELLIGENT COMPUTING AND INTELLIGENT SYSTEMS, PROCEEDINGS, VOL 4, 2009, : 184 - +
  • [27] Using SAX representation for human action recognition
    Junejo, Imran N.
    Al Aghbari, Zaher
    [J]. JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2012, 23 (06) : 853 - 861
  • [28] Dynamic Motion Representation for Human Action Recognition
    Asghari-Esfeden, Sadjad
    Sznaier, Mario
    Camps, Octavia
    [J]. 2020 IEEE WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2020, : 546 - 555
  • [29] Structured sparse representation for human action recognition
    Moayedi, F.
    Azimifara, Z.
    Boostani, R.
    [J]. NEUROCOMPUTING, 2015, 161 : 38 - 46
  • [30] Video Analytics Framework for Human Action Recognition
    Khan, Muhammad Attique
    Alhaisoni, Majed
    Armghan, Ammar
    Alenezi, Fayadh
    Tariq, Usman
    Nam, Yunyoung
    Akram, Tallha
    [J]. CMC-COMPUTERS MATERIALS & CONTINUA, 2021, 68 (03): : 3841 - 3859