Dense saliency-based spatiotemporal feature points for action recognition

被引:0
|
作者
Rapantzikos, Konstantinos [1 ]
Avrithis, Yannis [1 ]
Kollias, Stefanos [1 ]
机构
[1] Natl Tech Univ Athens, Sch Elect & Comp Engn, GR-10682 Athens, Greece
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Several spatiotemporal feature point detectors have been recently used in video analysis for action recognition. Feature points are detected using a number of measures, namely saliency, cornerness, periodicity, motion activity etc. Each of these measures is usually intensity-based and provides a different trade-off between density and informativeness. In this paper, we use saliency for feature point detection in videos and incorporate color and motion apart from intensity. Our method uses a multi-scale volumetric representation of the video and involves spatiotemporal operations at the voxel level. Saliency is computed by a global minimization process constrained by pure volumetric constraints, each of them being related to an informative visual aspect, namely spatial proximity, scale and feature similarity (intensity, color, motion). Points are selected as the extrema of the saliency response and prove to balance well between density and informativeness. We provide an intuitive view of the detected points and visual comparisons against state-of-the-art space-time detectors. Our detector outperforms them on the KTH dataset using Nearest-Neighbor classifiers and ranks among the top using different classification frameworks. Statistics and comparisons are also performed on the more difficult Hollywood Human Actions (HOHA) dataset increasing the performance compared to current published results.
引用
收藏
页码:1454 / 1461
页数:8
相关论文
共 50 条
  • [1] Action recognition by saliency-based dense sampling
    Xu, Zengmin
    Hu, Ruimin
    Chen, Jun
    Chen, Chen
    Chen, Huafeng
    Li, Hongyang
    Sun, Qingquan
    NEUROCOMPUTING, 2017, 236 : 82 - 92
  • [2] SALIENCY-BASED SELECTION OF SPARSE DESCRIPTORS FOR ACTION RECOGNITION
    Vig, Eleonora
    Dorr, Michael
    Cox, David D.
    2012 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2012), 2012, : 1405 - 1408
  • [3] Saliency-based dense trajectories for action recognition using low-rank matrix decomposition
    Wang, Xiaofang
    Qi, Chun
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2016, 41 : 361 - 374
  • [4] Spatiotemporal saliency-based multi-stream networks with attention-aware LSTM for action recognition
    Liu, Zhenbing
    Li, Zeya
    Wang, Ruili
    Zong, Ming
    Ji, Wanting
    NEURAL COMPUTING & APPLICATIONS, 2020, 32 (18): : 14593 - 14602
  • [5] Spatiotemporal saliency-based multi-stream networks with attention-aware LSTM for action recognition
    Zhenbing Liu
    Zeya Li
    Ruili Wang
    Ming Zong
    Wanting Ji
    Neural Computing and Applications, 2020, 32 : 14593 - 14602
  • [6] Saliency-Based Spatiotemporal Attention for Video Captioning
    Chen, Yangyu
    Zhang, Weigang
    Wang, Shuhui
    Li, Liang
    Huang, Qingming
    2018 IEEE FOURTH INTERNATIONAL CONFERENCE ON MULTIMEDIA BIG DATA (BIGMM), 2018,
  • [7] A saliency-based approach to event recognition
    Ahmad, Kashif
    Conci, Nicola
    De Natale, F. G. B.
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2018, 60 : 42 - 51
  • [8] Spatiotemporal saliency for human action recognition
    Oikonomopoulos, A
    Patras, I
    Pantic, M
    2005 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO (ICME), VOLS 1 AND 2, 2005, : 430 - 433
  • [9] SPARSE CODING-BASED SPATIOTEMPORAL SALIENCY FOR ACTION RECOGNITION
    Zhang, Tao
    Xu, Long
    Yang, Jie
    Shi, Pengfei
    Jia, Wenjing
    2015 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2015, : 2045 - 2049
  • [10] Saliency-based framework for facial expression recognition
    Rizwan Ahmed Khan
    Alexandre Meyer
    Hubert Konik
    Saida Bouakaz
    Frontiers of Computer Science, 2019, 13 : 183 - 198