Learning View-invariant Sparse Representations for Cross-view Action Recognition

被引:55
|
作者
Zheng, Jingjing [1 ]
Jiang, Zhuolin [2 ]
机构
[1] Univ Maryland, College Pk, MD 20742 USA
[2] Huawei Technol, Noahs Ark Lab, Shenzhen, Peoples R China
关键词
D O I
10.1109/ICCV.2013.394
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
We present an approach to jointly learn a set of view-specific dictionaries and a common dictionary for cross-view action recognition. The set of view-specific dictionaries is learned for specific views while the common dictionary is shared across different views. Our approach represents videos in each view using both the corresponding view-specific dictionary and the common dictionary. More importantly, it encourages the set of videos taken from different views of the same action to have similar sparse representations. In this way, we can align view-specific features in the sparse feature spaces spanned by the view-specific dictionary set and transfer the view-shared features in the sparse feature space spanned by the common dictionary. Meanwhile, the incoherence between the common dictionary and the view-specific dictionary set enables us to exploit the discrimination information encoded in view-specific features and view-shared features separately. In addition, the learned common dictionary not only has the capability to represent actions from unseen views, but also makes our approach effective in a semi-supervised setting where no correspondence videos exist and only a few labels exist in the target view. Extensive experiments using the multi-view IXMAS dataset demonstrate that our approach outperforms many recent approaches for cross-view action recognition.
引用
收藏
页码:3176 / 3183
页数:8
相关论文
共 50 条
  • [21] View-invariant representation and learning of human action
    Rao, C
    Shah, M
    [J]. IEEE WORKSHOP ON DETECTION AND RECOGNITION OF EVENTS IN VIDEO, PROCEEDINGS, 2001, : 55 - 63
  • [22] View-Invariant Action Recognition from Point Triplets
    Shen, Yuping
    Foroosh, Hassan
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2009, 31 (10) : 1898 - 1905
  • [23] View-invariant action recognition using Fundamental Ratios
    Shen, Yuping
    Foroosh, Hassan
    [J]. 2008 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, VOLS 1-12, 2008, : 3216 - 3221
  • [24] CROSS-VIEW ACTION RECOGNITION VIA TRANSDUCTIVE TRANSFER LEARNING
    Qin, Jie
    Zhang, Zhaoxiang
    Wang, Yunhong
    [J]. 2013 20TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP 2013), 2013, : 3582 - 3586
  • [25] Learning Discriminative Transferable Sparse Coding for Cross-View Action Recognition in Wireless Sensor Networks
    Zhang, Zhong
    Liu, Shuang
    [J]. INTERNATIONAL JOURNAL OF DISTRIBUTED SENSOR NETWORKS, 2015,
  • [26] Towards Fast, View-Invariant Human Action Recognition
    Cherla, Srikanth
    Kulkarni, Kaustubh
    Kale, Amit
    Ramasubramanian, V.
    [J]. 2008 IEEE COMPUTER SOCIETY CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, VOLS 1-3, 2008, : 1650 - 1657
  • [27] Cross-View Action Recognition via Transferable Dictionary Learning
    Zheng, Jingjing
    Jiang, Zhuolin
    Chellappa, Rama
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2016, 25 (06) : 2542 - 2556
  • [28] Cross-View Action Recognition via View Knowledge Transfer
    Liu, Jingen
    Shah, Mubarak
    Kuipers, Benjamin
    Savarese, Silvio
    [J]. 2011 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2011,
  • [29] On Temporal Order Invariance for View-Invariant Action Recognition
    Anwaar-ul-Haq
    Gondal, Iqbal
    Murshed, Manzur
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2013, 23 (02) : 203 - 211
  • [30] A survey about view-invariant human action recognition
    Nghia Pham Trong
    Anh Truong Minh
    Nguyen, Hung
    Kazunori, Kotani
    Bac Le Hoai
    [J]. 2017 56TH ANNUAL CONFERENCE OF THE SOCIETY OF INSTRUMENT AND CONTROL ENGINEERS OF JAPAN (SICE), 2017, : 699 - 704