Hierarchically Learned View-Invariant Representations for Cross-View Action Recognition

被引:43
|
作者
Liu, Yang [1 ]
Lu, Zhaoyang [1 ]
Li, Jing [1 ]
Yang, Tao [2 ]
机构
[1] Xidian Univ, Sch Telecommun Engn, Xian 710071, Shaanxi, Peoples R China
[2] Northwestern Polytech Univ, Sch Comp Sci, Xian 710072, Shaanxi, Peoples R China
基金
中国国家自然科学基金;
关键词
Action recognition; cross-view; dictionary learning; distribution adaptation; SURVEILLANCE;
D O I
10.1109/TCSVT.2018.2868123
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Recognizing human actions from varied views is challenging due to huge appearance variations in different views. The key to this problem is to learn discriminant view-invariant representations generalizing well across views. In this paper, we address this problem by learning view-invariant representations hierarchically using a novel method, referred to as joint sparse representation and distribution adaptation. To obtain robust and informative feature representations, we first incorporate a sample-affinity matrix into the marginalized Stacked Denoising Autoencoder to obtain shared features that are then combined with the private features. In order to make the feature representations of videos across views transferable, we then learn a transferable dictionary pair simultaneously from pairs of videos taken at different views to encourage each action video across views to have the same sparse representation. However, the distribution difference across views still exists because a unified subspace, where the sparse representations of one action across views are the same, may not exist when the view difference is large. Therefore, we propose a novel unsupervised distribution adaptation method that learns a set of projections that project the source and target views data into respective low-dimensional subspaces, where the marginal and conditional distribution differences are reduced simultaneously. Therefore, the finally learned feature representation is view-invariant and robust for substantial distribution difference across views even though the view difference is large. Experimental results on four multi-view datasets show that our approach outperforms the state-of-the-art approaches.
引用
收藏
页码:2416 / 2430
页数:15
相关论文
共 50 条
  • [1] Deeply Learned View-Invariant Features for Cross-View Action Recognition
    Kong, Yu
    Ding, Zhengming
    Li, Jun
    Fu, Yun
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (06) : 3028 - 3037
  • [2] Learning View-invariant Sparse Representations for Cross-view Action Recognition
    Zheng, Jingjing
    Jiang, Zhuolin
    [J]. 2013 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV), 2013, : 3176 - 3183
  • [3] Cross-domain learned view-invariant representation for cross-view action recognition
    Li, Yandi
    Li, Mengdi
    Zhao, Zhihao
    [J]. JOURNAL OF ELECTRONIC IMAGING, 2022, 31 (06)
  • [4] Deep Cross-view Convolutional Features for View-invariant Action Recognition
    Ulhaq, Anwaar
    [J]. 2018 IEEE THIRD INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, APPLICATIONS AND SYSTEMS (IPAS), 2018, : 137 - 142
  • [5] A New View-Invariant Feature for Cross-View Gait Recognition
    Kusakunniran, Worapan
    Wu, Qiang
    Zhang, Jian
    Ma, Yi
    Li, Hongdong
    [J]. IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2013, 8 (10) : 1642 - 1653
  • [6] Global-Local Cross-View Fisher Discrimination for View-invariant Action Recognition
    Gao, Lingling
    Ji, Yanli
    Yang, Yang
    Shen, Heng Tao
    [J]. PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 5255 - 5264
  • [7] Cross-View Action Recognition Using View-Invariant Pose Feature Learned from Synthetic Data with Domain Adaptation
    Yang, Yu-Huan
    Liu, An-Sheng
    Liu, Yu-Hung
    Yeh, Tso-Hsin
    Li, Zi-Jun
    Fu, Li-Chen
    [J]. COMPUTER VISION - ACCV 2018, PT II, 2019, 11362 : 431 - 446
  • [8] Novel Cross-View Human Action Model Recognition Based on the Powerful View-Invariant Features Technique
    Mambou, Sebastien
    Krejcar, Ondrej
    Kuca, Kamil
    Selamat, Ali
    [J]. FUTURE INTERNET, 2018, 10 (09)
  • [9] Unsupervised Learning of View-invariant Action Representations
    Li, Junnan
    Wong, Yongkang
    Zhao, Qi
    Kankanhalli, Mohan S.
    [J]. ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [10] View-invariant Action Recognition in Surveillance Videos
    Zhang, Fang
    Wang, Yunhong
    Zhang, Zhaoxiang
    [J]. 2011 FIRST ASIAN CONFERENCE ON PATTERN RECOGNITION (ACPR), 2011, : 580 - 583