Structure-Preserving Binary Representations for RGB-D Action Recognition

被引:81
|
作者
Yu, Mengyang [1 ]
Liu, Li [1 ]
Shao, Ling [1 ]
机构
[1] Northumbria Univ, Dept Comp Sci & Digital Technol, Newcastle Upon Tyne NE1 8ST, Tyne & Wear, England
关键词
RGB-D fusion; flux; binary; structure-preserving; dimensionality reduction; local feature; DIMENSIONALITY; FEATURES;
D O I
10.1109/TPAMI.2015.2491925
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In this paper, we propose a novel binary local representation for RGB-D video data fusion with a structure-preserving projection. Our contribution consists of two aspects. Toacquire a general feature for the video data, we convert the problem to describing the gradient fields of RGB and depth information of video sequences. With the local fluxes of the gradient fields, which include the orientation and the magnitude of the neighborhood of each point, a new kind of continuous local descriptor called Local Flux Feature(LFF) is obtained. Then the LFFs from RGB and depth channels are fused into a Hamming space via the Structure Preserving Projection (SPP). Specifically, an orthogonal projection matrix is applied to preserve the pairwise structure with a shape constraint to avoid the collapse of data structure in the projected space. Furthermore, a bipartite graph structure of data is taken into consideration, which is regarded as a higher level connection between samples and classes than the pairwise structure of local features. The extensive experiments show not only the high efficiency of binary codes and the effectiveness of combining LFFs from RGB-D channels via SPP on various action recognition benchmarks of RGB-D data, but also the potential power of LFF for general action recognition.
引用
收藏
页码:1651 / 1664
页数:14
相关论文
共 50 条
  • [1] Child Action Recognition in RGB and RGB-D Data
    Turarova, Aizada
    Zhanatkyzy, Aida
    Telisheva, Zhansaule
    Sabyrov, Arman
    Sandygulova, Anara
    [J]. HRI'20: COMPANION OF THE 2020 ACM/IEEE INTERNATIONAL CONFERENCE ON HUMAN-ROBOT INTERACTION, 2020, : 491 - 492
  • [2] ACTION RECOGNITION IN RGB-D EGOCENTRIC VIDEOS
    Tang, Yansong
    Tian, Yi
    Lu, Jiwen
    Feng, Jianjiang
    Zhou, Jie
    [J]. 2017 24TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2017, : 3410 - 3414
  • [3] Structured Images for RGB-D Action Recognition
    Wang, Pichao
    Wang, Shuang
    Gao, Zhimin
    Hou, Yonghong
    Li, Wanqing
    [J]. 2017 IEEE INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW 2017), 2017, : 1005 - 1014
  • [4] Learning Effective RGB-D Representations for Scene Recognition
    Song, Xinhang
    Jiang, Shuqiang
    Herranz, Luis
    Chen, Chengpeng
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (02) : 980 - 993
  • [5] Fusion of Skeleton and RGB Features for RGB-D Human Action Recognition
    Weiyao, Xu
    Muqing, Wu
    Min, Zhao
    Ting, Xia
    [J]. IEEE SENSORS JOURNAL, 2021, 21 (17) : 19157 - 19164
  • [6] RGB-D action recognition based on discriminative common structure learning model
    Liu, Tianshan
    Kong, Jun
    Jiang, Min
    Huo, Hongtao
    [J]. JOURNAL OF ELECTRONIC IMAGING, 2019, 28 (02)
  • [7] Deep Bilinear Learning for RGB-D Action Recognition
    Hu, Jian-Fang
    Zheng, Wei-Shi
    Pan, Jiahui
    Lai, Jianhuang
    Zhang, Jianguo
    [J]. COMPUTER VISION - ECCV 2018, PT VII, 2018, 11211 : 346 - 362
  • [8] Learning local representations for scalable RGB-D face recognition
    Grati, Nesrine
    Ben-Hamadou, Achraf
    Hammami, Mohamed
    [J]. EXPERT SYSTEMS WITH APPLICATIONS, 2020, 150
  • [9] Joint Deep Learning for RGB-D Action Recognition
    Qin, Xiaolei
    Ge, Yongxin
    Zhan, Liuwei
    Li, Guangrui
    Huang, Sheng
    Wang, Hongxing
    Chen, Feiyu
    Wang, Hongxing
    [J]. 2018 IEEE INTERNATIONAL CONFERENCE ON VISUAL COMMUNICATIONS AND IMAGE PROCESSING (IEEE VCIP), 2018,
  • [10] Viewpoint Invariant RGB-D Human Action Recognition
    Liu, Jian
    Akhtar, Naveed
    Mian, Ajmal
    [J]. 2017 INTERNATIONAL CONFERENCE ON DIGITAL IMAGE COMPUTING - TECHNIQUES AND APPLICATIONS (DICTA), 2017, : 261 - 268