Learning multi-temporal-scale deep information for action recognition

被引:25
|
作者
Yao, Guangle [1 ,2 ,3 ]
Lei, Tao [1 ]
Zhong, Jiandan [1 ,2 ,3 ]
Jiang, Ping [1 ]
机构
[1] Chinese Acad Sci, Inst Opt & Elect, Chengdu, Sichuan, Peoples R China
[2] Univ Elect Sci & Technol China, Chengdu, Sichuan, Peoples R China
[3] Univ Chinese Acad Sci, Beijing, Peoples R China
关键词
Action recognition; Convolutional neural networks; Deep learning; Spatiotemporal information; HISTOGRAMS; NETWORKS;
D O I
10.1007/s10489-018-1347-3
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Action recognition in video is widely applied in video indexing, intelligent surveillance, multimedia understanding, and other fields. A typical human action contains the spatiotemporal information from various scales. Learning and fusing the multi-temporal-scale information make action recognition more reliable in terms of recognition accuracy. To demonstrate this argument, in this paper, we use Res3D, a 3D Convolution Neural Network (CNN) architecture, to extract information in multiple temporal scales. And in each temporal scale, we transfer the knowledge learned from RGB to 3-channel optical flow (OF) and learn information from RGB and OF fields. We also propose Parallel Pair Discriminant Correlation Analysis (PPDCA) to fuse the multi-temporal-scale information into action representation with a lower dimension. Experimental results show that compared with single-temporal-scale method, the proposed multi-temporal-scale method gains higher recognition accuracy, and spends more time on feature extraction, but less time on classification due to the representation with lower dimension. Moreover, the proposed method achieves recognition performance comparable to that of the state-of-the-art methods. The source code and 3D filter animations are available online: https://github.com/JerryYaoGl/multi-temporal-scale.
引用
收藏
页码:2017 / 2029
页数:13
相关论文
共 50 条
  • [1] Learning multi-temporal-scale deep information for action recognition
    Guangle Yao
    Tao Lei
    Jiandan Zhong
    Ping Jiang
    Applied Intelligence, 2019, 49 : 2017 - 2029
  • [2] Temporal Modeling on Multi-Temporal-Scale Spatiotemporal Atoms for Action Recognition
    Yao, Guangle
    Lei, Tao
    Liu, Xianyuan
    Jiang, Ping
    APPLIED SCIENCES-BASEL, 2018, 8 (10):
  • [3] Action Recognition with Temporal Scale-Invariant Deep Learning Framework
    Chen, Huafeng
    Chen, Jun
    Hu, Ruimin
    Chen, Chen
    Wang, Zhongyuan
    CHINA COMMUNICATIONS, 2017, 14 (02) : 163 - 172
  • [4] Action Recognition with Temporal Scale-Invariant Deep Learning Framework
    Huafeng Chen
    Jun Chen
    Ruimin Hu
    Chen Chen
    Zhongyuan Wang
    China Communications, 2017, 14 (02) : 163 - 172
  • [5] Multi-scale spatialtemporal information deep fusion network with temporal pyramid mechanism for video action recognition
    Ou, Hongshi
    Sun, Jifeng
    JOURNAL OF INTELLIGENT & FUZZY SYSTEMS, 2021, 41 (03) : 4533 - 4545
  • [6] Spatio-Temporal Information for Action Recognition in Thermal Video Using Deep Learning Model
    Srihari, P.
    Harikiran, J.
    INTERNATIONAL JOURNAL OF ELECTRICAL AND COMPUTER ENGINEERING SYSTEMS, 2022, 13 (08) : 669 - 680
  • [7] Action Recognition by Learning Deep Multi-Granular Spatio-Temporal Video Representation
    Li, Qing
    Qiu, Zhaofan
    Yao, Ting
    Mei, Tao
    Rui, Yong
    Luo, Jiebo
    ICMR'16: PROCEEDINGS OF THE 2016 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2016, : 159 - 166
  • [8] Improved SSD using deep multi-scale attention spatial–temporal features for action recognition
    Shuren Zhou
    Jia Qiu
    Arun Solanki
    Multimedia Systems, 2022, 28 : 2123 - 2131
  • [9] MULTI-SCALE TEMPORAL INFORMATION EXTRACTOR FOR GAIT RECOGNITION
    Lin, Beibei
    Zhang, Shunli
    Liu, Yu
    Qin, Shengdi
    2021 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2021, : 2998 - 3002
  • [10] Deep Fusion of Skeleton Spatial–Temporal and Dynamic Information for Action Recognition
    Gao, Song
    Zhang, Dingzhuo
    Tang, Zhaoming
    Wang, Hongyan
    Sensors, 2024, 24 (23)