A Guided Method for Improving the Video Human Action Classification in Convolutional Neural Networks

被引:0
|
作者
Mao L. [1 ]
Chen S. [1 ]
Yang D. [1 ]
机构
[1] College of Mechanical and Electronic Engineering, Dalian Minzu University, Dalian
关键词
3D convolutional neural networks; Dynamic information learning ability; Guided optimization; Temporal domain dynamic information understanding ability; Video human action classification;
D O I
10.13203/j.whugis20190101
中图分类号
学科分类号
摘要
Objectives: In order to improve the ability of convolutional neural networks (CNNs) of understanding temporal dynamic information, this paper proposes a dominant layer optimization module. Methods: The new module uses the dominant layer to guide and optimize the update gradient of convolutional layer weights, and assist the difference estimation with the maximum mean difference algorithm of a reproducing Hilbert space. Results: In continuous training, the network can improve the learning ability of temporal dynamic information, and the dynamic information similarity between the features learned by convolutional layer and the input data is also increased. Conclusions: This module enhances the performance of the CNNs model on video human action classification and achieves improvements to the network. © 2021, Editorial Board of Geomatics and Information Science of Wuhan University. All right reserved.
引用
收藏
页码:1241 / 1246
页数:5
相关论文
共 21 条
  • [21] Xie S, Girshick R, Dollar P, Et al., Aggregated Residual Transformations for Deep Neural Networks, IEEE Conference on Computer Vision and Pattern Recognition, (2017)