Converting video classification problem to image classification with global descriptors and pre-trained network

被引:7
|
作者
Zebhi, Saeedeh [1 ]
Al-Modarresi, S. M. T. [1 ]
Abootalebi, Vahid [1 ]
机构
[1] Yazd Univ, Elect Engn Dept, Yazd, Iran
关键词
ACTION RECOGNITION;
D O I
10.1049/iet-cvi.2019.0625
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Motion history image (MHI) is a spatio-temporal template that temporal motion information is collapsed into a single image where intensity is a function of recency of motion. Also, it consists of spatial information. Energy image (EI) based on the magnitude of optical flow is a temporal template that shows only temporal information of motion. Each video can be described in these templates. So, four new methods are introduced in this study. The first three methods are called basic methods. In method 1, each video splits into N groups of consecutive frames and MHI is calculated for each group. Transfer learning with fine-tuning technique has been used for classifying these templates. EIs are used for classifying in method 2 similar to method 1. Fusing two streams of these templates is introduced as method 3. Finally, spatial information is added in method 4. Among these methods, method 4 outperforms others and it is called the proposed method. It achieves the recognition accuracy of 92.30 and 94.50% for UCF Sport and UCF-11 action data sets, respectively. Also, the proposed method is compared with the state-of-the-art approaches and the results show that it has the best performance.
引用
收藏
页码:614 / 624
页数:11
相关论文
共 50 条
  • [31] Comparative Ship Classification in Heterogeneous Dataset with Pre-trained Models
    Tienin, Bole Wilfried
    Cui Guolong
    Esidang, Roldan Mba
    2022 IEEE RADAR CONFERENCE (RADARCONF'22), 2022,
  • [32] A survey of text classification based on pre-trained language model
    Wu, Yujia
    Wan, Jun
    NEUROCOMPUTING, 2025, 616
  • [33] Pre-trained Language Models with Limited Data for Intent Classification
    Kasthuriarachchy, Buddhika
    Chetty, Madhu
    Karmakar, Gour
    Walls, Darren
    2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [34] Tabular Cell Classification Using Pre-Trained Cell Embeddings
    Ghasemi-Gol, Majid
    Pujara, Jay
    Szekely, Pedro
    2019 19TH IEEE INTERNATIONAL CONFERENCE ON DATA MINING (ICDM 2019), 2019, : 230 - 239
  • [35] Speech Topic Classification Based on Pre-trained and Graph Networks
    Niu, Fangjing
    Cao, Tengfei
    Hu, Ying
    Huang, Hao
    He, Liang
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1721 - 1726
  • [36] Spanish Pre-Trained CaTrBETO Model for Sentiment Classification in Twitter
    Pijal, Washington
    Armijos, Arianna
    Llumiquinga, Jose
    Lalvay, Sebastian
    Allauca, Steven
    Cuenca, Erick
    2022 THIRD INTERNATIONAL CONFERENCE ON INFORMATION SYSTEMS AND SOFTWARE TECHNOLOGIES, ICI2ST, 2022, : 93 - 98
  • [37] Focused Contrastive Loss for Classification With Pre-Trained Language Models
    He, Jiayuan
    Li, Yuan
    Zhai, Zenan
    Fang, Biaoyan
    Thorne, Camilo
    Druckenbrodt, Christian
    Akhondi, Saber
    Verspoor, Karin
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (07) : 3047 - 3061
  • [38] Performance Evaluation of CNN and Pre-trained Models for Malware Classification
    Omar Habibi
    Mohammed Chemmakha
    Mohamed Lazaar
    Arabian Journal for Science and Engineering, 2023, 48 : 10355 - 10369
  • [39] Performance Evaluation of CNN and Pre-trained Models for Malware Classification
    Habibi, Omar
    Chemmakha, Mohammed
    Lazaar, Mohamed
    ARABIAN JOURNAL FOR SCIENCE AND ENGINEERING, 2023, 48 (08) : 10355 - 10369
  • [40] Analyzing the Potential of Pre-Trained Embeddings for Audio Classification Tasks
    Grollmisch, Sascha
    Cano, Estefania
    Kehling, Christian
    Taenzer, Michael
    28TH EUROPEAN SIGNAL PROCESSING CONFERENCE (EUSIPCO 2020), 2021, : 790 - 794