MetaVD: A Meta Video Dataset for enhancing human action recognition datasets

被引:6
|
作者
Yoshikawa, Yuya [1 ]
Shigeto, Yutaro [1 ]
Takeuchi, Akikazu [1 ]
机构
[1] Chiba Inst Technol, Software Technol & Artificial Intelligence Res La, Chiba, Japan
关键词
Human action recognition; Video datasets;
D O I
10.1016/j.cviu.2021.103276
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Numerous practical datasets have been developed to recognize human actions from videos. However, many of them were constructed by collecting videos within a limited domain; thus, a model trained using one of the existing datasets often fails to classify videos in a different domain accurately. A possible solution for this drawback is to enhance the domain of each action label, i.e., to import videos associated with a given action label from the other datasets, and then, to train a model using the enhanced dataset. To realize this solution, we constructed a meta video dataset from the existing datasets for human action recognition, referred to as MetaVD. MetaVD comprises six popular human action recognition datasets, which we integrated by annotating 568,015 relation labels in total. These relation labels reflect equality, similarity, and hierarchy between action labels of the original datasets. We further present simple yet effective dataset enhancement methods using MetaVD, which are useful for training models with higher generalization performance, as established by experiments on human action classification. As a further contribution of MetaVD, we show that its analysis can provide useful insight into the datasets.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Stereoscopic Video Description for Human Action Recognition
    Mademlis, Ioannis
    Iosifidis, Alexandros
    Tefas, Anastasios
    Nikolaidis, Nikos
    Pitas, Ioannis
    2014 IEEE SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE FOR MULTIMEDIA, SIGNAL AND VISION PROCESSING (CIMSIVP), 2014, : 1 - 6
  • [22] Automatic Video Descriptor for Human Action Recognition
    Perera, Minoli
    Farook, Cassim
    Madurapperuma, A. P.
    2017 NATIONAL INFORMATION TECHNOLOGY CONFERENCE (NITC), 2017, : 61 - 66
  • [23] Aeriform in-action: A novel dataset for human action recognition in aerial videos
    Kapoor, Surbhi
    Sharma, Akashdeep
    Verma, Amandeep
    Singh, Sarbjeet
    PATTERN RECOGNITION, 2023, 140
  • [24] Spatio-Temporal Action Localization for Human Action Recognition in Large Dataset
    Megrhi, Sameh
    Jmal, Marwa
    Beghdadi, Azeddine
    Mseddi, Wided
    VIDEO SURVEILLANCE AND TRANSPORTATION IMAGING APPLICATIONS 2015, 2015, 9407
  • [25] The Johns Hopkins University Multimodal Dataset for Human Action Recognition
    Murray, Thomas S.
    Mendat, Daniel R.
    Pouliquen, Philippe O.
    Andreou, Andreas G.
    RADAR SENSOR TECHNOLOGY XIX; AND ACTIVE AND PASSIVE SIGNATURES VI, 2015, 9461
  • [26] Benchmarking a Multimodal and Multiview and Interactive Dataset for Human Action Recognition
    Liu, An-An
    Xu, Ning
    Nie, Wei-Zhi
    Su, Yu-Ting
    Wong, Yongkang
    Kankanhalli, Mohan
    IEEE TRANSACTIONS ON CYBERNETICS, 2017, 47 (07) : 1781 - 1794
  • [27] A large-scale fMRI dataset for human action recognition
    Zhou, Ming
    Gong, Zhengxin
    Dai, Yuxuan
    Wen, Yushan
    Liu, Youyi
    Zhen, Zonglei
    SCIENTIFIC DATA, 2023, 10 (01)
  • [28] A large-scale fMRI dataset for human action recognition
    Ming Zhou
    Zhengxin Gong
    Yuxuan Dai
    Yushan Wen
    Youyi Liu
    Zonglei Zhen
    Scientific Data, 10
  • [29] Neural Architecture Search for Enhancing Action Video Recognition in Compressed Domains
    Lamkowski, Pedro
    Rodrigues, Douglas
    Passos, Leandro A.
    Papa, Joao P.
    Almeida, Jurandy
    2024 31ST INTERNATIONAL CONFERENCE ON SYSTEMS, SIGNALS AND IMAGE PROCESSING, IWSSIP 2024, 2024,
  • [30] Okutama-Action: An Aerial View Video Dataset for Concurrent Human Action Detection
    Barekatain, Mohammadamin
    Marti, Miquel
    Shih, Hsueh-Fu
    Murray, Samuel
    Nakayama, Kotaro
    Matsuo, Yutaka
    Prendinger, Helmut
    2017 IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW), 2017, : 2153 - 2160