A multi-modal multi-view dataset for human fall analysis and preliminary investigation on modality

被引:0
|
作者
Thanh-Hai Tran [1 ]
Thi-Lan Le [1 ]
Dinh-Tan Pham [1 ,2 ]
Van-Nam Hoang [1 ]
Van-Minh Khong [1 ]
Quoc-Toan Tran [1 ]
Thai-Son Nguyen [3 ]
Cuong Pham [3 ]
机构
[1] Hanoi Univ Sci & Technol, Int Res Inst MICA, Hanoi, Vietnam
[2] Hanoi Univ Min & Geol, Fac Informat Technol, Hanoi, Vietnam
[3] Posts & Telecommun Inst Technol, Fac Informat Technol, Hanoi, Vietnam
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Over the last decade, a large number of methods have been proposed for human fall detection. Most existing methods were evaluated based on trimmed datasets. More importantly, these datasets lack variety of falls, subjects, views and modalities. This paper makes two contributions in the topic of automatic human fall detection. Firstly, to address the above issues, we introduce a large continuous multimodal multivew dataset of human fall, namely CMDFALL. Our CMDFALL dataset was built by capturing activities from 50 subjects, with seven overlapped Kinect sensors and two wearable accelerometers. Each subject performs 20 activities including 8 falls of different styles and 12 daily activities. All multi-modal multi-view data (RGB, depth, skeleton, acceleration) are time-synchronized and annotated for evaluating performance of recognition algorithms of human activities or human fall in indoor environment. Secondly, based on the multimodal property of the dataset, we investigate the role of each modality to get the best results in the context of human activity recognition. To this end, we adopt existing baseline techniques which have been shown to be very efficient for each data modality such as C3D convnet on RGB; DMM-KDES on depth; Res-TCN on skeleton and 2D convnet on acceleration data. We analyze to show which modalities and their combination give the best performance.
引用
收藏
页码:1947 / 1952
页数:6
相关论文
共 50 条
  • [31] Multi-View Multi-Modal Head-Gaze Estimation for Advanced Indoor User Interaction
    Kim, Jung-Hwa
    Jeong, Jin-Woo
    CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 70 (03): : 5107 - 5132
  • [32] MMNeRF: Multi-Modal and Multi-View Optimized Cross-Scene Neural Radiance Fields
    Zhang, Qi
    Wang, Bo Han
    Yang, Ming Chuan
    Zou, Hang
    IEEE ACCESS, 2023, 11 : 27401 - 27413
  • [33] Graph-Based Multi-Modal Multi-View Fusion for Facial Action Unit Recognition
    Chen, Jianrong
    Dey, Sujit
    IEEE ACCESS, 2024, 12 : 69310 - 69324
  • [34] MULTI-VIEW AND MULTI-MODAL EVENT DETECTION UTILIZING TRANSFORMER-BASED MULTI-SENSOR FUSION
    Yasuda, Masahiro
    Ohishi, Yasunori
    Saito, Shoichiro
    Harado, Noboru
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 4638 - 4642
  • [35] Cross Modality Knowledge Distillation for Multi-modal Aerial View Object Classification
    Yang, Lehan
    Xu, Kele
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 382 - 387
  • [36] A multi-modal open dataset for mental-disorder analysis
    Cai, Hanshu
    Yuan, Zhenqin
    Gao, Yiwen
    Sun, Shuting
    Li, Na
    Tian, Fuze
    Xiao, Han
    Li, Jianxiu
    Yang, Zhengwu
    Li, Xiaowei
    Zhao, Qinglin
    Liu, Zhenyu
    Yao, Zhijun
    Yang, Minqiang
    Peng, Hong
    Zhu, Jing
    Zhang, Xiaowei
    Gao, Guoping
    Zheng, Fang
    Li, Rui
    Guo, Zhihua
    Ma, Rong
    Yang, Jing
    Zhang, Lan
    Hu, Xiping
    Li, Yumin
    Hu, Bin
    SCIENTIFIC DATA, 2022, 9 (01)
  • [37] A multi-modal open dataset for mental-disorder analysis
    Hanshu Cai
    Zhenqin Yuan
    Yiwen Gao
    Shuting Sun
    Na Li
    Fuze Tian
    Han Xiao
    Jianxiu Li
    Zhengwu Yang
    Xiaowei Li
    Qinglin Zhao
    Zhenyu Liu
    Zhijun Yao
    Minqiang Yang
    Hong Peng
    Jing Zhu
    Xiaowei Zhang
    Guoping Gao
    Fang Zheng
    Rui Li
    Zhihua Guo
    Rong Ma
    Jing Yang
    Lan Zhang
    Xiping Hu
    Yumin Li
    Bin Hu
    Scientific Data, 9
  • [38] Modality Mixer for Multi-modal Action Recognition
    Lee, Sumin
    Woo, Sangmin
    Park, Yeonju
    Nugroho, Muhammad Adi
    Kim, Changick
    2023 IEEE/CVF WINTER CONFERENCE ON APPLICATIONS OF COMPUTER VISION (WACV), 2023, : 3297 - 3306
  • [39] Collaborative recommendation model based on multi-modal multi-view attention network: Movie and literature cases
    Hu, Zheng
    Cai, Shi-Min
    Wang, Jun
    Zhou, Tao
    APPLIED SOFT COMPUTING, 2023, 144
  • [40] M3V: Multi-modal Multi-view Context Embedding for Repair Operator Prediction
    Xu, Xuezheng
    Wang, Xudong
    Xue, Jingling
    CGO 2022 - Proceedings of the 2022 IEEE/ACM International Symposium on Code Generation and Optimization, 2022, : 266 - 277