A multi-modal multi-view dataset for human fall analysis and preliminary investigation on modality

被引:0
|
作者
Thanh-Hai Tran [1 ]
Thi-Lan Le [1 ]
Dinh-Tan Pham [1 ,2 ]
Van-Nam Hoang [1 ]
Van-Minh Khong [1 ]
Quoc-Toan Tran [1 ]
Thai-Son Nguyen [3 ]
Cuong Pham [3 ]
机构
[1] Hanoi Univ Sci & Technol, Int Res Inst MICA, Hanoi, Vietnam
[2] Hanoi Univ Min & Geol, Fac Informat Technol, Hanoi, Vietnam
[3] Posts & Telecommun Inst Technol, Fac Informat Technol, Hanoi, Vietnam
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Over the last decade, a large number of methods have been proposed for human fall detection. Most existing methods were evaluated based on trimmed datasets. More importantly, these datasets lack variety of falls, subjects, views and modalities. This paper makes two contributions in the topic of automatic human fall detection. Firstly, to address the above issues, we introduce a large continuous multimodal multivew dataset of human fall, namely CMDFALL. Our CMDFALL dataset was built by capturing activities from 50 subjects, with seven overlapped Kinect sensors and two wearable accelerometers. Each subject performs 20 activities including 8 falls of different styles and 12 daily activities. All multi-modal multi-view data (RGB, depth, skeleton, acceleration) are time-synchronized and annotated for evaluating performance of recognition algorithms of human activities or human fall in indoor environment. Secondly, based on the multimodal property of the dataset, we investigate the role of each modality to get the best results in the context of human activity recognition. To this end, we adopt existing baseline techniques which have been shown to be very efficient for each data modality such as C3D convnet on RGB; DMM-KDES on depth; Res-TCN on skeleton and 2D convnet on acceleration data. We analyze to show which modalities and their combination give the best performance.
引用
收藏
页码:1947 / 1952
页数:6
相关论文
共 50 条
  • [21] Multi-modal multi-view Bayesian semantic embedding for community question answering
    Sang, Lei
    Xu, Min
    Qian, ShengSheng
    Wu, Xindong
    NEUROCOMPUTING, 2019, 334 : 44 - 58
  • [22] MMGCN: Multi-modal multi-view graph convolutional networks for cancer prognosis prediction
    Yang, Ping
    Chen, Wengxiang
    Qiu, Hang
    COMPUTER METHODS AND PROGRAMS IN BIOMEDICINE, 2024, 257
  • [23] Multi-modal Multi-view Clustering based on Non-negative Matrix Factorization
    Khalafaoui, Yasser
    Grozavu, Nistor
    Matei, Basarab
    Goix, Laurent-Walter
    2022 IEEE SYMPOSIUM SERIES ON COMPUTATIONAL INTELLIGENCE (SSCI), 2022, : 1386 - 1391
  • [24] Automatic Medical Image Report Generation with Multi-view and Multi-modal Attention Mechanism
    Yang, Shaokang
    Niu, Jianwei
    Wu, Jiyan
    Liu, Xuefeng
    ALGORITHMS AND ARCHITECTURES FOR PARALLEL PROCESSING, ICA3PP 2020, PT III, 2020, 12454 : 687 - 699
  • [25] Multi-view Multi-modal Person Authentication from a Single Walking Image Sequence
    Muramatsu, Daigo
    Iwama, Haruyuki
    Makihara, Yasushi
    Yagi, Yasushi
    2013 INTERNATIONAL CONFERENCE ON BIOMETRICS (ICB), 2013,
  • [26] H4M: Heterogeneous, Multi-source, Multi-modal, Multi-view and Multi-distributional Dataset for Socioeconomic Analytics in the Case of Beijing
    Zhao, Yaping
    Shi, Shuhui
    Ravi, Ramgopal
    Wang, Zhongrui
    Lam, Edmund Y.
    Zhao, Jichang
    2022 IEEE 9TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA), 2022, : 740 - 749
  • [27] MULTI-VIEW VARIATIONAL RECURRENT NEURAL NETWORK FOR HUMAN EMOTION RECOGNITION USING MULTI-MODAL BIOLOGICAL SIGNALS
    Moroto, Yuya
    Maeda, Keisuke
    Ogawa, Takahiro
    Haseyama, Miki
    2023 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2023, : 2925 - 2929
  • [28] A MULTI-MODAL VIEW OF MEMORY
    HERRMANN, DJ
    SEARLEMAN, A
    BULLETIN OF THE PSYCHONOMIC SOCIETY, 1988, 26 (06) : 503 - 503
  • [29] Beyond Emotion: A Multi-Modal Dataset for Human Desire Understanding
    Jia, Ao
    He, Yu
    Zhang, Yazhou
    Uprety, Sagar
    Song, Dawei
    Lioma, Christina
    NAACL 2022: THE 2022 CONFERENCE OF THE NORTH AMERICAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: HUMAN LANGUAGE TECHNOLOGIES, 2022, : 1512 - 1522
  • [30] Multi Visual Modality Fall Detection Dataset
    Denkovski, Stefan
    Khan, Shehroz S.
    Malamis, Brandon
    Moon, Sae Young
    Ye, Bing
    Mihailidis, Alex
    IEEE ACCESS, 2022, 10 : 106422 - 106435