Different gait combinations based on multi-modal deep CNN architectures

被引:0
|
作者
Yaprak, Busranur [1 ]
Gedikli, Eyup [2 ]
机构
[1] Gumushane Univ, Dept Software Engn, TR-29100 Gumushane, Turkiye
[2] Karadeniz Tech Univ, Dept Software Engn, TR-61080 Trabzon, Turkiye
关键词
Gait recognition; Multi-modal deep CNN; Gait Combination; GEI; Silhouette; RECOGNITION; FUSION; MOTION; IMAGE;
D O I
10.1007/s11042-024-18859-9
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Gait recognition is the process of identifying a person from a distance based on their walking patterns. However, the recognition rate drops significantly under cross-view angle and appearance-based variations. In this study, the effectiveness of the most well-known gait representations in solving this problem is investigated based on deep learning. For this purpose, a comprehensive performance evaluation is performed by combining different modalities, including silhouettes, optical flows, and concatenated image of the Gait Energy Image (GEI) head and leg region, with GEI itself. This evaluation is carried out across different multimodal deep convolutional neural network (CNN) architectures, namely fine-tuned EfficientNet-B0, MobileNet-V1, and ConvNeXt-base models. These models are trained separately on GEIs, silhouettes, optical flows, and concatenated image of GEI head and leg regions, and then extracted GEI features are fused in pairs with other extracted modality features to find the most effective gait combination. Experimental results on the two different datasets CASIA-B and Outdoor-Gait show that the concatenated image of GEI head and leg regions significantly increased the recognition rate of the networks compared to other modalities. Moreover, this modality demonstrates greater robustness under varied carrying (BG) and clothing (CL) conditions compared to optical flows (OF) and silhouettes (SF). Codes available at https://github.com/busrakckugurlu/Different-gait-combinations-based-on-multi-modal-deep-CNN-architectures.git
引用
收藏
页码:83403 / 83425
页数:23
相关论文
共 50 条
  • [21] Deep Learning Based Multi-modal Registration for Retinal Imaging
    Arikan, Mustafa
    Sadeghipour, Amir
    Gerendas, Bianca
    Told, Reinhard
    Schmidt-Erfurt, Ursula
    INTERPRETABILITY OF MACHINE INTELLIGENCE IN MEDICAL IMAGE COMPUTING AND MULTIMODAL LEARNING FOR CLINICAL DECISION SUPPORT, 2020, 11797 : 75 - 82
  • [22] Multi-Modal Pedestrian Detection Algorithm Based on Deep Learning
    Li X.
    Fu H.
    Niu W.
    Wang P.
    Lü Z.
    Wang W.
    Hsi-An Chiao Tung Ta Hsueh/Journal of Xi'an Jiaotong University, 2022, 56 (10): : 61 - 70
  • [23] A multi-modal dataset for gait recognition under occlusion
    Li, Na
    Zhao, Xinbo
    APPLIED INTELLIGENCE, 2023, 53 (02) : 1517 - 1534
  • [24] Gait and Postural Sway Analysis, A Multi-Modal System
    Ismail, Hafsa
    ICMI'15: PROCEEDINGS OF THE 2015 ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2015, : 629 - 633
  • [25] A multi-modal dataset for gait recognition under occlusion
    Na Li
    Xinbo Zhao
    Applied Intelligence, 2023, 53 : 1517 - 1534
  • [26] A Multi-modal Gait Based Human Identity Recognition System Based on Surveillance Videos
    Hossain, Emdad
    Chetty, Girija
    6TH INTERNATIONAL CONFERENCE ON SIGNAL PROCESSING AND COMMUNICATION SYSTEMS (ICSPCS'2012), 2012,
  • [27] A Multi-Modal Gait Analysis-Based Detection System of the Risk of Depression
    Shao, Wei
    You, Zhiyang
    Liang, Lesheng
    Hu, Xiping
    Li, Chengming
    Wang, Wei
    Hu, Bin
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2022, 26 (10) : 4859 - 4868
  • [28] A deep-CNN based low-cost, multi-modal sensing system for efficient walking activity identification
    Chakraborty, Amartya
    Mukherjee, Nandini
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (11) : 16741 - 16766
  • [29] A deep-CNN based low-cost, multi-modal sensing system for efficient walking activity identification
    Amartya Chakraborty
    Nandini Mukherjee
    Multimedia Tools and Applications, 2023, 82 : 16741 - 16766
  • [30] Multi-input CNN: a deep learning-based approach for predicting breast cancer prognosis using multi-modal data
    Shamita Uma Kandan
    Mariam Mohamed Alketbi
    Zaher Al Aghbari
    Discover Data, 3 (1):