Different gait combinations based on multi-modal deep CNN architectures

被引:0
|
作者
Yaprak, Busranur [1 ]
Gedikli, Eyup [2 ]
机构
[1] Gumushane Univ, Dept Software Engn, TR-29100 Gumushane, Turkiye
[2] Karadeniz Tech Univ, Dept Software Engn, TR-61080 Trabzon, Turkiye
关键词
Gait recognition; Multi-modal deep CNN; Gait Combination; GEI; Silhouette; RECOGNITION; FUSION; MOTION; IMAGE;
D O I
10.1007/s11042-024-18859-9
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Gait recognition is the process of identifying a person from a distance based on their walking patterns. However, the recognition rate drops significantly under cross-view angle and appearance-based variations. In this study, the effectiveness of the most well-known gait representations in solving this problem is investigated based on deep learning. For this purpose, a comprehensive performance evaluation is performed by combining different modalities, including silhouettes, optical flows, and concatenated image of the Gait Energy Image (GEI) head and leg region, with GEI itself. This evaluation is carried out across different multimodal deep convolutional neural network (CNN) architectures, namely fine-tuned EfficientNet-B0, MobileNet-V1, and ConvNeXt-base models. These models are trained separately on GEIs, silhouettes, optical flows, and concatenated image of GEI head and leg regions, and then extracted GEI features are fused in pairs with other extracted modality features to find the most effective gait combination. Experimental results on the two different datasets CASIA-B and Outdoor-Gait show that the concatenated image of GEI head and leg regions significantly increased the recognition rate of the networks compared to other modalities. Moreover, this modality demonstrates greater robustness under varied carrying (BG) and clothing (CL) conditions compared to optical flows (OF) and silhouettes (SF). Codes available at https://github.com/busrakckugurlu/Different-gait-combinations-based-on-multi-modal-deep-CNN-architectures.git
引用
收藏
页码:83403 / 83425
页数:23
相关论文
共 50 条
  • [1] Deep Learning Based Multi-Modal Fusion Architectures for Maritime Vessel Detection
    Farahnakian, Fahimeh
    Heikkonen, Jukka
    REMOTE SENSING, 2020, 12 (16)
  • [2] A Multi-Modal Emotion Recognition System Based on CNN-Transformer Deep Learning Technique
    Karatay, Busra
    Bestepe, Deniz
    Sailunaz, Kashfia
    Ozyer, Tansel
    Alhajj, Reda
    2022 7TH INTERNATIONAL CONFERENCE ON DATA SCIENCE AND MACHINE LEARNING APPLICATIONS (CDMA 2022), 2022, : 145 - 150
  • [3] A Deep Multi-Modal CNN for Multi-Instance Multi-Label Image Classification
    Song, Lingyun
    Liu, Jun
    Qian, Buyue
    Sun, Mingxuan
    Yang, Kuan
    Sun, Meng
    Abbas, Samar
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2018, 27 (12) : 6025 - 6038
  • [4] Tactile texture recognition of multi-modal bionic finger based on multi-modal CBAM-CNN interpretable method
    Ma, Feihong
    Li, Yuliang
    Chen, Meng
    DISPLAYS, 2024, 83
  • [5] Multi-modal advanced deep learning architectures for breast cancer survival prediction
    Arya, Nikhilanand
    Saha, Sriparna
    KNOWLEDGE-BASED SYSTEMS, 2021, 221
  • [6] Deep Learning-Based CNN Multi-Modal Camera Model Identification for Video Source Identification
    Singh S.
    Sehgal V.K.
    Informatica (Slovenia), 2023, 47 (03): : 417 - 430
  • [7] Memory based fusion for multi-modal deep learning
    Priyasad, Darshana
    Fernando, Tharindu
    Denman, Simon
    Sridharan, Sridha
    Fookes, Clinton
    INFORMATION FUSION, 2021, 67 : 136 - 146
  • [8] DEEP MULTI-MODAL SCHIZOPHRENIA DISORDER DIAGNOSIS VIA A GRU-CNN ARCHITECTURE
    Masoudi, B.
    Danishvar, S.
    NEURAL NETWORK WORLD, 2022, 32 (03) : 147 - 161
  • [9] Deep learning architectures for Parkinson's disease detection by using multi-modal features
    Pahuja, Gunjan
    Prasad, Bhanu
    COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 146
  • [10] Integrating Deep and Shallow Models for Multi-Modal Depression Analysis-Hybrid Architectures
    Yang, Le
    Jiang, Dongmei
    Sahli, Hichem
    IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2021, 12 (01) : 239 - 253