WiVi: WiFi-Video Cross-Modal Fusion based Multi-Path Gait Recognition System

被引:3
|
作者
Fan, Jinmeng [1 ,2 ,3 ]
Zhou, Hao [1 ,2 ,3 ]
Zhou, Fengyu [1 ,2 ,3 ]
Wang, Xiaoyan [4 ]
Liu, Zhi [5 ]
Li, Xiang-Yang [1 ,2 ,3 ]
机构
[1] Univ Sci & Technol China, Sch Comp Sci & Technol, LINKE Lab, Hefei, Peoples R China
[2] Univ Sci & Technol China, CAS Key Lab Wireless Opt Commun, Hefei, Peoples R China
[3] Deqing Alpha Innovat Inst, Huzhou, Zhejiang, Peoples R China
[4] Ibaraki Univ, Grad Sch Sci & Engn, Ibaraki, Japan
[5] Univ Electrocommun, Dept Comp & Network Engn, Tokyo, Japan
基金
国家重点研发计划;
关键词
D O I
10.1109/IWQoS54832.2022.9812893
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
WiFi-based gait recognition is an attractive method for device-free user identification, but path-sensitive Channel State Information (CSI) hinders its application in multi-path environments, which exacerbates sampling and deployment costs (i.e., large number of samples and multiple specially placed devices). On the other hand, although video-based ideal CSI generation is promising for dramatically reducing samples, the missing environment-related information in the ideal CSI makes it unsuitable for general indoor scenarios with multiple walking paths. In this paper, we propose WiVi, a WiFi-video cross-modal fusion based multi-path gait recognition system which needs fewer samples and fewer devices simultaneously. When the subject walks naturally in the room, we determine whether he/she is walking on the predefined judgment paths with a K-Nearest Neighbors (KNN) classifier working on the WiFi-based human localization results. For each judgment path, we generate the ideal CSI through video-based simulation to decrease the number of needed samples, and adopt two separated neural networks (NNs) to fulfill environment-aware comparison among the ideal and measured CSIs. The first network is supervised by measured CSI samples, and learns to obtain the semi-ideal CSI features which contain the room-specific 'accent', i.e., the long-term environment influence normally caused by room layout. The second network is trained for similarity evaluation between the semi-ideal and measured features, with the existence of short-term environment influence such as channel variation or noises. We implement the prototype system and conduct extensive experiments to evaluate the performance. Experimental results show that WiVi's recognition accuracy ranges from 85.4% for a 6-person group to 98.0% for a 3-person group. As compared with single-path gait recognition systems, we achieve average 113.8% performance improvement. As compared with the other multi-path gait recognition systems, we achieve similar or even better performance with needed samples being reduced by 57.1-93.7%
引用
下载
收藏
页数:10
相关论文
共 50 条
  • [1] IMFi: IMU-WiFi based Cross-modal Gait Recognition System with Hot-Deployment
    Song, Zengyu
    Zhou, Hao
    Wang, Shan
    Fan, Jinmeng
    Guo, Kaiwen
    Zhou, Wangqiu
    Wang, Xiaoyan
    Li, Xiang-Yang
    2021 17TH INTERNATIONAL CONFERENCE ON MOBILITY, SENSING AND NETWORKING (MSN 2021), 2021, : 279 - 286
  • [2] Metaphor recognition based on cross-modal multi-level information fusion
    Qimeng Yang
    Yuanbo Yan
    Xiaoyu He
    Shisong Guo
    Complex & Intelligent Systems, 2025, 11 (1)
  • [3] Multi-modal fusion network with multi-scale multi-path and cross-modal interactions for RGB-D salient object detection
    Chen, Hao
    Li, Youfu
    Su, Dan
    PATTERN RECOGNITION, 2019, 86 : 376 - 385
  • [4] A Short Video Classification Framework Based on Cross-Modal Fusion
    Pang, Nuo
    Guo, Songlin
    Yan, Ming
    Chan, Chien Aun
    SENSORS, 2023, 23 (20)
  • [5] Research on cross-modal emotion recognition based on multi-layer semantic fusion
    Xu Z.
    Gao Y.
    Mathematical Biosciences and Engineering, 2024, 21 (02) : 2488 - 2514
  • [6] Parallel learned generative adversarial network with multi-path subspaces for cross-modal retrieval
    Li, Zhuoyi
    Lu, Huibin
    Fu, Hao
    Gu, Guanghua
    INFORMATION SCIENCES, 2023, 620 : 84 - 104
  • [7] Cross-modal learning with multi-modal model for video action recognition based on adaptive weight training
    Zhou, Qingguo
    Hou, Yufeng
    Zhou, Rui
    Li, Yan
    Wang, Jinqiang
    Wu, Zhen
    Li, Hung-Wei
    Weng, Tien-Hsiung
    CONNECTION SCIENCE, 2024, 36 (01)
  • [8] Gated Multi-modal Fusion with Cross-modal Contrastive Learning for Video Question Answering
    Lyu, Chenyang
    Li, Wenxi
    Ji, Tianbo
    Zhou, Liting
    Gurrin, Cathal
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING, ICANN 2023, PT VII, 2023, 14260 : 427 - 438
  • [9] Cross-Modal Semantic Fusion Video Emotion Analysis Based on Attention Mechanism
    Zhao, Lianfen
    Pan, Zhengjun
    2023 8TH INTERNATIONAL CONFERENCE ON CLOUD COMPUTING AND BIG DATA ANALYTICS, ICCCBDA, 2023, : 381 - 386
  • [10] Label graph learning for multi-label image recognition with cross-modal fusion
    Xie, Yanzhao
    Wang, Yangtao
    Liu, Yu
    Zhou, Ke
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (18) : 25363 - 25381