Privacy-Safe Action Recognition via Cross-Modality Distillation

被引:0
|
作者
Kim, Yuhyun [1 ]
Jung, Jinwook [1 ]
Noh, Hyeoncheol [1 ]
Ahn, Byungtae [2 ]
Kwon, Junghye [3 ]
Choi, Dong-Geol [1 ]
机构
[1] Hanbat Natl Univ, Dept Informat & Commun Engn, Daejeon 34158, South Korea
[2] Korea Inst Machinery & Mat, Daejeon 34103, South Korea
[3] Chungnam Natl Univ, Coll Med, Dept Internal Med, Div Hematol Oncol, Daejeon 34134, South Korea
来源
IEEE ACCESS | 2024年 / 12卷
关键词
Action recognition; knowledge distillation; cross-modality distillation; deep learning; multi modal; privacy-safe;
D O I
10.1109/ACCESS.2024.3431227
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Human action recognition systems enhance public safety by detecting abnormal behavior autonomously. RGB sensors commonly used in such systems capture personal information of subjects and, as a result, run the risk of potential privacy leakage. On the other hand, privacy-safe alternatives, such as depth or thermal sensors, exhibit poorer performance because they lack the semantic context provided by RGB sensors. Moreover, the data availability of privacy-safe alternatives is significantly lower than RGB sensors. To address these problems, we explore effective cross-modality distillation methods in this paper, aiming to distill the knowledge of context-rich large-scale pre-trained RGB-based models into privacy-safe depth-based models. Based on extensive experiments on multiple architectures and benchmark datasets, we propose an effective method for training privacy-safe depth-based action recognition models via cross-modality distillation: cross-modality mixing distillation. This approach improves both the performance and efficiency by enabling interaction between depth and RGB modalities through a linear combination of their features. By utilizing the proposed cross-modal mixing distillation approach, we achieve state-of-the-art accuracy in two depth-based action recognition benchmarks. The code and the pre-trained models will be available upon publication.
引用
收藏
页码:125955 / 125965
页数:11
相关论文
共 50 条
  • [1] Cross-modality online distillation for multi-view action recognition
    Xu, Chao
    Wu, Xia
    Li, Yachun
    Jin, Yining
    Wang, Mengmeng
    Liu, Yong
    NEUROCOMPUTING, 2021, 456 : 384 - 393
  • [2] Efficient RGB-T Tracking via Cross-Modality Distillation
    Zhang, Tianlu
    Guo, Hongyuan
    Jiao, Qiang
    Zhang, Qiang
    Han, Jungong
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR, 2023, : 5404 - 5413
  • [3] Cross-Modality Face Recognition via Heterogeneous Joint Bayesian
    Shi, Hailin
    Wang, Xiaobo
    Yi, Dong
    Lei, Zhen
    Zhu, Xiangyu
    Li, Stan Z.
    IEEE SIGNAL PROCESSING LETTERS, 2017, 24 (01) : 81 - 85
  • [4] Dynamic Knowledge Distillation with Cross-Modality Knowledge Transfer
    Wang, Guangzhi
    PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 2974 - 2978
  • [5] CROSS-MODALITY DISTILLATION: A CASE FOR CONDITIONAL GENERATIVE ADVERSARIAL NETWORKS
    Roheda, Siddharth
    Riggan, Benjamin S.
    Krim, Hamid
    Dai, Liyi
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 2926 - 2930
  • [6] Semi-Supervised Cross-Modality Action Recognition by Latent Tensor Transfer Learning
    Jia, Chengcheng
    Ding, Zhengming
    Kong, Yu
    Fu, Yun
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2020, 30 (09) : 2801 - 2814
  • [7] Cross-Modality Compensation Convolutional Neural Networks for RGB-D Action Recognition
    Cheng, Jun
    Ren, Ziliang
    Zhang, Qieshi
    Gao, Xiangyang
    Hao, Fusheng
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2022, 32 (03) : 1498 - 1509
  • [8] Accurate Positioning via Cross-Modality Training
    Papaioannou, Savvas
    Wen, Hongkai
    Xiao, Zhuoling
    Markham, Andrew
    Trigoni, Niki
    SENSYS'15: PROCEEDINGS OF THE 13TH ACM CONFERENCE ON EMBEDDED NETWORKED SENSOR SYSTEMS, 2015, : 239 - 251
  • [9] Cross-Modality Gesture Recognition With Complete Representation Projection
    Liu, Xiaokai
    Li, Mingyue
    Zhang, Boyi
    Hao, Luyuan
    Ma, Xiaorui
    Wang, Jie
    IEEE INTERNET OF THINGS JOURNAL, 2024, 11 (09): : 16184 - 16195
  • [10] Incremental Cross-Modality Deep Learning for Pedestrian Recognition
    Pop, Danut Ovidiu
    Rogozan, Alexandrina
    Nashashibi, Fawzi
    Bensrhair, Abdelaziz
    2017 28TH IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV 2017), 2017, : 523 - 528