CROSS-MODAL KNOWLEDGE DISTILLATION FOR ACTION RECOGNITION

被引:0
|
作者
Thoker, Fida Mohammad [1 ]
Gall, Juergen [1 ]
机构
[1] Univ Bonn, Bonn, Germany
基金
欧洲研究理事会;
关键词
Knowledge Distillation; Action Recognition; Transfer Learning; Cross-Modal Action Recognition;
D O I
10.1109/icip.2019.8802909
中图分类号
TB8 [摄影技术];
学科分类号
0804 ;
摘要
In this work, we address the problem how a network for action recognition that has been trained on a modality like RGB videos can be adapted to recognize actions for another modality like sequences of 3D human poses. To this end, we extract the knowledge of the trained teacher network for the source modality and transfer it to a small ensemble of student networks for the target modality. For the cross-modal knowledge distillation, we do not require any annotated data. Instead we use pairs of sequences of both modalities as supervision, which are straightforward to acquire. In contrast to previous works for knowledge distillation that use a KL-loss, we show that the cross-entropy loss together with mutual learning of a small ensemble of student networks performs better. In fact, the proposed approach for cross-modal knowledge distillation nearly achieves the accuracy of a student network trained with full supervision.
引用
收藏
页码:6 / 10
页数:5
相关论文
共 50 条
  • [31] Cross-Modal Graph Knowledge Representation and Distillation Learning for Land Cover Classification
    Wang, Wenzhen
    Liu, Fang
    Liao, Wenzhi
    Xiao, Liang
    [J]. IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [32] A Precise and Scalable Indoor Positioning System Using Cross-Modal Knowledge Distillation
    Rizk, Hamada
    Elmogy, Ahmed
    Rihan, Mohamed
    Yamaguchi, Hirozumi
    [J]. Sensors, 2024, 24 (22)
  • [33] Unsupervised domain adaptation for lip reading based on cross-modal knowledge distillation
    Yuki Takashima
    Ryoichi Takashima
    Ryota Tsunoda
    Ryo Aihara
    Tetsuya Takiguchi
    Yasuo Ariki
    Nobuaki Motoyama
    [J]. EURASIP Journal on Audio, Speech, and Music Processing, 2021
  • [34] CKDH: CLIP-Based Knowledge Distillation Hashing for Cross-Modal Retrieval
    Li, Jiaxing
    Wong, Wai Keung
    Jiang, Lin
    Fang, Xiaozhao
    Xie, Shengli
    Xu, Yong
    [J]. IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (07) : 6530 - 6541
  • [35] Cross-modal distillation for flood extent mapping
    Garg, Shubhika
    Feinstein, Ben
    Timnat, Shahar
    Batchu, Vishal
    Dror, Gideon
    Rosenthal, Adi Gerzi
    Gulshan, Varun
    [J]. ENVIRONMENTAL DATA SCIENCE, 2023, 2
  • [36] Cross-modal Representation Learning for Zero-shot Action Recognition
    Lin, Chung-Ching
    Lin, Kevin
    Wang, Lijuan
    Liu, Zicheng
    Li, Linjie
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 19946 - 19956
  • [37] Text-Derived Knowledge Helps Vision: A Simple Cross-modal Distillation for Video-based Action Anticipation
    Ghosh, Sayontan
    Aggarwal, Tanvi
    Hoai, Minh
    Balasubramanian, Niranjan
    [J]. 17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 1882 - 1897
  • [38] Multimodal Fusion with Cross-Modal Attention for Action Recognition in Still Images
    Tsai, Jia-Hua
    Chu, Wei-Ta
    [J]. PROCEEDINGS OF THE 4TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA IN ASIA, MMASIA 2022, 2022,
  • [39] CROSS-MODAL KNOWLEDGE DISTILLATION FOR FINE-GRAINED ONE-SHOT CLASSIFICATION
    Zhao, Jiabao
    Lin, Xin
    Yang, Yifan
    Yang, Jing
    He, Liang
    [J]. 2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 4295 - 4299
  • [40] Social Image-Text Sentiment Classification With Cross-Modal Consistency and Knowledge Distillation
    Liu, Huan
    Li, Ke
    Fan, Jianping
    Yan, Caixia
    Qin, Tao
    Zheng, Qinghua
    [J]. IEEE TRANSACTIONS ON AFFECTIVE COMPUTING, 2023, 14 (04) : 3332 - 3344