Event Camera Data Pre-training

被引:0
|
作者
Yang, Yan [1 ]
Pan, Liyuan [2 ,3 ]
Liu, Liu [4 ]
机构
[1] Australian Natl Univ, BDSI, Canberra, ACT, Australia
[2] BITSZ, Beijing, Peoples R China
[3] BIT, Sch CSAT, Beijing, Peoples R China
[4] Huawei, Cyberverse Dept, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.00982
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes a pre-trained neural network for handling event camera data. Our model is a self-supervised learning framework, and uses paired event camera data and natural RGB images for training. Our method contains three modules connected in a sequence: i) a family of event data augmentations, generating meaningful event images for self-supervised training; ii) a conditional masking strategy to sample informative event patches from event images, encouraging our model to capture the spatial layout of a scene and accelerating training; iii) a contrastive learning approach, enforcing the similarity of embeddings between matching event images, and between paired event and RGB images. An embedding projection loss is proposed to avoid the model collapse when enforcing the event image embedding similarities. A probability distribution alignment loss is proposed to encourage the event image to be consistent with its paired RGB image in the feature space. Transfer learning performance on downstream tasks shows the superiority of our method over state-of-the-art methods. For example, we achieve top-1 accuracy at 64.83% on the N-ImageNet dataset. Our code is available at https://github.com/Yan98/Event-Camera-Data-Pre-training.
引用
下载
收藏
页码:10665 / 10675
页数:11
相关论文
共 50 条
  • [1] Pre-training in Medical Data: A Survey
    Qiu, Yixuan
    Lin, Feng
    Chen, Weitong
    Xu, Miao
    MACHINE INTELLIGENCE RESEARCH, 2023, 20 (02) : 147 - 179
  • [2] Pre-Training Model of Public Opinion Event Vector
    Wang, Nan
    Tan, Shuru
    Xie, Xiaolan
    Li, Hairong
    Computer Engineering and Applications, 2024, 60 (18) : 189 - 197
  • [3] The ImageNet Shuffle: Reorganized Pre-training for Video Event Detection
    Mettes, Pascal
    Koelma, Dennis C.
    Snoek, Cees G. M.
    ICMR'16: PROCEEDINGS OF THE 2016 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, 2016, : 175 - 182
  • [4] Application Specificity of Data for Pre-Training in Computer Vision
    Peters, Gabriel G.
    Couwenhoven, Scott D.
    Walvoord, Derek J.
    Salvaggio, Carl
    DISRUPTIVE TECHNOLOGIES IN INFORMATION SCIENCES VIII, 2024, 13058
  • [5] ELLE: Efficient Lifelong Pre-training for Emerging Data
    Qin, Yujia
    Zhang, Jiajie
    Lin, Yankai
    Liu, Zhiyuan
    Li, Peng
    Sun, Maosong
    Zhou, Jie
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 2789 - 2810
  • [6] Event Feature Pre-training Model Based on Public Opinion Evolution
    Wang, Nan
    Tan, Shu-Ru
    Xie, Xiao-Lan
    Li, Hai-Rong
    Jiang, Jia-Hui
    INTERNATIONAL JOURNAL OF ADVANCED COMPUTER SCIENCE AND APPLICATIONS, 2023, 14 (04) : 197 - 206
  • [7] Unifying Event Detection and Captioning as Sequence Generation via Pre-training
    Zhang, Qi
    Song, Yuqing
    Jin, Qin
    COMPUTER VISION, ECCV 2022, PT XXXVI, 2022, 13696 : 363 - 379
  • [8] PTEKC: pre-training with event knowledge of ConceptNet for cross-lingual event causality identification
    Enchang Zhu
    Zhengtao Yu
    Yuxin Huang
    Shengxiang Gao
    Yantuan Xian
    International Journal of Machine Learning and Cybernetics, 2025, 16 (3) : 1859 - 1872
  • [9] Unifying Structured Data as Graph for Data-to-Text Pre-Training
    Li, Shujie
    Li, Liang
    Geng, Ruiying
    Yang, Min
    Li, Binhua
    Yuan, Guanghu
    He, Wanwei
    Yuan, Shao
    Ma, Can
    Huang, Fei
    Li, Yongbin
    TRANSACTIONS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, 2024, 12 : 210 - 228
  • [10] Multi-stage Pre-training over Simplified Multimodal Pre-training Models
    Liu, Tongtong
    Feng, Fangxiang
    Wang, Xiaojie
    59TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS AND THE 11TH INTERNATIONAL JOINT CONFERENCE ON NATURAL LANGUAGE PROCESSING, VOL 1 (ACL-IJCNLP 2021), 2021, : 2556 - 2565