Event Camera Data Pre-training

被引:0
|
作者
Yang, Yan [1 ]
Pan, Liyuan [2 ,3 ]
Liu, Liu [4 ]
机构
[1] Australian Natl Univ, BDSI, Canberra, ACT, Australia
[2] BITSZ, Beijing, Peoples R China
[3] BIT, Sch CSAT, Beijing, Peoples R China
[4] Huawei, Cyberverse Dept, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
10.1109/ICCV51070.2023.00982
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes a pre-trained neural network for handling event camera data. Our model is a self-supervised learning framework, and uses paired event camera data and natural RGB images for training. Our method contains three modules connected in a sequence: i) a family of event data augmentations, generating meaningful event images for self-supervised training; ii) a conditional masking strategy to sample informative event patches from event images, encouraging our model to capture the spatial layout of a scene and accelerating training; iii) a contrastive learning approach, enforcing the similarity of embeddings between matching event images, and between paired event and RGB images. An embedding projection loss is proposed to avoid the model collapse when enforcing the event image embedding similarities. A probability distribution alignment loss is proposed to encourage the event image to be consistent with its paired RGB image in the feature space. Transfer learning performance on downstream tasks shows the superiority of our method over state-of-the-art methods. For example, we achieve top-1 accuracy at 64.83% on the N-ImageNet dataset. Our code is available at https://github.com/Yan98/Event-Camera-Data-Pre-training.
引用
下载
收藏
页码:10665 / 10675
页数:11
相关论文
共 50 条
  • [21] Pre-training phenotyping classifiers
    Dligach, Dmitriy
    Afshar, Majid
    Miller, Timothy
    JOURNAL OF BIOMEDICAL INFORMATICS, 2021, 113 (113)
  • [22] Rethinking Pre-training and Self-training
    Zoph, Barret
    Ghiasi, Golnaz
    Lin, Tsung-Yi
    Cui, Yin
    Liu, Hanxiao
    Cubuk, Ekin D.
    Le, Quoc V.
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [23] Sparse attention is all you need for pre-training on tabular data
    Tokimasa Isomura
    Ryotaro Shimizu
    Masayuki Goto
    Neural Computing and Applications, 2025, 37 (3) : 1509 - 1522
  • [24] Early Rumor Detection based on Data Augmentation and Pre-training Transformer
    Hu, Yanjun
    Ju, Xinyi
    Ye, Zhousheng
    Khan, Sulaiman
    Yuan, Chengwu
    Lai, Qiran
    Liu, Junqiang
    2022 IEEE 12TH ANNUAL COMPUTING AND COMMUNICATION WORKSHOP AND CONFERENCE (CCWC), 2022, : 152 - 158
  • [25] Historical document image analysis using controlled data for pre-training
    Najoua Rahal
    Lars Vögtlin
    Rolf Ingold
    International Journal on Document Analysis and Recognition (IJDAR), 2023, 26 : 241 - 254
  • [26] Too Large; Data Reduction for Vision-Language Pre-Training
    Wang, Alex Jinpeng
    Lin, Kevin Qinghong
    Zhang, David Junhao
    Lei, Stan Weixian
    Shou, Mike Zheng
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 3124 - 3134
  • [27] Historical document image analysis using controlled data for pre-training
    Rahal, Najoua
    Vogtlin, Lars
    Ingold, Rolf
    INTERNATIONAL JOURNAL ON DOCUMENT ANALYSIS AND RECOGNITION, 2023, 26 (03) : 241 - 254
  • [28] Unsupervised Pre-Training of Image Features on Non-Curated Data
    Caron, Mathilde
    Bojanowski, Piotr
    Mairal, Julien
    Joulin, Armand
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 2959 - 2968
  • [29] Improved Fine-Tuning by Better Leveraging Pre-Training Data
    Liu, Ziquan
    Xu, Yi
    Xu, Yuanhong
    Qian, Qi
    Li, Hao
    Ji, Xiangyang
    Chan, Antoni B.
    Jin, Rong
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,
  • [30] Trajectory-based fish event classification through pre-training with diffusion models
    Canovi, Noemi
    Ellis, Benjamin A.
    Sordalen, Tonje K.
    Allken, Vaneeda
    Halvorsen, Kim T.
    Malde, Ketil
    Beyan, Cigdem
    ECOLOGICAL INFORMATICS, 2024, 82