E2PNet: Event to Point Cloud Registration with Spatio-Temporal Representation Learning

被引:0
|
作者
Lin, Xiuhong [1 ,2 ]
Qiu, Changjie [1 ,2 ]
Cai, Zhipeng [3 ]
Shen, Siqi [1 ,2 ]
Zang, Yu [1 ,2 ]
Liu, Weiquan [1 ,2 ]
Bian, Xuesheng [1 ,5 ]
Mueller, Matthias [4 ]
Wang, Cheng [1 ,2 ]
机构
[1] Xiamen Univ XMU, Sch Informat, Fujian Key Lab Sensing & Comp Smart Cities, Xiamen, Peoples R China
[2] XMU, Key Lab Multimedia Trusted Percept & Efficient Co, Xiamen, Peoples R China
[3] Intel Labs, Hillsboro, OR USA
[4] Apple Inc, Cupertino, CA USA
[5] Yancheng Inst Technol, Yancheng, Peoples R China
基金
中国国家自然科学基金; 中国博士后科学基金;
关键词
VISION;
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Event cameras have emerged as a promising vision sensor in recent years due to their unparalleled temporal resolution and dynamic range. While registration of 2D RGB images to 3D point clouds is a long-standing problem in computer vision, no prior work studies 2D-3D registration for event cameras. To this end, we propose E2PNet, the first learning-based method for event-to-point cloud registration. The core of E2PNet is a novel feature representation network called Event-Points-to-Tensor (EP2T), which encodes event data into a 2D grid-shaped feature tensor. This grid-shaped feature enables matured RGB-based frameworks to be easily used for event-to-point cloud registration, without changing hyper-parameters and the training procedure. EP2T treats the event input as spatio-temporal point clouds. Unlike standard 3D learning architectures that treat all dimensions of point clouds equally, the novel sampling and information aggregation modules in EP2T are designed to handle the inhomogeneity of the spatial and temporal dimensions. Experiments on the MVSEC and VECtor datasets demonstrate the superiority of E2PNet over hand-crafted and other learning-based methods. Compared to RGB-based registration, E2PNet is more robust to extreme illumination or fast motion due to the use of event data. Beyond 2D-3D registration, we also show the potential of EP2T for other vision tasks such as flow estimation, event-to-image reconstruction and object recognition. The source code can be found at: E2PNet.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Deep Hierarchical Representation of Point Cloud Videos via Spatio-Temporal Decomposition
    Fan, Hehe
    Yu, Xin
    Yang, Yi
    Kankanhalli, Mohan
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (12) : 9918 - 9930
  • [2] Registration of spatio-temporal point clouds of plants for phenotyping
    Chebrolu, Nived
    Magistri, Federico
    Labe, Thomas
    Stachniss, Cyrill
    PLOS ONE, 2021, 16 (02):
  • [3] Learning to rank spatio-temporal event hotspots
    Mohler, George
    Porter, Michael
    Carter, Jeremy
    LaFree, Gary
    CRIME SCIENCE, 2020, 9 (01)
  • [4] Learning to rank spatio-temporal event hotspots
    George Mohler
    Michael Porter
    Jeremy Carter
    Gary LaFree
    Crime Science, 9
  • [5] Dual Contrastive Learning for Spatio-temporal Representation
    Ding, Shuangrui
    Qian, Rui
    Xiong, Hongkai
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022, : 5649 - 5658
  • [6] Event stream learning using spatio-temporal event surface
    Dong, Junfei
    Jiang, Runhao
    Xiao, Rong
    Yan, Rui
    Tang, Huajin
    NEURAL NETWORKS, 2022, 154 : 543 - 559
  • [7] Beyond supervision: An unsupervised spatio-temporal point cloud noise modeling for event vision sensor
    Annamalai, Lakshmi
    Thakur, Chetan Singh
    PATTERN RECOGNITION LETTERS, 2024, 184 : 162 - 168
  • [8] Point Spatio-Temporal Pyramid Network for Point Cloud Video Understanding
    Shen, Zhiqiang
    Wang, Longguang
    Guo, Yulan
    Liu, Qiong
    Zhou, Xi
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 (1209-1213) : 1209 - 1213
  • [9] Point Spatio-Temporal Transformer Networks for Point Cloud Video Modeling
    Fan, Hehe
    Yang, Yi
    Kankanhalli, Mohan
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2023, 45 (02) : 2181 - 2192
  • [10] Spatio-Temporal Crop Aggregation for Video Representation Learning
    Sameni, Sepehr
    Jenni, Simon
    Favaro, Paolo
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION, ICCV, 2023, : 5641 - 5651