ImVoteNet: Boosting 3D Object Detection in Point Clouds with Image Votes

被引:159
|
作者
Qi, Charles R.
Chen, Xinlei [1 ]
Litany, Or [1 ,2 ]
Guibas, Leonidas J. [1 ,2 ]
机构
[1] Facebook AI, Menlo Pk, CA 94025 USA
[2] Stanford Univ, Stanford, CA 94305 USA
关键词
HOUGH TRANSFORM; DATABASE;
D O I
10.1109/CVPR42600.2020.00446
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
3D object detection has seen quick progress thanks to advances in deep learning on point clouds. A few recent works have even shown state-of-the-art performance with just point clouds input (e.g. VOTENET). However, point cloud data have inherent limitations. They are sparse, lack color information and often suffer from sensor noise. Images, on the other hand, have high resolution and rich texture. Thus they can complement the 3D geometry provided by point clouds. Yet how to effectively use image information to assist point cloud based detection is still an open question. In this work, we build on top of VOTENET and propose a 3D detection architecture called IMVOTENET specialized for RGB-D scenes. IMVOTENET is based on fusing 2D votes in images and 3D votes in point clouds. Compared to prior work on multi-modal detection, we explicitly extract both geometric and semantic features from the 2D images. We leverage camera parameters to lift these features to 3D. To improve the synergy of 2D-3D feature fusion, we also propose a multi-tower training scheme. We validate our model on the challenging SUN RGB-D dataset, advancing state-of-the-art results by 5.7 mAP. We also provide rich ablation studies to analyze the contribution of each design choice.
引用
收藏
页码:4403 / 4412
页数:10
相关论文
共 50 条
  • [21] Leveraging Dynamic Occupancy Grids for 3D Object Detection in Point Clouds
    Sierra-Gonzalez, David
    Paigwar, Anshul
    Erkent, Ozgur
    Dibangoye, Jilles
    Laugier, Christian
    16TH IEEE INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION, ROBOTICS AND VISION (ICARCV 2020), 2020, : 1188 - 1193
  • [22] Object Detection and Classification in 3D Point Clouds for Mobile Robot Manipulation
    Cupec, Robert
    PROCEEDINGS OF 2019 INTERNATIONAL CONFERENCE ON SYSTEMS, SIGNALS AND IMAGE PROCESSING (IWSSIP 2019), 2019, : 19 - 19
  • [23] SWFormer: Sparse Window Transformer for 3D Object Detection in Point Clouds
    Sun, Pei
    Tan, Mingxing
    Wang, Weiyue
    Liu, Chenxi
    Xia, Fei
    Leng, Zhaoqi
    Anguelov, Dragomir
    COMPUTER VISION, ECCV 2022, PT X, 2022, 13670 : 426 - 442
  • [24] Attentional Keypoint Detection on Point Clouds for 3D Object Part Segmentation
    Zhou, Feng
    Zhang, Qi
    Zhu, He
    Liu, Shibo
    Jiang, Na
    Cai, Xingquan
    Qi, Qianfang
    Hu, Yong
    APPLIED SCIENCES-BASEL, 2023, 13 (23):
  • [25] 3D Cascade RCNN: High Quality Object Detection in Point Clouds
    Cai, Qi
    Pan, Yingwei
    Yao, Ting
    Mei, Tao
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2022, 31 : 5706 - 5719
  • [26] DISCOVERING CORRESPONDENCE AMONG IMAGE SETS WITH PROJECTION VIEW PRESERVATION FOR 3D OBJECT DETECTION IN POINT CLOUDS
    Yamazaki, Tomoaki
    Sugimura, Daisuke
    Hamamoto, Takayuki
    2018 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2018, : 3111 - 3115
  • [27] Interactive Object Segmentation in 3D Point Clouds
    Kontogianni, Theodora
    Celikkan, Ekin
    Tang, Siyu
    Schindler, Konrad
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 2891 - 2897
  • [28] SOA: Seed point offset attention for indoor 3D object detection in point clouds
    Shu, Jun
    Yu, Shiqi
    Shu, Xinyi
    Hu, Jiewen
    COMPUTERS & GRAPHICS-UK, 2024, 123
  • [29] Scale invariant point feature (SIPF) for 3D point clouds and 3D multi-scale object detection
    Lin, Baowei
    Wang, Fasheng
    Zhao, Fangda
    Sun, Yi
    NEURAL COMPUTING & APPLICATIONS, 2018, 29 (05): : 1209 - 1224
  • [30] Exploiting Label Uncertainty for Enhanced 3D Object Detection From Point Clouds
    Sun, Yang
    Lu, Bin
    Liu, Yonghuai
    Yang, Zhenyu
    Behera, Ardhendu
    Song, Ran
    Yuan, Hejin
    Jiang, Haiyan
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2024, 25 (06) : 6074 - 6089