MixVPR: Feature Mixing for Visual Place Recognition

被引:72
|
作者
Ali-bey, Amar [1 ]
Chaib-draa, Brahim [1 ]
Giguere, Philippe [1 ]
机构
[1] Univ Laval, Quebec City, PQ, Canada
关键词
MODEL;
D O I
10.1109/WACV56688.2023.00301
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Visual Place Recognition (VPR) is a crucial part of mobile robotics and autonomous driving as well as other computer vision tasks. It refers to the process of identifying a place depicted in a query image using only computer vision. At large scale, repetitive structures, weather and illumination changes pose a real challenge, as appearances can drastically change over time. Along with tackling these challenges, an efficient VPR technique must also be practical in real-world scenarios where latency matters. To address this, we introduce MixVPR, a new holistic feature aggregation technique that takes feature maps from pre-trained backbones as a set of global features. Then, it incorporates a global relationship between elements in each feature map in a cascade of feature mixing, eliminating the need for local or pyramidal aggregation as done in NetVLAD or TransVPR. We demonstrate the effectiveness of our technique through extensive experiments on multiple large-scale benchmarks. Our method outperforms all existing techniques by a large margin while having less than half the number of parameters compared to CosPlace and NetVLAD. We achieve a new all-time high recall@1 score of 94.6% on Pitts250k-test, 88.0% on MapillarySLS, and more importantly, 58.4% on Nordland. Finally, our method outperforms two-stage retrieval techniques such as Patch-NetVLAD, TransVPR and SuperGLUE all while being orders of magnitude faster.
引用
收藏
页码:2997 / 3006
页数:10
相关论文
共 50 条
  • [31] Feature Weighting in Visual Product Recognition
    Zhang, Wen
    Yap, Kim-Hui
    Zhang, Da-Jiang
    Miao, Zhen-Wei
    2015 IEEE INTERNATIONAL SYMPOSIUM ON CIRCUITS AND SYSTEMS (ISCAS), 2015, : 734 - 737
  • [32] Visual Place Recognition in Long-term and Large-scale Environment based on CNN Feature
    Zhu, Jianliang
    Ai, Yunfeng
    Tian, Bin
    Cao, Dongpu
    Scherer, Sebastian
    2018 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2018, : 1679 - 1685
  • [33] Erratum to “A Robust Feature Matching Strategy for Fast and Effective Visual Place Recognition in Challenging Environmental Conditions”
    Saba Arshad
    Gon-Woo Kim
    International Journal of Control, Automation and Systems, 2023, 21 : 1726 - 1726
  • [34] Fast Dual-Feature Extraction Based on Tightly Coupled Lightweight Network for Visual Place Recognition
    Hu, Xiaofei
    Zhou, Yang
    Lyu, Liang
    Lan, Chaozhen
    Shi, Qunshan
    Hou, Mingbo
    IEEE ACCESS, 2023, 11 : 127855 - 127865
  • [35] Deep Learning Off-the-shelf Holistic Feature Descriptors for Visual Place Recognition in Challenging Conditions
    Aliajni, Farid
    Rahtu, Esa
    2020 IEEE 22ND INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP), 2020,
  • [36] HBST: A Hamming Distance Embedding Binary Search Tree for Feature-Based Visual Place Recognition
    Schlegel, Dominik
    Grisetti, Giorgio
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2018, 3 (04): : 3741 - 3748
  • [37] Visual Navigation using Place Recognition with Visual Line Words
    Kim, Yong Nyeon
    Kol, Dong Wook
    Suh, Il Hong
    2014 11TH INTERNATIONAL CONFERENCE ON UBIQUITOUS ROBOTS AND AMBIENT INTELLIGENCE (URAI), 2014, : 676 - 676
  • [38] Building Location Models for Visual Place Recognition
    Stumm, Elena S.
    Mei, Christopher
    Lacroix, Simon
    INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH, 2016, 35 (04): : 334 - 356
  • [39] On Motion Blur and Deblurring in Visual Place Recognition
    Ismagilov, Timur
    Ferrarini, Bruno
    Milford, Michael
    Tuyen, Nguyen Tan Viet
    Ramchurn, Sarvapali D.
    Ehsan, Shoaib
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2025, 10 (05): : 4746 - 4753
  • [40] DEEP VISUAL PLACE RECOGNITION FOR WATERBORNE DOMAINS
    Thomas, Luke
    Edwards, Michael
    Capsey, Austin
    Rahat, Alma
    Roach, Matt
    2022 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING, ICIP, 2022, : 3546 - 3550