MatchFormer: Interleaving Attention in Transformers for Feature Matching

被引:23
|
作者
Wang, Qing [1 ]
Zhang, Jiaming [1 ]
Yang, Kailun [1 ]
Peng, Kunyu [1 ]
Stiefelhagen, Rainer [1 ]
机构
[1] Karlsruhe Inst Technol, Karlsruhe, Germany
来源
关键词
Feature matching; Vision transformers;
D O I
10.1007/978-3-031-26313-2_16
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Local feature matching is a computationally intensive task at the subpixel level. While detector-based methods coupled with feature descriptors struggle in low-texture scenes, CNN-based methods with a sequential extract-to-match pipeline, fail to make use of the matching capacity of the encoder and tend to overburden the decoder for matching. In contrast, we propose a novel hierarchical extract-and-match transformer, termed as MatchFormer. Inside each stage of the hierarchical encoder, we interleave self-attention for feature extraction and cross-attention for feature matching, yielding a human-intuitive extract-and-match scheme. Such a match-aware encoder releases the overloaded decoder and makes the model highly efficient. Further, combining self- and cross-attention on multi-scale features in a hierarchical architecture improves matching robustness, particularly in low-texture indoor scenes or with less outdoor training data. Thanks to such a strategy, MatchFormer is a multi-win solution in efficiency, robustness, and precision. Compared to the previous best method in indoor pose estimation, our lite MatchFormer has only 45% GFLOPs, yet achieves a +1.3% precision gain and a 41% running speed boost. The large MatchFormer reaches state-of-the-art on four different benchmarks, including indoor pose estimation (ScanNet), outdoor pose estimation (MegaDepth), homography estimation and image matching (HPatch), and visual localization (InLoc).
引用
收藏
页码:256 / 273
页数:18
相关论文
共 50 条
  • [11] Transformer With Linear-Window Attention for Feature Matching
    Shen, Zhiwei
    Kong, Bin
    Dong, Xiaoyu
    IEEE ACCESS, 2023, 11 : 121202 - 121211
  • [12] Learning for Feature Matching via Graph Context Attention
    Guo, Junwen
    Xiao, Guobao
    Tang, Zhimin
    Chen, Shunxing
    Wang, Shiping
    Ma, Jiayi
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2023, 61
  • [13] Modeling Selective Feature Attention for Lightweight Text Matching
    Zang, Jianxiang
    Liu, Hui
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 6624 - 6632
  • [14] Attention-Based Feature Fusion With External Attention Transformers for Breast Cancer Histopathology Analysis
    Vanitha, K.
    Manimaran, A.
    Chokkanathan, K.
    Anitha, K.
    Mahesh, T. R.
    Vinoth Kumar, V.
    Vivekananda, G. N.
    IEEE ACCESS, 2024, 12 : 126296 - 126312
  • [15] ReViT: Enhancing vision transformers feature diversity with attention residual connections
    Diko, Anxhelo
    Avola, Danilo
    Cascio, Marco
    Cinque, Luigi
    PATTERN RECOGNITION, 2024, 156
  • [16] Leakage Inductance Determination for Transformers with Interleaving of Windings
    Doebbelin, R.
    Lindemann, A.
    PIERS 2010 CAMBRIDGE: PROGRESS IN ELECTROMAGNETICS RESEARCH SYMPOSIUM PROCEEDINGS, VOLS 1 AND 2, 2010, : 439 - 443
  • [17] Correspondence Transformers with Asymmetric Feature Learning and Matching Flow Super-Resolution
    Sun, Yixuan
    Zhao, Dongyang
    Yin, Zhangyue
    Huang, Yiwen
    Gui, Tao
    Zhang, Wenqiang
    Ge, Weifeng
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 17787 - 17796
  • [18] Fast Image Matching Based on Channel Attention and Feature Slicing
    Gai Shaoyan
    Huang Yanyan
    Da Feipeng
    ACTA OPTICA SINICA, 2023, 43 (22)
  • [19] Local feature matching transformers for intraoperative OCT en face and fundus image registration
    Matten, Philipp
    Sommersperger, Michael
    Dehghani, Shervin
    Roodaki, Hessam
    Drexler, Wolfgang
    Leitgeb, Rainer A.
    Schmoll, Tilman
    Navab, Nassir
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2024, 65 (07)
  • [20] AAPMatcher: Adaptive attention pruning matcher for accurate local feature matching
    Fan, Xuan
    Liu, Sijia
    Liu, Shuaiyan
    Zhao, Lijun
    Li, Ruifeng
    Neural Networks, 2025, 188