Transformer With Linear-Window Attention for Feature Matching

被引:0
|
作者
Shen, Zhiwei [1 ,2 ]
Kong, Bin [1 ,3 ,4 ]
Dong, Xiaoyu [1 ,2 ]
机构
[1] Chinese Acad Sci, Hefei Inst Intelligent Machines, Hefei 230031, Peoples R China
[2] Univ Sci & Technol China, Hefei Inst Phys Sci, Hefei 230026, Peoples R China
[3] Anhui Engn Lab Intelligent Driving Technol & Appli, Hefei 230088, Peoples R China
[4] Chinese Acad Sci, Innovat Res Inst Robot & Intelligent Mfg Hefei, Hefei 230088, Peoples R China
关键词
Feature extraction; Transformers; Task analysis; Computational modeling; Computational efficiency; Memory management; Visualization; Feature matching; visual transformer; detector-free; computational complexity; low-texture;
D O I
10.1109/ACCESS.2023.3328855
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
A transformer can capture long-term dependencies through an attention mechanism, and hence, can be applied to various vision tasks. However, its secondary computational complexity is a major obstacle in vision tasks that require accurate predictions. To address this limitation, this study introduces linear-window attention (LWA), a new attention model for a vision transformer. The transformer computes self-attention that is restricted to nonoverlapping local windows and represented as a linear dot product of kernel feature mappings. Furthermore, the computational complexity of each window is reduced to linear from quadratic using the constraint property of matrix products. In addition, we applied the LWA to feature matching to construct a coarse-to-fine-level detector-free feature matching method, called transformer with linear-window attention for feature matching TRLWAM. At the coarse level, we extracted the dense pixel-level matches, and at the fine level, we obtained the final matching results via multi-head multilayer perceptron refinement. We demonstrated the effectiveness of LWA through Replace experiments. The results showed that the TRLWAM could extract dense matches from low-texture or repetitive pattern regions in indoor environments, and exhibited excellent results with a low computational cost for MegaDepth and HPatches datasets. We believe the proposed LWA can provide new conceptions for transformer applications in visual tasks.
引用
收藏
页码:121202 / 121211
页数:10
相关论文
共 50 条
  • [41] LINEAR FEATURE COMPATIBILITY FOR PATTERN-MATCHING RELAXATION
    CUCKA, P
    ROSENFELD, A
    PATTERN RECOGNITION, 1992, 25 (02) : 189 - 196
  • [42] A Morphing of Linear Feature Based on Shape Context Matching
    Fang W.
    Li J.
    Li, Jingzhong (lilideyx@126.com), 1600, Editorial Board of Medical Journal of Wuhan University (42): : 963 - 967
  • [43] Feature Detection and Matching With Linear Adjustment and Adaptive Thresholding
    Cai, Zhiming
    Ou, Yiwen
    Ling, Yufeng
    Dong, Jian
    Lu, Jian
    Lee, Howard
    IEEE ACCESS, 2020, 8 : 189735 - 189746
  • [44] A lightweight transformer with linear self-attention for defect recognition
    Zhai, Yuwen
    Li, Xinyu
    Gao, Liang
    Gao, Yiping
    ELECTRONICS LETTERS, 2024, 60 (17)
  • [45] Vertical Disparity Correction of Stereoscopic Video Using Fast Feature Window Matching
    Jung, Eun-Kyung
    Kim, Chang-Il
    Park, Soon-Yong
    2012 IEEE INTERNATIONAL CONFERENCE ON CONSUMER ELECTRONICS (ICCE), 2012, : 463 - +
  • [46] Multi-Scale Feature Attention-DEtection TRansformer: Multi-Scale Feature Attention for security check object detection
    Sima, Haifeng
    Chen, Bailiang
    Tang, Chaosheng
    Zhang, Yudong
    Sun, Junding
    IET COMPUTER VISION, 2024, 18 (05) : 613 - 625
  • [47] PPLA-Transformer: An Efficient Transformer for Defect Detection with Linear Attention Based on Pyramid Pooling
    Song, Xiaona
    Tian, Yubo
    Liu, Haichao
    Wang, Lijun
    Niu, Jinxing
    SENSORS, 2025, 25 (03)
  • [48] AAPMatcher: Adaptive attention pruning matcher for accurate local feature matching
    Fan, Xuan
    Liu, Sijia
    Liu, Shuaiyan
    Zhao, Lijun
    Li, Ruifeng
    Neural Networks, 2025, 188
  • [49] SWAformer: A novel shifted window attention Transformer model for accurate power distribution prediction
    Cao, Yizhi
    Liao, Yilin
    Liu, Zhaoran
    Ma, Xiang
    Liu, Xinggao
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 265
  • [50] Semantic Segmentation Method of UAV Image Based on Window Attention Aggregation Swin Transformer
    Li, Junjie
    Yi, Shi
    He, Runhua
    Liu, Xi
    Computer Engineering and Applications, 2024, 60 (15) : 198 - 210