Cross-Modal feature description for remote sensing image matching

被引:10
|
作者
Li, Liangzhi [1 ]
Liu, Ming [2 ]
Ma, Lingfei [3 ]
Han, Ling [1 ,2 ]
机构
[1] Changan Univ, Coll Geol Engn & Geomatics, Xian 710064, Peoples R China
[2] Changan Univ, Sch Land Engn, Xian 710064, Peoples R China
[3] Cent Univ Finance & Econ, Sch Stat & Math, Beijing 102206, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature description; Cross; -modal; Self; -attention; Remote sensing image matching; Information interaction; REGISTRATION; FRAMEWORK;
D O I
10.1016/j.jag.2022.102964
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
Effective feature description for cross-modal remote sensing matching is challenging due to the complex geo-metric and radiometric differences between multimodal images. Currently, Siamese or pseudo-Siamese networks directly describe features from multimodal remote sensing images at the fully connected layer, however, the similarity of cross-modal features during feature extraction is barely considered. Therefore, we construct a cross -modal feature description matching network (CM-Net) for remote sensing image matching in this paper. First, a contextual self-attention module is proposed to add semantic global dependency information using candidate and non-candidate keypoint patches. Then, a cross-fusion module is designed to obtain cross-modal feature de-scriptions through information interaction. Finally, a similarity matching loss function is presented to optimize discriminative feature representations, converting a matching task into a classification task. The proposed CM -Net model is evaluated by qualitative and quantitative experiments on four multimodal image datasets, which achieves the average Matching score (M.S.), Mean Matching Accuracy (MMA), and average Root-mean-square error (aRMSE) of 0.781, 0.275, and 1.726, respectively. The comparative study demonstrates the superior per-formance of the proposed CM-Net for the remote sensing image matching.
引用
收藏
页数:12
相关论文
共 50 条
  • [41] Cross-Modal Attention With Semantic Consistence for Image-Text Matching
    Xu, Xing
    Wang, Tan
    Yang, Yang
    Zuo, Lin
    Shen, Fumin
    Shen, Heng Tao
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2020, 31 (12) : 5412 - 5425
  • [42] MEMORY EFFECTS IN CROSS-MODAL MATCHING
    JONES, B
    CONNOLLY, K
    BRITISH JOURNAL OF PSYCHOLOGY, 1970, 61 (02) : 267 - &
  • [43] Geometric Matching for Cross-Modal Retrieval
    Wang, Zheng
    Gao, Zhenwei
    Yang, Yang
    Wang, Guoqing
    Jiao, Chengbo
    Shen, Heng Tao
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, : 1 - 13
  • [44] CROSS-MODAL MATCHING BY YOUNG CHILDREN
    MILNER, AD
    BRYANT, PE
    JOURNAL OF COMPARATIVE AND PHYSIOLOGICAL PSYCHOLOGY, 1970, 71 (03): : 453 - &
  • [45] CROSS-MODAL MATCHING BY AMNESIC SUBJECTS
    SHAW, C
    KENTRIDGE, RW
    AGGLETON, JP
    NEUROPSYCHOLOGIA, 1990, 28 (07) : 665 - 671
  • [46] CROSS-MODAL MATCHING BY SPASTIC CHILDREN
    JONES, B
    ALEXANDER, R
    DEVELOPMENTAL MEDICINE AND CHILD NEUROLOGY, 1974, 16 (01): : 40 - 46
  • [47] Joint feature approach for image-text cross-modal retrieval
    Gao, Dihui
    Sheng, Lijie
    Xu, Xiaodong
    Miao, Qiguang
    Xi'an Dianzi Keji Daxue Xuebao/Journal of Xidian University, 2024, 51 (04): : 128 - 138
  • [48] Stacked cross-modal feature consolidation attention networks for image captioning
    Pourkeshavarz, Mozhgan
    Nabavi, Shahabedin
    Moghaddam, Mohsen Ebrahimi
    Shamsfard, Mehrnoush
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (04) : 12209 - 12233
  • [49] CROSS-MODAL TRANSFER WITH NEURAL WORD VECTORS FOR IMAGE FEATURE LEARNING
    Irie, Go
    Asami, Taichi
    Tarashima, Shuhei
    Kurozumi, Takayuki
    Kinebuchi, Tetsuya
    2017 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2017, : 2916 - 2920
  • [50] Stacked cross-modal feature consolidation attention networks for image captioning
    Mozhgan Pourkeshavarz
    Shahabedin Nabavi
    Mohsen Ebrahimi Moghaddam
    Mehrnoush Shamsfard
    Multimedia Tools and Applications, 2024, 83 : 12209 - 12233