MSAN: Multiscale self-attention network for pansharpening

被引:0
|
作者
Lu, Hangyuan [1 ,2 ]
Yang, Yong [3 ]
Huang, Shuying [4 ]
Liu, Rixian [1 ,2 ]
Guo, Huimin [5 ]
机构
[1] Jinhua Univ Vocat Technol, Coll Informat Engn, Jinhua 321007, Peoples R China
[2] Jinhua Univ Vocat Technol, Key Lab Crop Harvesting Equipment Technol Zhejiang, Jinhua 321007, Peoples R China
[3] Tiangong Univ, Sch Comp Sci & Technol, Tianjin 300387, Peoples R China
[4] Tiangong Univ, Sch Software, Tianjin 300387, Peoples R China
[5] Xiamen Univ, Sch Informat, Xiamen 361005, Peoples R China
基金
中国国家自然科学基金;
关键词
Pansharpening; Multiscale; Self-attention; Swin Transformer; FUSION; IMAGES;
D O I
10.1016/j.patcog.2025.111441
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Effective extraction of spectral-spatial features from multispectral (MS) and panchromatic (PAN) images is critical for high-quality pansharpening. However, existing deep learning methods often overlook local misalignment and struggle to integrate local and long-range features effectively, resulting in spectral and spatial distortions. To address these challenges, this paper proposes a refined detail injection model that adaptively learns injection coefficients using long-range features. Building upon this model, a multiscale self-attention network (MSAN) is proposed, consisting of a feature extraction branch and a self-attention mechanism branch. In the former branch, a two-stage multiscale convolution network is designed to fully extract detail features with multiple receptive fields. In the latter branch, a streamlined Swin Transformer (SST) is proposed to efficiently generate multiscale self-attention maps by learning the correlation between local and long-range features. To better preserve spectral-spatial information, a revised Swin Transformer block is proposed by incorporating spectral and spatial attention within the block. The obtained self-attention maps from SST serve as the injection coefficients to refine the extracted details, which are then injected into the upsampled MS image to produce the final fused image. Experimental validation demonstrates the superiority of MSAN over traditional and state-of-the-art methods, with competitive efficiency. The code of this work will be released on GitHub once the paper is accepted.
引用
收藏
页数:17
相关论文
共 50 条
  • [31] TESANet: Self-attention network for olfactory EEG classification
    Tong, Chengxuan
    Ding, Yi
    Liang, Kevin Lim Jun
    Zhang, Zhuo
    Zhang, Haihong
    Guan, Cuntai
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [32] Grassmannian Manifold Self-Attention Network for Signal Classification
    Wang, Rui
    Hu, Chen
    Chen, Ziheng
    Wu, Xiao-Jun
    Song, Xiaoning
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 5099 - 5107
  • [33] Quantum mixed-state self-attention network
    Chen, Fu
    Zhao, Qinglin
    Feng, Li
    Chen, Chuangtao
    Lin, Yangbin
    Lin, Jianhong
    NEURAL NETWORKS, 2025, 185
  • [34] A visual self-attention network for facial expression recognition
    Yu, Naigong
    Bai, Deguo
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [35] A Self-attention Based LSTM Network for Text Classification
    Jing, Ran
    2019 3RD INTERNATIONAL CONFERENCE ON CONTROL ENGINEERING AND ARTIFICIAL INTELLIGENCE (CCEAI 2019), 2019, 1207
  • [36] SELF-ATTENTION GENERATIVE ADVERSARIAL NETWORK FOR SPEECH ENHANCEMENT
    Huy Phan
    Nguyen, Huy Le
    Chen, Oliver Y.
    Koch, Philipp
    Duong, Ngoc Q. K.
    McLoughlin, Ian
    Mertins, Alfred
    2021 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP 2021), 2021, : 7103 - 7107
  • [37] Diversifying Search Results using Self-Attention Network
    Qin, Xubo
    Dou, Zhicheng
    Wen, Ji-Rong
    CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 1265 - 1274
  • [38] Hierarchical Self-Attention Network for Action Localization in Videos
    Pramono, Rizard Renanda Adhi
    Chen, Yie-Tarng
    Fang, Wen-Hsien
    2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2019), 2019, : 61 - 70
  • [39] Self-attention Based Collaborative Neural Network for Recommendation
    Ma, Shengchao
    Zhu, Jinghua
    WIRELESS ALGORITHMS, SYSTEMS, AND APPLICATIONS, WASA 2019, 2019, 11604 : 235 - 246
  • [40] A pagerank self-attention network for traffic flow prediction
    Kang, Ting
    Wang, Huaizhi
    Wu, Ting
    Peng, Jianchun
    Jiang, Hui
    FRONTIERS IN ENERGY RESEARCH, 2022, 10