DED-SAM:Adapting Segment Anything Model 2 for Dual Encoder-Decoder Change Detection

被引:1
|
作者
Qiu, Junlong [1 ]
Liu, Wei [1 ]
Zhang, Xin [2 ]
Li, Erzhu [1 ]
Zhang, Lianpeng [1 ]
Li, Xing [1 ]
机构
[1] Jiangsu Normal Univ, Sch Geog Geomat & Planning, Xuzhou 221116, Peoples R China
[2] Chinese Acad Sci, Aerosp Informat Res Inst, Beijing 100094, Peoples R China
基金
国家重点研发计划; 中国国家自然科学基金;
关键词
Feature extraction; Remote sensing; Decoding; Data models; Visualization; Transformers; Object oriented modeling; Accuracy; Semantics; Semantic segmentation; Change detection; remote sensing; segment anything model; vision foundation model; NETWORK;
D O I
10.1109/JSTARS.2024.3490754
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Change detection has become a crucial topic in the field of remote sensing deep learning due to its extensive application in earth observation. However, real remote sensing images often contain multiple land cover classes with significant intraclass variability and interclass similarity, limiting the performance of change detection in complex scenarios. To address this, we leverage the capabilities of vision foundation models by applying the segment anything model (SAM) to remote sensing change detection, and we name this method dual encoder-decoder SAM (DED-SAM). Specifically, we construct a DED framework, utilizing a small-scale change detection model in both branches to generate mixed prompts including image features, mask prompts, and box prompts. The SAM 2 model is used for fine-grained recognition of dual-temporal images, generating accurate, and stable feature boundaries, which are then used as constraints to generate the final change mask. To validate the effectiveness of DED-SAM across various application scenarios, we conduct quantitative experiments on three public datasets: Levir-CD, SYSU-CD, and CDD, testing its detection capabilities under single change categories, multiple change categories, and seasonal pseudochange interference. The results show that the proposed DED-SAM achieved state-of-the-art F1 scores and IoUs on these three datasets: LEVIR-CD (92.00%, 85.11%), SYSU-CD (84.15%, 72.01%), and CDD (97.72%, 95.47%).
引用
收藏
页码:995 / 1006
页数:12
相关论文
共 50 条
  • [41] Crater Detection and Population Statistics in Tianwen-1 Landing Area Based on Segment Anything Model (SAM)
    Zhao, Yaqi
    Ye, Hongxia
    REMOTE SENSING, 2024, 16 (10)
  • [42] ED2LM: Encoder-Decoder to Language Model for Faster Document Re-ranking Inference
    Hui, Kai
    Zhuang, Honglei
    Chen, Tao
    Qin, Zhen
    Lu, Jing
    Bahri, Dara
    Ma, Ji
    Gupta, Jai
    dos Santos, Cicero Nogueira
    Tay, Yi
    Metzler, Donald
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), 2022, : 3747 - 3758
  • [43] An encoder-decoder ConvLSTM surrogate model for simulating geological CO2 sequestration with dynamic well controls
    Feng, Zhao
    Tariq, Zeeshan
    Shen, Xianda
    Yan, Bicheng
    Tang, Xuhai
    Zhang, Fengshou
    GAS SCIENCE AND ENGINEERING, 2024, 125
  • [44] Dual attention guided deep encoder-decoder network for change analysis in land use/land cover for Dakshina Kannada District, Karnataka, India
    Nitesh Naik
    Kandasamy Chandrasekaran
    Venkatesan Meenakshi Sundaram
    Prabhavathy Panneer
    Environmental Earth Sciences, 2023, 82
  • [45] Dual attention guided deep encoder-decoder network for change analysis in land use/land cover for Dakshina Kannada District, Karnataka, India
    Naik, Nitesh
    Chandrasekaran, Kandasamy
    Sundaram, Venkatesan Meenakshi
    Panneer, Prabhavathy
    ENVIRONMENTAL EARTH SCIENCES, 2023, 82 (01)
  • [46] A grid-level segmentation model based on encoder-decoder structure with multi-source features for crop lodging detection
    Wang, Lihui
    Xiao, Huidi
    APPLIED SOFT COMPUTING, 2024, 151
  • [47] EM-SAM: Eye-Movement-Guided Segment Anything Model for Object Detection and Recognition in Complex Scenes
    Li, Jinqi
    Yu, Yang
    Zhou, Junfan
    Wang, Chinan
    Ling-li Zeng
    2024 WRC SYMPOSIUM ON ADVANCED ROBOTICS AND AUTOMATION, WRC SARA, 2024, : 401 - 408
  • [48] Seq2Code: Transformer-Based Encoder-Decoder Model for Python']Python Source Code Generation
    Laskari, Naveen Kumar
    Reddy, K. Adi Narayana
    Reddy, M. Indrasena
    THIRD CONGRESS ON INTELLIGENT SYSTEMS, CIS 2022, VOL 1, 2023, 608 : 301 - 309
  • [49] High Quality Image Steganography Model Based on Encoder-Decoder Networks and 2D Logistic Chaotic Encryption
    Zhang, Qiu-Yu
    Hu, Xue-Wen
    Wang, Zhen
    International Journal of Network Security, 2023, 25 (03) : 394 - 408
  • [50] Road-SAM: Adapting the Segment Anything Model to Road Extraction From Large Very-High-Resolution Optical Remote Sensing Images
    Feng, Wenqing
    Guan, Fangli
    Sun, Chenhao
    Xu, Wei
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21