Deep Learning Multimodal Fusion for Road Network Extraction: Context and Contour Improvement

被引:4
|
作者
Filho Antonio [1 ]
Shimabukuro, Milton [2 ]
Poz, Aluir Dal [1 ]
机构
[1] Sao Paulo State Univ Unesp, Sch Technol & Sci, Dept Cartog, BR-19060900 Presidente Prudente, SP, Brazil
[2] Sao Paulo State Univ, Dept Math & Comp Sci, BR-01049010 Presidente Prudente, SP, Brazil
基金
巴西圣保罗研究基金会;
关键词
Deep learning; depth models; multimodal fusion; road network extraction; Unet; SEMANTIC SEGMENTATION; RGB;
D O I
10.1109/LGRS.2023.3291656
中图分类号
P3 [地球物理学]; P59 [地球化学];
学科分类号
0708 ; 070902 ;
摘要
Road extraction is still a challenging topic for researchers. Currently, deep convolution neural networks are state-of-the-art in road network segmentation and are known for their remarkable ability to explore multilevel contexts. Despite this, the architectures still suffer from occlusion and obstructions that cause discontinuities and omissions in extracted road networks. Generally, these effects are minimized with strategies to obtain the context of the scene and not explore the complementarity of knowledge from a diversity of sources. We propose an early fusion network with RGB and surface model images that provide complementary geometric data to improve road surface extraction. Our results demonstrate that Unet_early reaches 71.01% intersection over union (IoU) and 81.95% F1, and the fusion strategy increases the IoU and F1 scores at 2.3% and 1.5%, respectively. Besides, it overpassed the best model without fusion (DeepLabv3+). The Brazilian dataset and architecture implementation are available at https://github.com/tunofilho/ieee_road_multimodal.
引用
收藏
页数:5
相关论文
共 50 条
  • [1] FRCFNet: Feature Reassembly and Context Information Fusion Network for Road Extraction
    Wang, Haijuan
    Bai, Lin
    Xue, Danni
    Momi, Moslema Chowdhuray
    Ye, Zhen
    Quan, Siwen
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2024, 21 : 1 - 5
  • [2] Research on Feature Extraction and Multimodal Fusion of Video Caption Based on Deep Learning
    Chen, Hongjun
    Li, Hengyi
    Wu, Xueqin
    2020 THE 4TH INTERNATIONAL CONFERENCE ON MANAGEMENT ENGINEERING, SOFTWARE ENGINEERING AND SERVICE SCIENCES (ICMSS 2020), 2020, : 73 - 76
  • [3] Exploration of Deep Learning-based Multimodal Fusion for Semantic Road Scene Segmentation
    Zhang, Yifei
    Morel, Olivier
    Blanchon, Marc
    Seulin, Ralph
    Rastgoo, Mojdeh
    Sidibe, Desire
    PROCEEDINGS OF THE 14TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS (VISAPP), VOL 5, 2019, : 336 - 343
  • [4] ROAD NETWORK EXTRACTION VIA DEEP LEARNING AND LINE INTEGRAL CONVOLUTION
    Li, Peikang
    Zang, Yu
    Wang, Cheng
    Li, Jonathan
    Cheng, Ming
    Luo, Lun
    Yu, Yao
    2016 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS), 2016, : 1599 - 1602
  • [5] An enhanced multimodal fusion deep learning neural network for lung cancer classification
    Sangeetha, S. K. B.
    Mathivanan, Sandeep Kumar
    Karthikeyan, P.
    Rajadurai, Hariharan
    Shivahare, Basu Dev
    Mallik, Saurav
    Qin, Hong
    SYSTEMS AND SOFT COMPUTING, 2024, 6
  • [6] EXTRACTION OF ROAD NETWORK USING A MODIFIED ACTIVE CONTOUR APPROACH
    Mssedi, Said
    Ben Salah, Mohamed
    Abdelfattah, Riadh
    Mitiche, Amar
    2011 18TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2011,
  • [7] A Survey on Deep Learning for Multimodal Data Fusion
    Gao, Jing
    Li, Peng
    Chen, Zhikui
    Zhang, Jianing
    NEURAL COMPUTATION, 2020, 32 (05) : 829 - 864
  • [8] Advances in deep learning for multimodal fusion and alignment
    Multimedia Tools and Applications, 2022, 81 : 11931 - 11931
  • [9] Deep Learning for HABs Prediction with Multimodal Fusion
    Zhao, Fei
    Zhang, Chengcui
    31ST ACM SIGSPATIAL INTERNATIONAL CONFERENCE ON ADVANCES IN GEOGRAPHIC INFORMATION SYSTEMS, ACM SIGSPATIAL GIS 2023, 2023, : 17 - 18
  • [10] Advances in deep learning for multimodal fusion and alignment
    Huang, Feiran
    Mumtaz, Shahid
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (09) : 11931 - 11931