Deep Learning for HABs Prediction with Multimodal Fusion

被引:0
|
作者
Zhao, Fei [1 ]
Zhang, Chengcui [1 ]
机构
[1] Univ Alabama Birmingham, Birmingham, AL 35294 USA
关键词
Geolocation; Computer Vision; Deep Learning; Harmful Algal Blooms;
D O I
10.1145/3589132.3628370
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Harmful Algal Blooms (HABs) present significant environmental and public health threats. Recent machine learning-based HABs monitoring methods often rely solely on unimodal data, e.g., satellite imagery, overlooking crucial environmental factors such as temperature. Moreover, existing multi-modal approaches grapple with real-time applicability and generalizability challenges due to the use of ensemble methodologies and hard-coded geolocation clusters. Addressing these gaps, this paper presents a novel deep learning model using a single-model-based multi-task framework. This framework is designed to segment water bodies and predict HABs severity levels concurrently, enabling the model to focus on areas of interest, thereby enhancing prediction accuracy. Our model integrates multimodal inputs, i.e., satellite imagery, elevation data, temperature readings, and geolocation details, via a dual-branch architecture: the Satellite-Elevation (SE) branch and the TemperatureGeolocation (TG) branch. Satellite and elevation data in the SE branch, being spatially coherent, assist in water area detection and feature extraction. Meanwhile, the TG branch, using sequential temperature data and geolocation information, captures temporal algal growth patterns and adjusts for temperature variations influenced by regional climatic differences, ensuring the model's adaptability across different geographic regions. Additionally, we propose a geometric multimodal focal loss to further enhance representation learning. On the Tick-Tick Bloom (TTB) dataset, our approach outperforms the SOTA methods by 15.65%.
引用
收藏
页码:17 / 18
页数:2
相关论文
共 50 条
  • [1] Deep Multimodal Fusion for Persuasiveness Prediction
    Nojavanasghari, Behnaz
    Gopinath, Deepak
    Koushik, Jayanth
    Baltrusaitis, Tadas
    Morency, Louis-Philippe
    [J]. ICMI'16: PROCEEDINGS OF THE 18TH ACM INTERNATIONAL CONFERENCE ON MULTIMODAL INTERACTION, 2016, : 284 - 288
  • [2] A Survey on Deep Learning for Multimodal Data Fusion
    Gao, Jing
    Li, Peng
    Chen, Zhikui
    Zhang, Jianing
    [J]. NEURAL COMPUTATION, 2020, 32 (05) : 829 - 864
  • [3] Advances in deep learning for multimodal fusion and alignment
    [J]. Multimedia Tools and Applications, 2022, 81 : 11931 - 11931
  • [4] Advances in deep learning for multimodal fusion and alignment
    Huang, Feiran
    Mumtaz, Shahid
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (09) : 11931 - 11931
  • [5] Soybean yield prediction from UAV using multimodal data fusion and deep learning
    Maimaitijiang, Maitiniyazi
    Sagan, Vasit
    Sidike, Paheding
    Hartling, Sean
    Esposito, Flavin
    Fritschi, Felix B.
    [J]. REMOTE SENSING OF ENVIRONMENT, 2020, 237
  • [6] Exploring Fusion Strategies in Deep Multimodal Affect Prediction
    Patania, Sabrina
    D'Amelio, Alessandro
    Lanzarotti, Raffaella
    [J]. IMAGE ANALYSIS AND PROCESSING, ICIAP 2022, PT II, 2022, 13232 : 730 - 741
  • [7] Visualizing Multimodal Deep Learning for Lesion Prediction
    Gillmann, Christina
    Peter, Lucas
    Schmidt, Carlo
    Saur, Dorothee
    Scheuermann, Gerik
    [J]. IEEE COMPUTER GRAPHICS AND APPLICATIONS, 2021, 41 (05) : 90 - 98
  • [8] Deep Learning with Multimodal Data for Prediction of Glaucoma
    Chen, Jimmy
    Fan, Rui
    Christopher, Mark
    Belghith, Akram
    Bowd, Christopher
    Goldbaum, Michael
    Fazio, Massimo
    Girkin, Christopher
    De Moraes, Gustavo
    Liebmann, Jeffrey
    Zangwill, Linda
    [J]. INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2023, 64 (09)
  • [9] Multimodal Deep Learning for Solar Irradiance Prediction
    Li, Zhuo
    Wang, Kejie
    Li, Chenchen
    Zhao, Miao
    Cao, Jiannong
    [J]. 2019 INTERNATIONAL CONFERENCE ON INTERNET OF THINGS (ITHINGS) AND IEEE GREEN COMPUTING AND COMMUNICATIONS (GREENCOM) AND IEEE CYBER, PHYSICAL AND SOCIAL COMPUTING (CPSCOM) AND IEEE SMART DATA (SMARTDATA), 2019, : 784 - 792
  • [10] Success Prediction on Crowdfunding with Multimodal Deep Learning
    Cheng, Chaoran
    Tan, Fei
    Hou, Xiurui
    Wei, Zhi
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 2158 - 2164