A deep multimodal generative and fusion framework for class-imbalanced multimodal data

被引:0
|
作者
Qing Li
Guanyuan Yu
Jun Wang
Yuehao Liu
机构
[1] Southwestern University of Finance and Economics,Fintech Innovation Center and School of Economic Information Engineering
来源
关键词
Multimodal classification; Class-imbalanced data; Deep multimodal generative adversarial network; Deep multimodal hybrid fusion network;
D O I
暂无
中图分类号
学科分类号
摘要
The purpose of multimodal classification is to integrate features from diverse information sources to make decisions. The interactions between different modalities are crucial to this task. However, common strategies in previous studies have been to either concatenate features from various sources into a single compound vector or input them separately into several different classifiers that are then assembled into a single robust classifier to generate the final prediction. Both of these approaches weaken or even ignore the interactions among different feature modalities. In addition, in the case of class-imbalanced data, multimodal classification becomes troublesome. In this study, we propose a deep multimodal generative and fusion framework for multimodal classification with class-imbalanced data. This framework consists of two modules: a deep multimodal generative adversarial network (DMGAN) and a deep multimodal hybrid fusion network (DMHFN). The DMGAN is used to handle the class imbalance problem. The DMHFN identifies fine-grained interactions and integrates different information sources for multimodal classification. Experiments on a faculty homepage dataset show the superiority of our framework compared to several start-of-the-art methods.
引用
下载
收藏
页码:25023 / 25050
页数:27
相关论文
共 50 条
  • [1] A deep multimodal generative and fusion framework for class-imbalanced multimodal data
    Li, Qing
    Yu, Guanyuan
    Wang, Jun
    Liu, Yuehao
    MULTIMEDIA TOOLS AND APPLICATIONS, 2020, 79 (33-34) : 25023 - 25050
  • [2] Deep Multimodal Data Fusion
    Zhao, Fei
    Zhang, Chengcui
    Geng, Baocheng
    ACM COMPUTING SURVEYS, 2024, 56 (09)
  • [3] Hybrid neural network with cost-sensitive support vector machine for class-imbalanced multimodal data
    Kim, Kyung Hye
    Sohn, So Young
    NEURAL NETWORKS, 2020, 130 : 176 - 184
  • [4] A Multi-label Multimodal Deep Learning Framework for Imbalanced Data Classification
    Pouyanfar, Samira
    Wang, Tianyi
    Chen, Shu-Ching
    2019 2ND IEEE CONFERENCE ON MULTIMEDIA INFORMATION PROCESSING AND RETRIEVAL (MIPR 2019), 2019, : 199 - 204
  • [5] A multimodal generative and fusion framework for recognizing faculty homepages
    Yu, Guanyuan
    Li, Qing
    Wang, Jun
    Zhang, Di
    Liu, Yuehao
    INFORMATION SCIENCES, 2020, 525 : 205 - 220
  • [6] A Hybrid Framework for Class-Imbalanced Classification
    Chen, Rui
    Luo, Lailong
    Chen, Yingwen
    Xia, Junxu
    Guo, Deke
    WIRELESS ALGORITHMS, SYSTEMS, AND APPLICATIONS, WASA 2021, PT I, 2021, 12937 : 301 - 313
  • [7] From Abstract to Details: A Generative Multimodal Fusion Framework for Recommendation
    Xiao, Fangxiong
    Deng, Lixi
    Chen, Jingjing
    Ji, Houye
    Yang, Xiaorui
    Ding, Zhuoye
    Long, Bo
    PROCEEDINGS OF THE 30TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2022, 2022,
  • [8] Multimodal and multicontrast image fusion via deep generative models
    Dimitri, Giovanna Maria
    Spasov, Simeon
    Duggento, Andrea
    Passamonti, Luca
    Lio, Pietro
    Toschi, Nicola
    INFORMATION FUSION, 2022, 88 : 146 - 160
  • [9] MultiVI: deep generative model for the integration of multimodal data
    Ashuach, Tal
    Gabitto, Mariano I.
    Koodli, Rohan V.
    Saldi, Giuseppe-Antonio
    Jordan, Michael I.
    Yosef, Nir
    NATURE METHODS, 2023, 20 (08) : 1222 - +
  • [10] MultiVI: deep generative model for the integration of multimodal data
    Tal Ashuach
    Mariano I. Gabitto
    Rohan V. Koodli
    Giuseppe-Antonio Saldi
    Michael I. Jordan
    Nir Yosef
    Nature Methods, 2023, 20 : 1222 - 1231