Reliable Multi-modal Learning: A Survey

被引:0
|
作者
Yang, Yang [1 ]
Zhan, De-Chuan [2 ]
Jiang, Yuan [2 ]
Xiong, Hui [3 ]
机构
[1] School of Computer Science and Engineering, Nanjing University of Science and Technology, Nanjing,210094, China
[2] State Key Laboratory for Novel Software Technology, Nanjing University, Nanjing,210023, China
[3] Rutgers Business School, Newark,NJ,07012, United States
来源
Ruan Jian Xue Bao/Journal of Software | 2021年 / 32卷 / 04期
基金
中国国家自然科学基金;
关键词
Modal analysis - Learning systems;
D O I
10.13328/j.cnki.jos.006167
中图分类号
学科分类号
摘要
Recently, multi-modal learning is one of the important research fields of machine learning and data mining, and it has a wide range of practical applications, such as cross-media search, multi-language processing, auxiliary information click-through rate estimation, etc. Traditional multi-modal learning methods usually use the consistency or complementarity among modalities to design corresponding loss functions or regularization terms for joint training, thereby improving the single-modal and ensemble performance. However, in the open environment, affected by factors such as data missing and noise, multi-modal data is imbalanced, specifically manifested as insufficient or incomplete, resulting in inconsistency modal feature representations and inconsistent modal alignment relationships. Direct use of traditional multi-modal methods will even degrade single-modal and ensemble performance. To solve these problems, reliable multi-modal learning has been proposed and studied. This paper systematically summarizes and analyzes the progress made by domestic and international scholars on reliable multi-modal research, and the challenges that future research may face. © Copyright 2021, Institute of Software, the Chinese Academy of Sciences. All rights reserved.
引用
收藏
页码:1067 / 1081
相关论文
共 50 条
  • [1] A survey of multi-modal learning theory(英文)
    HUANG Yu
    HUANG Longbo
    [J]. 中山大学学报(自然科学版)(中英文), 2023, 62 (05) : 38 - 49
  • [2] Multi-modal anchor adaptation learning for multi-modal summarization
    Chen, Zhongfeng
    Lu, Zhenyu
    Rong, Huan
    Zhao, Chuanjun
    Xu, Fan
    [J]. NEUROCOMPUTING, 2024, 570
  • [3] A Survey on Multi-modal Summarization
    Jangra, Anubhav
    Mukherjee, Sourajit
    Jatowt, Adam
    Saha, Sriparna
    Hasanuzzaman, Mohammad
    [J]. ACM COMPUTING SURVEYS, 2023, 55 (13S)
  • [4] Unsupervised Multi-modal Learning
    Iqbal, Mohammed Shameer
    [J]. ADVANCES IN ARTIFICIAL INTELLIGENCE (AI 2015), 2015, 9091 : 343 - 346
  • [5] Learning Multi-modal Similarity
    McFee, Brian
    Lanckriet, Gert
    [J]. JOURNAL OF MACHINE LEARNING RESEARCH, 2011, 12 : 491 - 523
  • [6] Multi-modal and multi-granular learning
    Zhang, Bo
    Zhang, Ling
    [J]. ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PROCEEDINGS, 2007, 4426 : 9 - +
  • [7] Learning in an Inclusive Multi-Modal Environment
    Graham, Deryn
    Benest, Ian
    Nicholl, Peter
    [J]. JOURNAL OF CASES ON INFORMATION TECHNOLOGY, 2010, 12 (03) : 28 - 44
  • [8] Learning of Multi-Modal Stimuli in Hawkmoths
    Balkenius, Anna
    Dacke, Marie
    [J]. PLOS ONE, 2013, 8 (07):
  • [9] Multi-Modal Meta Continual Learning
    Gai, Sibo
    Chen, Zhengyu
    Wang, Donglin
    [J]. 2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [10] Imagery in multi-modal object learning
    Jüttner, M
    Rentschler, I
    [J]. BEHAVIORAL AND BRAIN SCIENCES, 2002, 25 (02) : 197 - +