Collaboration based multi-modal multi-label learning

被引:0
|
作者
Yi Zhang
Yinlong Zhu
Zhecheng Zhang
Chongjung Wang
机构
[1] Nanjing University,Department of Computer Science and Technology, State Key Laboratory for Novel Software Technology
来源
Applied Intelligence | 2022年 / 52卷
关键词
Multi-modal; Multi-label; Collaboration; Label correlations;
D O I
暂无
中图分类号
学科分类号
摘要
Complex objects can be represented as multiple modal features and associated with multiple labels. The major challenge of complex object classification is how to jointly utilize heterogeneous modals in a mutually beneficial way. Besides, how to effectively utilize label correlations is also a challenging issue. Previous methods model the label correlations by requiring that any two label-specific classifiers behave similarly on the same modal if the associated labels are similar. To address the above challenges, we propose a novel modal-oriented deep learning framework named Collaboration based Multi-modal Multi-label Learning (CoM3L). With the help of memory structure in LSTM, CoM3L handles modalities sequentially, which predicts next modal to be extracted and learns label correlations simultaneously. On the one hand, CoM3L can extract the most useful modal sequence, which extracts different modal sequences for different instances. On the other hand, for each label, CoM3L combines the collaboration between its own prediction and the prediction of other labels. Extensive experiments on 5 multi-modal multi-label datasets validate the effectiveness of the proposed CoM3L approach.
引用
收藏
页码:14204 / 14217
页数:13
相关论文
共 50 条
  • [31] FusionM4Net: A multi-stage multi-modal learning algorithm for multi-label skin lesion classification
    Tang, Peng
    Yan, Xintong
    Nan, Yang
    Xiang, Shao
    Krammer, Sebastian
    Lasser, Tobias
    [J]. MEDICAL IMAGE ANALYSIS, 2022, 76
  • [32] Experts Collaboration Learning for Continual Multi-Modal Reasoning
    Xu, Li
    Liu, Jun
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2023, 32 : 5087 - 5098
  • [33] M3LA: A Novel Approach Based on Encoder-Decoder with Attention Framework for Multi-modal Multi-label Learning
    Zhu, Yinlong
    Zhang, Yi
    [J]. 2020 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2020,
  • [34] Imbalanced Multi-Modal Multi-Label Learning for Subcellular Localization Prediction of Human Proteins with Both Single and Multiple Sites
    He, Jianjun
    Gu, Hong
    Liu, Wenqi
    [J]. PLOS ONE, 2012, 7 (06):
  • [35] Deep Multi-Modal Hashing With Semantic Enhancement for Multi-Label Micro-Video Retrieval
    Jing, Peiguang
    Sun, Haoyi
    Nie, Liqiang
    Li, Yun
    Su, Yuting
    [J]. IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (10) : 5080 - 5091
  • [36] Multi-modal bilinear fusion with hybrid attention mechanism for multi-label skin lesion classification
    Wei, Yun
    Ji, Lin
    [J]. MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (24) : 65221 - 65247
  • [37] Context Recognition In-the-Wild: Unified Model for Multi-Modal Sensors and Multi-Label Classification
    Vaizman, Yonatan
    Weibel, Nadir
    Lanckriet, Gert
    [J]. Proceedings of the ACM on Interactive, Mobile, Wearable and Ubiquitous Technologies, 2017, 1 (04)
  • [38] Multi-modal anchor adaptation learning for multi-modal summarization
    Chen, Zhongfeng
    Lu, Zhenyu
    Rong, Huan
    Zhao, Chuanjun
    Xu, Fan
    [J]. NEUROCOMPUTING, 2024, 570
  • [39] Collaboration Based Multi-Label Propagation for Fraud Detection
    Wang, Haobo
    Li, Zhao
    Huang, Jiaming
    Hui, Pengrui
    Liu, Weiwei
    Hu, Tianlei
    Chen, Gang
    [J]. PROCEEDINGS OF THE TWENTY-NINTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2020, : 2477 - 2483
  • [40] M3ED: Multi-modal Multi-scene Multi-label Emotional Dialogue Database
    Zhao, Jinming
    Zhang, Tenggan
    Hu, Jingwen
    Liu, Yuchen
    Jin, Qin
    Wang, Xinchao
    Li, Haizhou
    [J]. PROCEEDINGS OF THE 60TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2022), VOL 1: (LONG PAPERS), 2022, : 5699 - 5710