SELF-AUGMENTED MULTI-MODAL FEATURE EMBEDDING

被引:0
|
作者
Matsuo, Shinnosuke [1 ]
Uchida, Seiichi [1 ]
Iwana, Brian Kenji [1 ]
机构
[1] Kyushu Univ, Fukuoka, Japan
关键词
Self-augmented multi-modality; multi-modal embedding; gating neural networks;
D O I
10.1109/ICASSP39728.2021.9413974
中图分类号
O42 [声学];
学科分类号
070206 ; 082403 ;
摘要
Oftentimes, patterns can be represented through different modalities. For example, leaf data can be in the form of images or contours. Handwritten characters can also be either online or offline. To exploit this fact, we propose the use of self-augmentation and combine it with multi-modal feature embedding. In order to take advantage of the complementary information from the different modalities, the self-augmented multi-modal feature embedding employs a shared feature space. Through experimental results on classification with online handwriting and leaf images, we demonstrate that the proposed method can create effective embeddings.
引用
收藏
页码:3995 / 3999
页数:5
相关论文
共 50 条
  • [31] MORE: Multi-mOdal REtrieval Augmented Generative Commonsense Reasoning
    Cui, Wanqing
    Bi, Keping
    Guo, Jiafeng
    Cheng, Xueqi
    FINDINGS OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS: ACL 2024, 2024, : 1178 - 1192
  • [32] SAIL: Self-Augmented Graph Contrastive Learning
    Yu, Lu
    Pei, Shichao
    Ding, Lizhong
    Zhou, Jun
    Li, Longfei
    Zhang, Chuxu
    Zhang, Xiangliang
    THIRTY-SIXTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTY-FOURTH CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE / TWELVETH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2022, : 8927 - 8935
  • [33] Multi-manifold Sparse Graph Embedding for Multi-modal Image Classification
    Li, Jingjing
    Wu, Yue
    Zhao, Jidong
    Lu, Ke
    NEUROCOMPUTING, 2016, 173 : 501 - 510
  • [34] Multi-Modal and Multi-Domain Embedding Learning for Fashion Retrieval and Analysis
    Gu, Xiaoling
    Wong, Yongkang
    Shou, Lidan
    Peng, Pai
    Chen, Gang
    Kankanhalli, Mohan S.
    IEEE TRANSACTIONS ON MULTIMEDIA, 2019, 21 (06) : 1524 - 1537
  • [35] Self-supervised multi-modal fusion network for multi-modal thyroid ultrasound image diagnosis
    Xiang, Zhuo
    Zhuo, Qiuluan
    Zhao, Cheng
    Deng, Xiaofei
    Zhu, Ting
    Wang, Tianfu
    Jiang, Wei
    Lei, Baiying
    COMPUTERS IN BIOLOGY AND MEDICINE, 2022, 150
  • [36] Multi-modal multi-task feature fusion for RGBT tracking
    Cai, Yujue
    Sui, Xiubao
    Gu, Guohua
    INFORMATION FUSION, 2023, 97
  • [37] Multi-modal Remote Sensing Image Description Based on Word Embedding and Self-Attention Mechanism
    Wang, Yuan
    Alifu, Kuerban
    Ma, Hongbing
    Li, Junli
    Halik, Umut
    Lv, Yalong
    3rd International Symposium on Autonomous Systems, ISAS 2019, 2019, : 358 - 363
  • [38] Multi-modal Remote Sensing Image Description Based on Word Embedding and Self-Attention Mechanism
    Wang, Yuan
    Alifu, Kuerban
    Ma, Hongbing
    Li, Junli
    Halik, Umut
    Lv, Yalong
    2019 3RD INTERNATIONAL SYMPOSIUM ON AUTONOMOUS SYSTEMS (ISAS 2019), 2019, : 358 - 363
  • [39] MFST: Multi-Modal Feature Self-Adaptive Transformer for Infrared and Visible Image Fusion
    Liu, Xiangzeng
    Gao, Haojie
    Miao, Qiguang
    Xi, Yue
    Ai, Yunfeng
    Gao, Dingguo
    REMOTE SENSING, 2022, 14 (13)
  • [40] Self-supervised multi-modal feature fusion for predicting early recurrence of hepatocellular carcinoma
    Wang, Sen
    Zhao, Ying
    Li, Jiayi
    Yi, Zongmin
    Li, Jun
    Zuo, Can
    Yao, Yu
    Liu, Ailian
    COMPUTERIZED MEDICAL IMAGING AND GRAPHICS, 2024, 118