LEARNING UNIFIED SPARSE REPRESENTATIONS FOR MULTI-MODAL DATA

被引:0
|
作者
Wang, Kaiye [1 ]
Wang, Wei [1 ]
Wang, Liang [1 ]
机构
[1] Chinese Acad Sci, Natl Lab Pattern Recognit, Ctr Res Intelligent Percept & Comp, Inst Automat, Beijing 100190, Peoples R China
关键词
Cross-modal retrieval; unified representation learning; joint dictionary learning; multi-modal data;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Cross-modal retrieval has become one of interesting and important research problem recently, where users can take one modality of data (e.g., text, image or video) as the query to retrieve relevant data of another modality. In this paper, we present a Multi-modal Unified Representation Learning (MURL) algorithm for cross-modal retrieval, which learns unified sparse representations for multi-modal data representing the same semantics via joint dictionary learning. The l(1)-norm is imposed on the unified representations to explicitly encourage sparsity, which makes our algorithm more robust. Furthermore, a constraint regularization term is imposed to force the representations to be similar if their corresponding multi-modal data have must-links or to be far apart if their corresponding multi-modal data have cannot-links. An iterative algorithm is also proposed to solve the objective function. The effectiveness of the proposed method is verified by extensive results on two real-world datasets.
引用
收藏
页码:3545 / 3549
页数:5
相关论文
共 50 条
  • [41] Multi-modal remote perception learning for object sensory data
    Almujally, Nouf Abdullah
    Rafique, Adnan Ahmed
    Al Mudawi, Naif
    Alazeb, Abdulwahab
    Alonazi, Mohammed
    Algarni, Asaad
    Jalal, Ahmad
    Liu, Hui
    FRONTIERS IN NEUROROBOTICS, 2024, 18
  • [42] Longitudinal and Multi-modal Data Learning via Joint Embedding and Sparse Regression for Parkinson's Disease Diagnosis
    Lei, Haijun
    Huang, Zhongwei
    Elazab, Ahmed
    Li, Hancong
    Lei, Baiying
    MACHINE LEARNING IN MEDICAL IMAGING: 9TH INTERNATIONAL WORKSHOP, MLMI 2018, 2018, 11046 : 310 - 318
  • [43] Unified reconstruction framework for multi-modal medical imaging
    Dong, Di
    Tian, Jie
    Dai, Yakang
    Yan, Guorui
    Yang, Fei
    Wu, Ping
    JOURNAL OF X-RAY SCIENCE AND TECHNOLOGY, 2011, 19 (01) : 111 - 126
  • [44] A Unified Framework for Multi-Modal Isolated Gesture Recognition
    Duan, Jiali
    Wan, Jun
    Zhou, Shuai
    Guo, Xiaoyuan
    Li, Stan Z.
    ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2018, 14 (01)
  • [45] Overview of Uni-modal and Multi-modal Representations for Classification Tasks
    Wiesen, Aryeh
    HaCohen-Kerner, Yaakov
    NATURAL LANGUAGE PROCESSING AND INFORMATION SYSTEMS (NLDB 2018), 2018, 10859 : 397 - 404
  • [46] UniColor: A Unified Framework for Multi-Modal Colorization with Transformer
    Huang, Zhitong
    Zhao, Nanxuan
    Liao, Jing
    ACM TRANSACTIONS ON GRAPHICS, 2022, 41 (06):
  • [47] Unified Multi-modal Learning for Any Modality Combinations in Alzheimer's Disease Diagnosis
    Feng, Yidan
    Gao, Bingchen
    Deng, Sen
    Qiu, Anqi
    Qin, Jing
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT III, 2024, 15003 : 487 - 497
  • [48] Unified losses for multi-modal pose coding and regression
    Johnson, Leif
    Cooper, Joseph
    Ballard, Dana
    2013 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2013,
  • [49] Multi-modal and multi-granular learning
    Zhang, Bo
    Zhang, Ling
    ADVANCES IN KNOWLEDGE DISCOVERY AND DATA MINING, PROCEEDINGS, 2007, 4426 : 9 - +
  • [50] Unified Multi-modal Unsupervised Representation Learning for Skeleton-based Action Understanding
    Sun, Shengkai
    Liu, Daizong
    Dong, Jianfeng
    Qu, Xiaoye
    Gao, Junyu
    Yang, Xun
    Wang, Xun
    Wang, Meng
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 2973 - 2984