LEARNING UNIFIED SPARSE REPRESENTATIONS FOR MULTI-MODAL DATA

被引:0
|
作者
Wang, Kaiye [1 ]
Wang, Wei [1 ]
Wang, Liang [1 ]
机构
[1] Chinese Acad Sci, Natl Lab Pattern Recognit, Ctr Res Intelligent Percept & Comp, Inst Automat, Beijing 100190, Peoples R China
关键词
Cross-modal retrieval; unified representation learning; joint dictionary learning; multi-modal data;
D O I
暂无
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Cross-modal retrieval has become one of interesting and important research problem recently, where users can take one modality of data (e.g., text, image or video) as the query to retrieve relevant data of another modality. In this paper, we present a Multi-modal Unified Representation Learning (MURL) algorithm for cross-modal retrieval, which learns unified sparse representations for multi-modal data representing the same semantics via joint dictionary learning. The l(1)-norm is imposed on the unified representations to explicitly encourage sparsity, which makes our algorithm more robust. Furthermore, a constraint regularization term is imposed to force the representations to be similar if their corresponding multi-modal data have must-links or to be far apart if their corresponding multi-modal data have cannot-links. An iterative algorithm is also proposed to solve the objective function. The effectiveness of the proposed method is verified by extensive results on two real-world datasets.
引用
收藏
页码:3545 / 3549
页数:5
相关论文
共 50 条
  • [1] Fast Multi-Modal Unified Sparse Representation Learning
    Verma, Mridula
    Shukla, Kaushal Kumar
    PROCEEDINGS OF THE 2017 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR'17), 2017, : 448 - 452
  • [2] Learning Common and Transferable Feature Representations for Multi-Modal Data
    Nitsch, Julia
    Nieto, Juan
    Siegwart, Roland
    Schmidt, Max
    Cadena, Cesar
    2020 IEEE INTELLIGENT VEHICLES SYMPOSIUM (IV), 2020, : 1595 - 1601
  • [3] A Unified Deep Learning Framework for Multi-Modal Multi-Dimensional Data
    Xi, Pengcheng
    Goubran, Rafik
    Shu, Chang
    2019 IEEE INTERNATIONAL SYMPOSIUM ON MEDICAL MEASUREMENTS AND APPLICATIONS (MEMEA), 2019,
  • [4] A unified framework for multi-modal federated learning
    Xiong, Baochen
    Yang, Xiaoshan
    Qi, Fan
    Xu, Changsheng
    NEUROCOMPUTING, 2022, 480 : 110 - 118
  • [5] Multi-Modal Representations for Improved Bilingual Lexicon Learning
    Vulic, Ivan
    Kiela, Douwe
    Clark, Stephen
    Moens, Marie-Francine
    PROCEEDINGS OF THE 54TH ANNUAL MEETING OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS (ACL 2016), VOL 2, 2016, : 188 - 194
  • [6] Sparse Relational Topical Coding on multi-modal data
    Song, Lingyun
    Liu, Jun
    Luo, Minnan
    Qian, Buyue
    Yang, Kuan
    PATTERN RECOGNITION, 2017, 72 : 368 - 380
  • [7] Low-Rank and Joint Sparse Representations for Multi-Modal Recognition
    Zhang, Heng
    Patel, Vishal M.
    Chellappa, Rama
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2017, 26 (10) : 4741 - 4752
  • [8] Sparse Multi-Modal Hashing
    Wu, Fei
    Yu, Zhou
    Yang, Yi
    Tang, Siliang
    Zhang, Yin
    Zhuang, Yueting
    IEEE TRANSACTIONS ON MULTIMEDIA, 2014, 16 (02) : 427 - 439
  • [9] METEOR: Learning Memory and Time Efficient Representations from Multi-modal Data Streams
    Silva, Amila
    Karunasekera, Shanika
    Leckie, Christopher
    Luo, Ling
    CIKM '20: PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON INFORMATION & KNOWLEDGE MANAGEMENT, 2020, : 1375 - 1384
  • [10] Multi-Modal Transportation Recommendation with Unified Route Representation Learning
    Liu, Hao
    Han, Jindong
    Fu, Yanjie
    Zhou, Jingbo
    Lu, Xinjiang
    Xiong, Hui
    PROCEEDINGS OF THE VLDB ENDOWMENT, 2020, 14 (03): : 342 - 350