Surface Material Retrieval Using Weakly Paired Cross-Modal Learning

被引:25
|
作者
Liu, Huaping [1 ,2 ]
Wang, Feng [1 ,2 ]
Sun, Fuchun [1 ,2 ]
Fang, Bin [1 ,2 ]
机构
[1] Tsinghua Univ, Dept Comp Sci & Technol, Beijing 100084, Peoples R China
[2] Tsinghua Univ, Beijing Natl Res Ctr Informat Sci & Technol, State Key Lab Intelligent Technol & Syst, Beijing 100084, Peoples R China
基金
中国国家自然科学基金; 美国国家科学基金会;
关键词
Cross-modal learning; multimodal data; surface material retrieval; MATERIAL RECOGNITION; FUSION;
D O I
10.1109/TASE.2018.2865000
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
In this paper, we investigate the cross-modal material retrieval problem, which permits the user to submit a multimodal query including tactile and auditory modalities, and retrieve the image results of visual modalities. Since multiple significantly different modalities are involved in this process, we encounter more challenges compared with the existing cross-modal retrieval tasks. Our focus is to learn cross-modal representations when the modalities are significantly different and with minimal supervision. A novelty is that we establish a framework that deals with weakly paired multimodal fusion method for heterogenous tactile and auditory modalities and weakly paired cross-modal transfer for visual modality. A structured dictionary learning method with a low rank and common classifier is developed to obtain the modal-invariant representation. Finally, some cross-modal validations on publicly available data sets are performed to show the advantages of the proposed method. Note to Practitioners-Cross-modal retrieval is an important task for industrial intelligence. In this paper, we establish a framework to effectively solve the cross-modal material retrieval problem. In the developed framework, the user may submit a multimodal query including acceleration and sound about an object, and the system may return the most relevant retrieved images. Such a framework may find extensive applications in many fields, because it can be flexible to deal with a multiple-modal query and uses the minimal category label supervision without the need of strong sample pairing information between modalities. Compared with the previous material analysis systems, this paper goes beyond previously proposed surface material classification approaches as it returns an ordered list of perceptually similar surface materials for a query.
引用
收藏
页码:781 / 791
页数:11
相关论文
共 50 条
  • [1] Weakly-paired deep dictionary learning for cross-modal retrieval
    Liu, Huaping
    Wang, Feng
    Zhang, Xinyu
    Sun, Fuchun
    [J]. PATTERN RECOGNITION LETTERS, 2020, 130 : 199 - 206
  • [2] Cross-Modal Surface Material Retrieval Using Discriminant Adversarial Learning
    Zheng, Wendong
    Liu, Huaping
    Wang, Bowen
    Sun, Fuchun
    [J]. IEEE TRANSACTIONS ON INDUSTRIAL INFORMATICS, 2019, 15 (09) : 4978 - 4987
  • [3] Audiovisual cross-modal material surface retrieval
    Liu, Zhuokun
    Liu, Huaping
    Huang, Wenmei
    Wang, Bowen
    Sun, Fuchun
    [J]. NEURAL COMPUTING & APPLICATIONS, 2020, 32 (18): : 14301 - 14309
  • [4] Audiovisual cross-modal material surface retrieval
    Zhuokun Liu
    Huaping Liu
    Wenmei Huang
    Bowen Wang
    Fuchun Sun
    [J]. Neural Computing and Applications, 2020, 32 : 14301 - 14309
  • [5] Cross-Modal Retrieval Using Deep Learning
    Malik, Shaily
    Bhardwaj, Nikhil
    Bhardwaj, Rahul
    Kumar, Saurabh
    [J]. PROCEEDINGS OF THIRD DOCTORAL SYMPOSIUM ON COMPUTATIONAL INTELLIGENCE, DOSCI 2022, 2023, 479 : 725 - 734
  • [6] Online weakly paired similarity learning for surface material retrieval
    Zheng, Wendong
    Liu, Huaping
    Wang, Bowen
    Sun, Fuchun
    [J]. INDUSTRIAL ROBOT-THE INTERNATIONAL JOURNAL OF ROBOTICS RESEARCH AND APPLICATION, 2019, 46 (03): : 396 - 403
  • [7] HCMSL: Hybrid Cross-modal Similarity Learning for Cross-modal Retrieval
    Zhang, Chengyuan
    Song, Jiayu
    Zhu, Xiaofeng
    Zhu, Lei
    Zhang, Shichao
    [J]. ACM TRANSACTIONS ON MULTIMEDIA COMPUTING COMMUNICATIONS AND APPLICATIONS, 2021, 17 (01)
  • [8] Learning DALTS for cross-modal retrieval
    Yu, Zheng
    Wang, Wenmin
    [J]. CAAI TRANSACTIONS ON INTELLIGENCE TECHNOLOGY, 2019, 4 (01) : 9 - 16
  • [9] Continual learning in cross-modal retrieval
    Wang, Kai
    Herranz, Luis
    van de Weijer, Joost
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 3623 - 3633
  • [10] Sequential Learning for Cross-modal Retrieval
    Song, Ge
    Tan, Xiaoyang
    [J]. 2019 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION WORKSHOPS (ICCVW), 2019, : 4531 - 4539