Semantic granularity metric learning for visual search

被引:6
|
作者
Manandhar, Dipu [1 ,3 ]
Bastan, Muhammet [2 ,4 ]
Yap, Kim-Hui [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore, Singapore
[2] Amazon, Palo Alto, CA USA
[3] Univ Surrey, Guildford, Surrey, England
[4] Nanyang Technol Univ, Singapore, Singapore
关键词
Deep learnin; Metric learning; Metric loss functions; Semantic similarity; Visual search; IMAGE SIMILARITY; DEEP; REPRESENTATION;
D O I
10.1016/j.jvcir.2020.102871
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Existing metric learning methods often do not consider different granularly in visual similarly. However, in many domains, images exhibit similarly at multiple granularities with visual semantic concepts, e.g. fashion demonstrates similarly ranging from clothing of the exact same instance to similar looks/design or common category. Therefore, training image triplets/pairs inherently possess different degree of information. Nevertheless, the existing methods often treat them with equal importance which hinder capturing underlying granularities in image similarly. In view of this, we propose a new semantic granularly metric learning (SGML) that develops a novel idea of detecting and leveraging attribute semantic space and integrating it into deep metric learning to capture multiple granularities of similarly. The proposed framework simultaneously learns image attributes and embeddings with multitask-CNN where the tasks are linked by semantic granularly similarly mapping to leverage correlations between the tasks. To this end, we propose a new soft-binomial deviance loss that effectively integrates informativeness of training samples into metric-learning on-the-fly during training. Compared to recent ensemble-based methods, SGML is conceptually elegant, computationally simple yet effective. Extensive experiments on benchmark datasets demonstrate its superiorly e.g., 1-4.5%-Recall@1 improvement over the state-of-the-arts (Kim a al., 2018; Cakir a al., 2019) on DeepFashion-Inshop
引用
收藏
页数:11
相关论文
共 50 条
  • [31] SEMANTIC EDGE DETECTION BASED ON DEEP METRIC LEARNING
    Cai, Shulian
    Huang, Jiabin
    Ding, Xinghao
    Zeng, Delu
    2017 INTERNATIONAL SYMPOSIUM ON INTELLIGENT SIGNAL PROCESSING AND COMMUNICATION SYSTEMS (ISPACS 2017), 2017, : 707 - 712
  • [32] Metric Learning for Semantic-Based Clothes Retrieval
    YANG Bo
    GUO Caili
    LI Zheng
    ZTE Communications, 2022, 20 (01) : 76 - 82
  • [33] Image Annotation Using Metric Learning in Semantic Neighbourhoods
    Verma, Yashaswi
    Jawahar, C. V.
    COMPUTER VISION - ECCV 2012, PT III, 2012, 7574 : 836 - 849
  • [34] Semantic Discriminative Metric Learning for Image Similarity Measurement
    Wang, Huibing
    Feng, Lin
    Zhang, Jing
    Liu, Yang
    IEEE TRANSACTIONS ON MULTIMEDIA, 2016, 18 (08) : 1579 - 1589
  • [35] Context enabled semantic granularity
    Albertoni, Riccardo
    Camossi, Elena
    De Martino, Monica
    Giannini, Franca
    Monti, Marina
    KNOWLEDGE-BASED INTELLIGENT INFORMATION AND ENGINEERING SYSTEMS, PT 2, PROCEEDINGS, 2008, 5178 : 682 - +
  • [36] Culture-specific strategies of visual semantic search
    Blinnikova, Irina
    Rabeson, Maria
    INTERNATIONAL JOURNAL OF PSYCHOLOGY, 2023, 58 : 217 - 217
  • [37] NBSearch: Semantic Search and Visual Exploration of Computational Notebooks
    Li, Xingjun
    Wang, Yuanxin
    Wang, Hong
    Wang, Yang
    Zhao, Jian
    CHI '21: PROCEEDINGS OF THE 2021 CHI CONFERENCE ON HUMAN FACTORS IN COMPUTING SYSTEMS, 2021,
  • [38] How do emotions affect visual semantic search?
    Blinnikova, Irina
    Izmalkova, Anna
    Marchenko, Olga
    PERCEPTION, 2015, 44 : 33 - 33
  • [39] Semantic Linking Maps for Active Visual Object Search
    Zeng, Zhen
    Roefer, Adrian
    Jenkins, Odest Chadwicke
    2020 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION (ICRA), 2020, : 1984 - 1990
  • [40] A multi-granularity semisupervised active learning for point cloud semantic segmentation
    Shanding Ye
    Zhe Yin
    Yongjian Fu
    Hu Lin
    Zhijie Pan
    Neural Computing and Applications, 2023, 35 : 15629 - 15645