Semantic granularity metric learning for visual search

被引:6
|
作者
Manandhar, Dipu [1 ,3 ]
Bastan, Muhammet [2 ,4 ]
Yap, Kim-Hui [1 ]
机构
[1] Nanyang Technol Univ, Sch Elect & Elect Engn, Singapore, Singapore
[2] Amazon, Palo Alto, CA USA
[3] Univ Surrey, Guildford, Surrey, England
[4] Nanyang Technol Univ, Singapore, Singapore
关键词
Deep learnin; Metric learning; Metric loss functions; Semantic similarity; Visual search; IMAGE SIMILARITY; DEEP; REPRESENTATION;
D O I
10.1016/j.jvcir.2020.102871
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Existing metric learning methods often do not consider different granularly in visual similarly. However, in many domains, images exhibit similarly at multiple granularities with visual semantic concepts, e.g. fashion demonstrates similarly ranging from clothing of the exact same instance to similar looks/design or common category. Therefore, training image triplets/pairs inherently possess different degree of information. Nevertheless, the existing methods often treat them with equal importance which hinder capturing underlying granularities in image similarly. In view of this, we propose a new semantic granularly metric learning (SGML) that develops a novel idea of detecting and leveraging attribute semantic space and integrating it into deep metric learning to capture multiple granularities of similarly. The proposed framework simultaneously learns image attributes and embeddings with multitask-CNN where the tasks are linked by semantic granularly similarly mapping to leverage correlations between the tasks. To this end, we propose a new soft-binomial deviance loss that effectively integrates informativeness of training samples into metric-learning on-the-fly during training. Compared to recent ensemble-based methods, SGML is conceptually elegant, computationally simple yet effective. Extensive experiments on benchmark datasets demonstrate its superiorly e.g., 1-4.5%-Recall@1 improvement over the state-of-the-arts (Kim a al., 2018; Cakir a al., 2019) on DeepFashion-Inshop
引用
收藏
页数:11
相关论文
共 50 条
  • [41] Time Varying Metric Learning for visual tracking
    Li, Jiatong
    Zhao, Baojun
    Deng, Chenwei
    Da Xu, Richard Yi
    PATTERN RECOGNITION LETTERS, 2016, 80 : 157 - 164
  • [42] Individual adaptive metric learning for visual tracking
    Yi, Sihua
    Jiang, Nan
    Wang, Xinggang
    Liu, Wenyu
    NEUROCOMPUTING, 2016, 191 : 273 - 285
  • [43] Learning Adaptive Metric for Robust Visual Tracking
    Jiang, Nan
    Liu, Wenyu
    Wu, Ying
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2011, 20 (08) : 2288 - 2300
  • [44] Effective visual tracking by pairwise metric learning
    Deng, Chenwei
    Wang, Baoxian
    Lin, Weisi
    Huang, Guang-Bin
    Zhao, Baojun
    NEUROCOMPUTING, 2017, 261 : 266 - 275
  • [45] A multi-granularity semisupervised active learning for point cloud semantic segmentation
    Ye, Shanding
    Yin, Zhe
    Fu, Yongjian
    Lin, Hu
    Pan, Zhijie
    NEURAL COMPUTING & APPLICATIONS, 2023, 35 (21): : 15629 - 15645
  • [46] A Machine Learning Technique for Semantic Search Engine
    Nagarajan, G.
    Thyagharajan, K. K.
    INTERNATIONAL CONFERENCE ON MODELLING OPTIMIZATION AND COMPUTING, 2012, 38 : 2164 - 2171
  • [47] Image Search Via Semantic Hashing Learning
    Sun, Weicheng
    Zhu, Songhao
    Cheng, Yanyun
    2017 29TH CHINESE CONTROL AND DECISION CONFERENCE (CCDC), 2017, : 1986 - 1990
  • [48] Dynamic Metric Learning: Towards a Scalable Metric Space to Accommodate Multiple Semantic Scales
    Sun, Yifan
    Zhu, Yuke
    Zhang, Yuhan
    Zheng, Pengkun
    Qiu, Xi
    Zhang, Chi
    Wei, Yichen
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 5389 - 5398
  • [49] Visual similarity is stronger than semantic similarity in guiding visual search for numbers
    Godwin, Hayward J.
    Hout, Michael C.
    Menneer, Tamaryn
    PSYCHONOMIC BULLETIN & REVIEW, 2014, 21 (03) : 689 - 695
  • [50] Visual similarity is stronger than semantic similarity in guiding visual search for numbers
    Hayward J. Godwin
    Michael C. Hout
    Tamaryn Menneer
    Psychonomic Bulletin & Review, 2014, 21 : 689 - 695