From Region to Patch: Attribute-Aware Foreground-Background Contrastive Learning for Fine-Grained Fashion Retrieval

被引:7
|
作者
Dong, Jianfeng [1 ]
Peng, Xiaoman [2 ]
Ma, Zhe [3 ]
Liu, Daizong [4 ]
Qu, Xiaoye [5 ]
Yang, Xun [6 ]
Zhu, Jixiang [2 ]
Liu, Baolong [1 ]
机构
[1] Zhejiang Gongshang Univ, Zhejiang Key Lab E Commerce, Hangzhou, Peoples R China
[2] Zhejiang Gongshang Univ, Hangzhou, Peoples R China
[3] Zhejiang Univ, Hangzhou, Peoples R China
[4] Peking Univ, Beijing, Peoples R China
[5] Huazhong Univ Sci & Technol, Wuhan, Peoples R China
[6] Univ Sci & Technol China, Hefei, Peoples R China
来源
PROCEEDINGS OF THE 46TH INTERNATIONAL ACM SIGIR CONFERENCE ON RESEARCH AND DEVELOPMENT IN INFORMATION RETRIEVAL, SIGIR 2023 | 2023年
关键词
Fashion Retrieval; Fine-Grained Similarity; Image Retrieval;
D O I
10.1145/3539618.3591690
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Attribute-specific fashion retrieval (ASFR) is a challenging information retrieval task, which has attracted increasing attention in recent years. Different from traditional fashion retrieval which mainly focuses on optimizing holistic similarity, the ASFR task concentrates on attribute-specific similarity, resulting in more finegrained and interpretable retrieval results. As the attribute-specific similarity typically corresponds to the specific subtle regions of images, we propose a Region-to-Patch Framework (RPF) that consists of a region-aware branch and a patch-aware branch to extract fine-grained attribute-related visual features for precise retrieval in a coarse-to-fine manner. In particular, the region-aware branch is first to be utilized to locate the potential regions related to the semantic of the given attribute. Then, considering that the located region is coarse and still contains the background visual contents, the patch-aware branch is proposed to capture patch-wise attributerelated details from the previous amplified region. Such a hybrid architecture strikes a proper balance between region localization and feature extraction. Besides, different from previous works that solely focus on discriminating the attribute-relevant foreground visual features, we argue that the attribute-irrelevant background features are also crucial for distinguishing the detailed visual contexts in a contrastive manner. Therefore, a novel E-InfoNCE loss based on the foreground and background representations is further proposed to improve the discrimination of attribute-specific representation. Extensive experiments on three datasets demonstrate the effectiveness of our proposed framework, and also show a decent generalization of our RPF on out-of-domain fashion images. Our source code is available at https://github.com/HuiGuanLab/RPF.
引用
收藏
页码:1273 / 1282
页数:10
相关论文
共 25 条
  • [21] Fine-grained Semantics-aware Representation Learning for Text-based Person Retrieval
    Wang, Di
    Yan, Feng
    Wang, Yifeng
    Zhao, Lin
    Liang, Xiao
    Zhong, Haodi
    Zhang, Ronghua
    PROCEEDINGS OF THE 4TH ANNUAL ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL, ICMR 2024, 2024, : 92 - 100
  • [22] Adaptive Multi-Dimensional Weighted Network With Category-Aware Contrastive Learning for Fine-Grained Hand Bone Segmentation
    Zeng, Bolun
    Chen, Li
    Zheng, Yuanyi
    Chen, Xiaojun
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2024, 28 (07) : 3985 - 3996
  • [23] Region-Based Global-Local Contrastive Representation Learning for Fine-Grained Ship Classification in Remote Sensing Images
    Li, Kun
    Liu, Zhunga
    Zhang, Zuowei
    IEEE TRANSACTIONS ON AEROSPACE AND ELECTRONIC SYSTEMS, 2025, 61 (02) : 2631 - 2643
  • [24] Partial-Label Contrastive Representation Learning for Fine-Grained Biomarkers Prediction From Histopathology Whole Slide Images
    Zheng, Yushan
    Wu, Kun
    Li, Jun
    Tang, Kunming
    Shi, Jun
    Wu, Haibo
    Jiang, Zhiguo
    Wang, Wei
    IEEE JOURNAL OF BIOMEDICAL AND HEALTH INFORMATICS, 2025, 29 (01) : 396 - 408
  • [25] Alleviating Over-Fitting in Hashing-Based Fine-Grained Image Retrieval: From Causal Feature Learning to Binary-Injected Hash Learning
    Xiang, Xinguang
    Ding, Xinhao
    Jin, Lu
    Li, Zechao
    Tang, Jinhui
    Jain, Ramesh
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 10665 - 10677