Improving generalized zero-shot learning via cluster-based semantic disentangling representation

被引:3
|
作者
Gao, Yi [1 ]
Feng, Wentao [1 ]
Xiao, Rong [1 ]
He, Lihuo [2 ]
He, Zhenan [1 ]
Lv, Jiancheng [1 ]
Tang, Chenwei [1 ]
机构
[1] Sichuan Univ, Coll Comp Sci, Chengdu 610065, Peoples R China
[2] Xidian Univ, Sch Elect Engn, Xian 710071, Peoples R China
基金
美国国家科学基金会;
关键词
Generalized zero-shot learning; Domain shift; Semantic gap; Cluster; Semantic disentangling representation;
D O I
10.1016/j.patcog.2024.110320
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Generalized Zero -Shot Learning (GZSL) aims to recognize both seen and unseen classes by training only the seen classes, in which the instances of unseen classes tend to be biased towards the seen class. In this paper, we propose a Cluster -based Semantic Disentangling Representation (CSDR) method to improve GZSL by alleviating the problems of domain shift and semantic gap. First, we cluster the seen data into multiple clusters, where the samples in each cluster belong to several original seen categories, so as to facilitate finegrained semantic disentangling of visual feature vectors. Then, we introduce representation random swapping and contrastive learning based on the clustering results to realize the disentangling semantic representations of semantic -unspecific, class -shared, and class -unique. The fine-grained semantic disentangling representations show high intra-class similarity and inter -class discriminability, which improve the performance of GZSL by alleviating the problem of domain shift. Finally, we construct the visual -semantic embedding space by the variational auto -encoder and alignment module, which can bridge the semantic gap by generating strongly discriminative unseen class samples. Extensive experimental results on four public data sets prove that our method significantly outperforms state-of-the-art methods in generalized and conventional settings.
引用
收藏
页数:14
相关论文
共 50 条
  • [21] Generalized Zero-Shot Recognition based on Visually Semantic Embedding
    Zhu, Pengkai
    Wang, Hanxiao
    Saligrama, Venkatesh
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 2990 - 2998
  • [22] Learning semantic ambiguities for zero-shot learning
    Hanouti, Celina
    Le Borgne, Herve
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 82 (26) : 40745 - 40759
  • [23] Learning semantic ambiguities for zero-shot learning
    Celina Hanouti
    Hervé Le Borgne
    Multimedia Tools and Applications, 2023, 82 : 40745 - 40759
  • [24] Superclass-aware visual feature disentangling for generalized zero-shot learning
    Niu, Chang
    Shang, Junyuan
    Zhou, Zhiheng
    Yang, Junmei
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 258
  • [25] Generative-based hybrid model with semantic representations for generalized zero-shot learning
    Akdemir, Emre
    Barisci, Necaattin
    SIGNAL IMAGE AND VIDEO PROCESSING, 2025, 19 (01)
  • [26] Contrastive semantic disentanglement in latent space for generalized zero-shot learning
    Fan, Wentao
    Liang, Chen
    Wang, Tian
    KNOWLEDGE-BASED SYSTEMS, 2022, 257
  • [27] Contrastive semantic disentanglement in latent space for generalized zero-shot learning
    Fan, Wentao
    Liang, Chen
    Wang, Tian
    Knowledge-Based Systems, 2022, 257
  • [28] Generation-based contrastive model with semantic alignment for generalized zero-shot learning
    Yang, Jingqi
    Shen, Qi
    Xie, Cheng
    IMAGE AND VISION COMPUTING, 2023, 137
  • [29] Zero-Shot Program Representation Learning
    Cui, Nan
    Jiang, Yuze
    Gu, Xiaodong
    Shen, Beijun
    30TH IEEE/ACM INTERNATIONAL CONFERENCE ON PROGRAM COMPREHENSION (ICPC 2022), 2022, : 60 - 70
  • [30] Zero-Shot Code Representation Learning via Prompt Tuning
    Cui, Nan
    Gu, Xiaodong
    Shen, Beijun
    arXiv,