Towards Effective Deep Embedding for Zero-Shot Learning

被引:1
|
作者
Zhang, Lei [1 ]
Wang, Peng [2 ]
Liu, Lingqiao [3 ,4 ]
Shen, Chunhua [3 ,4 ]
Wei, Wei [1 ,5 ,6 ]
Zhang, Yanning [1 ,5 ]
van den Hengel, Anton [3 ,4 ]
机构
[1] Northwestern Polytech Univ, Sch Comp Sci, Shaanxi Prov Key Lab Speech & Image Informat Proc, Xian 710072, Peoples R China
[2] Univ Wollongong, Sch Comp & Informat Technol, Wollongong, NSW 2522, Australia
[3] Univ Adelaide, Sch Comp Sci, Adelaide, SA 5005, Australia
[4] Australian Inst Machine Learning, Adelaide, SA 5005, Australia
[5] Northwestern Polytech Univ, Natl Engn Lab Integrated AeroSp Ground Ocean Big, Sch Comp Sci, Xian 710072, Peoples R China
[6] Northwestern Polytech Univ Shenzhen, Res & Dev Inst, Shenzhen 518057, Peoples R China
基金
中国国家自然科学基金;
关键词
Visualization; Semantics; Training; Testing; Labeling; Computer science; Zero-shot learning; Deep embedding; Deep neural network;
D O I
10.1109/TCSVT.2020.2984666
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
Zero-shot learning (ZSL) can be formulated as a cross-domain matching problem: after being projected into a joint embedding space, a visual sample will match against all candidate class-level semantic descriptions and be assigned to the nearest class. In this process, the embedding space underpins the success of such matching and is crucial for ZSL. In this paper, we conduct an in-depth study on the construction of embedding space for ZSL and posit that an ideal embedding space should satisfy two criteria: intra-class compactness and inter-class separability. While the former encourages the embeddings of visual samples of one class to distribute tightly close to the semantic description embedding of this class, the latter requires embeddings from different classes to be well separated from each other. Towards this goal, we present a simple but effective two-branch network to simultaneously map semantic descriptions and visual samples into a joint space, on which visual embeddings are forced to regress to their class-level semantic embeddings and the embeddings crossing classes are required to be distinguishable by a trainable classifier. Furthermore, we extend our method to a transductive setting to better handle the model bias problem in ZSL (i.e., samples from unseen classes tend to be categorized into seen classes) with minimal extra supervision. Specifically, we propose a pseudo labeling strategy to progressively incorporate the testing samples into the training process and thus balance the model between seen and unseen classes. Experimental results on five standard ZSL datasets show the superior performance of the proposed method and its transductive extension.
引用
收藏
页码:2843 / 2852
页数:10
相关论文
共 50 条
  • [1] Learning a Deep Embedding Model for Zero-Shot Learning
    Zhang, Li
    Xiang, Tao
    Gong, Shaogang
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3010 - 3019
  • [2] Deep Unbiased Embedding Transfer for Zero-Shot Learning
    Jia, Zhen
    Zhang, Zhang
    Wang, Liang
    Shan, Caifeng
    Tan, Tieniu
    [J]. IEEE TRANSACTIONS ON IMAGE PROCESSING, 2020, 29 : 1958 - 1971
  • [3] A Variational Autoencoder with Deep Embedding Model for Generalized Zero-Shot Learning
    Ma, Peirong
    Hu, Xiao
    [J]. THIRTY-FOURTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THE THIRTY-SECOND INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE AND THE TENTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2020, 34 : 11733 - 11740
  • [4] Contrastive Embedding for Generalized Zero-Shot Learning
    Han, Zongyan
    Fu, Zhenyong
    Chen, Shuo
    Yang, Jian
    [J]. 2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 2371 - 2381
  • [5] Transductive Unbiased Embedding for Zero-Shot Learning
    Song, Jie
    Shen, Chengchao
    Yang, Yezhou
    Liu, Yang
    Song, Mingli
    [J]. 2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 1024 - 1033
  • [6] Disentangled Ontology Embedding for Zero-shot Learning
    Geng, Yuxia
    Chen, Jiaoyan
    Zhang, Wen
    Xu, Yajing
    Chen, Zhuo
    Pan, Jeff Z.
    Huang, Yufeng
    Xiong, Feiyu
    Chen, Huajun
    [J]. PROCEEDINGS OF THE 28TH ACM SIGKDD CONFERENCE ON KNOWLEDGE DISCOVERY AND DATA MINING, KDD 2022, 2022, : 443 - 453
  • [7] Towards Open Zero-Shot Learning
    Marmoreo, Federico
    Carrazco, Julio Ivan Davila
    Cavazza, Jacopo
    Murino, Vittorio
    [J]. IMAGE ANALYSIS AND PROCESSING, ICIAP 2022, PT II, 2022, 13232 : 564 - 575
  • [8] Region interaction and attribute embedding for zero-shot learning
    Hu, Zhengwei
    Zhao, Haitao
    Peng, Jingchao
    Gu, Xiaojing
    [J]. INFORMATION SCIENCES, 2022, 609 : 984 - 995
  • [9] Incremental Embedding Learning via Zero-Shot Translation
    Wei, Kun
    Deng, Cheng
    Yang, Xu
    Li, Maosen
    [J]. THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 10254 - 10262
  • [10] ENCYCLOPEDIA ENHANCED SEMANTIC EMBEDDING FOR ZERO-SHOT LEARNING
    Jia, Zhen
    Zhang, Junge
    Huang, Kaiqi
    Tan, Tieniu
    [J]. 2017 24TH IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2017, : 1287 - 1291