Semantic Autoencoder for Zero-Shot Learning

被引:563
|
作者
Kodirov, Elyor [1 ]
Xiang, Tao [1 ]
Gong, Shaogang [1 ]
机构
[1] Queen Mary Univ London, London, England
基金
欧洲研究理事会;
关键词
RECOGNITION;
D O I
10.1109/CVPR.2017.473
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Existing zero-shot learning (ZSL) models typically learn a projection function from a feature space to a semantic embedding space (e.g. attribute space). However, such a projection function is only concerned with predicting the training seen class semantic representation (e.g. attribute prediction) or classification. When applied to test data, which in the context of ZSL contains different (unseen) classes without training data, a ZSL model typically suffers from the project domain shift problem. In this work, we present a novel solution to ZSL based on learning a Semantic AutoEncoder (SAE). Taking the encoder-decoder paradigm, an encoder aims to project a visual feature vector into the semantic space as in the existing ZSL models. However, the decoder exerts an additional constraint, that is, the projection/code must be able to reconstruct the original visual feature. We show that with this additional reconstruction constraint, the learned projection function from the seen classes is able to generalise better to the new unseen classes. Importantly, the encoder and decoder are linear and symmetric which enable us to develop an extremely efficient learning algorithm. Extensive experiments on six benchmark datasets demonstrate that the proposed SAE outperforms significantly the existing ZSL models with the additional benefit of lower computational cost. Furthermore, when the SAE is applied to supervised clustering problem, it also beats the state-of-the-art.
引用
收藏
页码:4447 / 4456
页数:10
相关论文
共 50 条
  • [21] Swap-Reconstruction Autoencoder for Compositional Zero-Shot Learning
    Guo, Ting
    Liang, Jiye
    Xie, Guo-Sen
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 438 - 443
  • [22] Discriminative Embedding Autoencoder With a Regressor Feedback for Zero-Shot Learning
    Shi, Ying
    Wei, Wei
    [J]. IEEE ACCESS, 2020, 8 : 11019 - 11030
  • [23] Graph and Autoencoder Based Feature Extraction for Zero-shot Learning
    Liu, Yang
    Xie, Deyan
    Gao, Quanxue
    Han, Jungong
    Wang, Shujian
    Gao, Xinbo
    [J]. PROCEEDINGS OF THE TWENTY-EIGHTH INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, 2019, : 3038 - 3044
  • [24] Learning exclusive discriminative semantic information for zero-shot learning
    Mi, Jian-Xun
    Zhang, Zhonghao
    Tai, Debao
    Zhou, Li-Fang
    Jia, Wei
    [J]. INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2023, 14 (03) : 761 - 772
  • [25] A meaningful learning method for zero-shot semantic segmentation
    Liu, Xianglong
    Bai, Shihao
    An, Shan
    Wang, Shuo
    Liu, Wei
    Zhao, Xiaowei
    Ma, Yuqing
    [J]. SCIENCE CHINA-INFORMATION SCIENCES, 2023, 66 (11)
  • [26] Zero-Shot Learning on Semantic Class Prototype Graph
    Fu, Zhenyong
    Xiang, Tao
    Kodirov, Elyor
    Gong, Shaogang
    [J]. IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2018, 40 (08) : 2009 - 2022
  • [27] Attentive Semantic Preservation Network for Zero-Shot Learning
    Lu, Ziqian
    Yu, Yunlong
    Lu, Zhe-Ming
    Shen, Feng-Li
    Zhang, Zhongfei
    [J]. 2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 2919 - 2925
  • [28] Joint Visual and Semantic Optimization for zero-shot learning
    Wu, Hanrui
    Yan, Yuguang
    Chen, Sentao
    Huang, Xiangkang
    Wu, Qingyao
    Ng, Michael K.
    [J]. KNOWLEDGE-BASED SYSTEMS, 2021, 215 (215)
  • [29] Semantic embeddings of generic objects for zero-shot learning
    Hascoet, Tristan
    Ariki, Yasuo
    Takiguchi, Tetsuya
    [J]. EURASIP JOURNAL ON IMAGE AND VIDEO PROCESSING, 2019, 2019 (1)
  • [30] Zero-Shot Classification with Discriminative Semantic Representation Learning
    Ye, Meng
    Guo, Yuhong
    [J]. 30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 5103 - 5111