SVDML: Semantic and Visual Space Deep Mutual Learning for Zero-Shot Learning

被引:0
|
作者
Lu, Nannan [1 ]
Luo, Yi [1 ]
Qiu, Mingkai [1 ]
机构
[1] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou 221100, Jiangsu, Peoples R China
基金
中国国家自然科学基金;
关键词
Zero-shot Learning; Semantic Representation; Visual Representation; Mutual Learning;
D O I
10.1007/978-981-99-8546-3_31
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The key challenge of zero-shot learning (ZSL) is how to identify novel objects for which no samples are available during the training process. Current approaches either align the global features of images to the corresponding class semantic vectors or use unidirectional attentions to locate the local visual features of images via semantic attributes to avoid interference from other noise in the image. However, they still have not found a way to establish a robust correlation between the semantic and visual representation. To solve the issue, we propose a Semantic and Visual space Deep Mutual Learning (SVDML), which consists of three modules: class representation learning, attribute embedding, and mutual learning, to establish the intrinsic semantic relations between visual features and attribute features. SVDML uses two kinds of prototype generators to separately guide the learning of global and local features of images and achieves interaction between two learning pipelines by mutual learning, so that promotes the recognition of the fine-grained features and strengthens the knowledge generalization ability in zero-shot learning. The proposed SVDML yields significant improvements over the strong baselines, leading to the new state-of the-art performances on three popular challenging benchmarks.
引用
收藏
页码:383 / 395
页数:13
相关论文
共 50 条
  • [31] Contrastive semantic disentanglement in latent space for generalized zero-shot learning
    Fan, Wentao
    Liang, Chen
    Wang, Tian
    KNOWLEDGE-BASED SYSTEMS, 2022, 257
  • [32] Contrastive semantic disentanglement in latent space for generalized zero-shot learning
    Fan, Wentao
    Liang, Chen
    Wang, Tian
    Knowledge-Based Systems, 2022, 257
  • [33] Adversarial Zero-Shot Learning with Semantic Augmentation
    Tong, Bin
    Klinkigt, Martin
    Chen, Junwen
    Cui, Xiankun
    Kong, Quan
    Murakami, Tomokazu
    Kobayashi, Yoshiyuki
    THIRTY-SECOND AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE / THIRTIETH INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE CONFERENCE / EIGHTH AAAI SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2018, : 2476 - 2483
  • [34] Preserving Semantic Relations for Zero-Shot Learning
    Annadani, Yashas
    Biswas, Soma
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 7603 - 7612
  • [35] Semantic softmax loss for zero-shot learning
    Ji, Zhong
    Sun, Yuxin
    Yu, Yunlong
    Guo, Jichang
    Pang, Yanwei
    NEUROCOMPUTING, 2018, 316 : 369 - 375
  • [36] Learning exclusive discriminative semantic information for zero-shot learning
    Mi, Jian-Xun
    Zhang, Zhonghao
    Tai, Debao
    Zhou, Li-Fang
    Jia, Wei
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2023, 14 (03) : 761 - 772
  • [37] Learning a Deep Embedding Model for Zero-Shot Learning
    Zhang, Li
    Xiang, Tao
    Gong, Shaogang
    30TH IEEE CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2017), 2017, : 3010 - 3019
  • [38] Learning exclusive discriminative semantic information for zero-shot learning
    Jian-Xun Mi
    Zhonghao Zhang
    Debao Tai
    Li-Fang Zhou
    Wei Jia
    International Journal of Machine Learning and Cybernetics, 2023, 14 : 761 - 772
  • [39] Learning cross-domain semantic-visual relationships for transductive zero-shot learning
    Lv, Fengmao
    Zhang, Jianyang
    Yang, Guowu
    Feng, Lei
    Yu, Yufeng
    Duan, Lixin
    PATTERN RECOGNITION, 2023, 141
  • [40] Scalable Zero-Shot Learning via Binary Visual-Semantic Embeddings
    Shen, Fumin
    Zhou, Xiang
    Yu, Jun
    Yang, Yang
    Liu, Li
    Shen, Heng Tao
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2019, 28 (07) : 3662 - 3674