Learning cross-domain semantic-visual relationships for transductive zero-shot learning

被引:14
|
作者
Lv, Fengmao [1 ]
Zhang, Jianyang [2 ]
Yang, Guowu [2 ]
Feng, Lei [3 ]
Yu, Yufeng [4 ]
Duan, Lixin [2 ]
机构
[1] Southwest Jiaotong Univ, West Pk High Tech Zone, Chengdu 611756, Sichuan, Peoples R China
[2] Univ Elect Sci & Technol China, 2006 Xiyuan Ave, Chengdu 611731, Sichuan, Peoples R China
[3] Nanyang Technol Univ, Sch Comp Sci & Engn, Singapore 639798, Singapore
[4] Guangzhou Univ, 230 Wai Huan Xi Rd, Guangzhou 510006, Guangdong, Peoples R China
基金
中国国家自然科学基金;
关键词
Zero -shot learning; Transfer learning; Domain adaptation;
D O I
10.1016/j.patcog.2023.109591
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Zero-Shot Learning (ZSL) learns models for recognizing new classes. One of the main challenges in ZSL is the domain discrepancy caused by the category inconsistency between training and testing data. Domain adaptation is the most intuitive way to address this challenge. However, existing domain adaptation tech-niques cannot be directly applied into ZSL due to the disjoint label space between source and target do-mains. This work proposes the Transferrable Semantic-Visual Relation (TSVR) approach towards transduc-tive ZSL. TSVR redefines image recognition as predicting the similarity/dissimilarity labels for semantic -visual fusions consisting of class attributes and visual features. After the above transformation, the source and target domains can have the same label space, which hence enables to quantify domain discrepancy. For the redefined problem, the number of similar semantic-visual pairs is significantly smaller than that of dissimilar ones. To this end, we further propose to use Domain-Specific Batch Normalization to align the domain discrepancy.(c) 2023 Elsevier Ltd. All rights reserved.
引用
下载
收藏
页数:9
相关论文
共 50 条
  • [1] Cross-domain mapping learning for transductive zero-shot learning
    Ding, Mingyu
    Wang, Zhe
    Lu, Zhiwu
    COMPUTER VISION AND IMAGE UNDERSTANDING, 2019, 187
  • [2] Semantic-visual shared knowledge graph for zero-shot learning
    Yu, Beibei
    Xie, Cheng
    Tang, Peng
    Li, Bin
    PEERJ COMPUTER SCIENCE, 2023, 9
  • [3] Semantic-Visual Combination Propagation Network for Zero-Shot Learning
    Song, Wenli
    Zhang, Lei
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS II-EXPRESS BRIEFS, 2022, 69 (04) : 2341 - 2345
  • [4] HSVA: Hierarchical Semantic-Visual Adaptation for Zero-Shot Learning
    Chen, Shiming
    Xie, Guo-Sen
    Liu, Yang
    Peng, Qinmu
    Sun, Baigui
    Li, Hao
    You, Xinge
    Shao, Ling
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [5] Semantic-visual shared knowledge graph for zero-shot learning
    Yu B.
    Xie C.
    Tang P.
    Li B.
    PeerJ Computer Science, 2023, 9
  • [6] Transductive Visual-Semantic Embedding for Zero-shot Learning
    Xu, Xing
    Shen, Fumin
    Yang, Yang
    Shao, Jie
    Huang, Zi
    PROCEEDINGS OF THE 2017 ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA RETRIEVAL (ICMR'17), 2017, : 41 - 49
  • [7] Progressive Semantic-Visual Mutual Adaption for Generalized Zero-Shot Learning
    Liu, Man
    Li, Feng
    Zhang, Chunjie
    Wei, Yunchao
    Bai, Huihui
    Zhao, Yao
    2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 15337 - 15346
  • [8] Cross-Domain Adversarial Learning for Zero-Shot Classification
    Liu H.
    Zheng Q.
    Luo M.
    Zhao H.
    Xiao Y.
    Lü Y.
    Jisuanji Yanjiu yu Fazhan/Computer Research and Development, 2019, 56 (12): : 2521 - 2535
  • [9] Transductive Zero-Shot Learning with Visual Structure Constraint
    Wan, Ziyu
    Chen, Dongdong
    Li, Yan
    Yan, Xingguang
    Zhang, Junge
    Yu, Yizhou
    Liao, Jing
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 32 (NIPS 2019), 2019, 32
  • [10] Transductive semantic knowledge graph propagation for zero-shot learning
    Zhang, Hai-gang
    Que, Hao-yi
    Ren, Jin
    Wu, Zheng-guang
    JOURNAL OF THE FRANKLIN INSTITUTE-ENGINEERING AND APPLIED MATHEMATICS, 2023, 360 (17): : 13108 - 13125