Open-set domain adaptation with visual-language foundation models

被引:0
|
作者
Yu, Qing [1 ]
Irie, Go [2 ]
Aizawa, Kiyoharu [1 ]
机构
[1] Univ Tokyo, Dept Informat & Commun Engn, Tokyo 1138656, Japan
[2] Tokyo Univ Sci, Dept Informat & Comp Technol, Tokyo 1258585, Japan
关键词
Deep learning; Cross-domain learning; Open-set recognition; Domain adaptation;
D O I
10.1016/j.cviu.2024.104230
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Unsupervised domain adaptation (UDA) has proven to be very effective in transferring knowledge obtained from a source domain with labeled data to a target domain with unlabeled data. Owing to the lack of labeled data in the target domain and the possible presence of unknown classes, open-set domain adaptation (ODA) has emerged as a potential solution to identify these classes during the training phase. Although existing ODA approaches aim to solve the distribution shifts between the source and target domains, most methods fine-tuned ImageNet pre-trained models on the source domain with the adaptation on the target domain. Recent visual- language foundation models (VLFM), such as Contrastive Language-Image Pre-Training (CLIP), are robust to many distribution shifts and, therefore, should substantially improve the performance of ODA. In this work, we explore generic ways to adopt CLIP, a popular VLFM, for ODA. We investigate the performance of zero-shot prediction using CLIP, and then propose an entropy optimization strategy to assist the ODA models with the outputs of CLIP. The proposed approach achieves state-of-the-art results on various benchmarks, demonstrating its effectiveness in addressing the ODA problem.
引用
收藏
页数:8
相关论文
共 50 条
  • [31] Fine-Grained Open-Set Deepfake Detection via Unsupervised Domain Adaptation
    Zhou, Xinye
    Han, Hu
    Shan, Shiguang
    Chen, Xilin
    IEEE TRANSACTIONS ON INFORMATION FORENSICS AND SECURITY, 2024, 19 : 7536 - 7547
  • [32] Self-Labeling Framework for Open-Set Domain Adaptation With Few Labeled Samples
    Yu, Qing
    Irie, Go
    Aizawa, Kiyoharu
    IEEE TRANSACTIONS ON MULTIMEDIA, 2024, 26 : 1474 - 1487
  • [33] Maximum open-set entropy optimization via uncertainty measure for universal domain adaptation
    Ai, Weiwei
    Yang, Zhao
    Chen, Zhiyong
    Hu, Xiao
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 101
  • [34] Deep Open-Set Segmentation in Visual Learning
    Nunes, Ian M.
    Poggi, Marcus
    Oliveira, Hugo
    Pereira, Matheus B.
    dos Santos, Jefersson A.
    2022 35TH SIBGRAPI CONFERENCE ON GRAPHICS, PATTERNS AND IMAGES (SIBGRAPI 2022), 2022, : 314 - 319
  • [35] Progressively Select and Reject Pseudo-labelled Samples for Open-Set Domain Adaptation
    Wang Q.
    Meng F.
    Breckon T.P.
    IEEE Transactions on Artificial Intelligence, 2024, 5 (09): : 1 - 12
  • [36] Reserve to Adapt: Mining Inter-Class Relations for Open-Set Domain Adaptation
    Tong, Yujun
    Chang, Dongliang
    Li, Da
    Wang, Xinran
    Liang, Kongming
    He, Zhongjiang
    Song, Yi-Zhe
    Ma, Zhanyu
    IEEE TRANSACTIONS ON IMAGE PROCESSING, 2025, 34 : 1382 - 1397
  • [37] Open-set domain adaptation for scene classification using multi-adversarial learning
    Zheng, Juepeng
    Wen, Yibin
    Chen, Mengxuan
    Yuan, Shuai
    Li, Weijia
    Zhao, Yi
    Wu, Wenzhao
    Zhang, Lixian
    Dong, Runmin
    Fu, Haohuan
    ISPRS JOURNAL OF PHOTOGRAMMETRY AND REMOTE SENSING, 2024, 208 (245-260) : 245 - 260
  • [38] A fault diagnosis method for rolling bearings in open-set domain adaptation with adversarial learning
    Lei, Tongfei
    Pan, Feng
    Hu, Jiabei
    He, Xu
    Li, Bing
    SCIENTIFIC REPORTS, 2025, 15 (01):
  • [39] Towards Multimodal Open-Set Domain Generalization and Adaptation Through Self-supervision
    Dong, Hao
    Chatziliadou, Eleni
    Fink, Olga
    COMPUTER VISION - ECCV 2024, PT LIX, 2025, 15117 : 270 - 287
  • [40] Prompting Visual-Language Models for Efficient Video Understanding
    Ju, Chen
    Han, Tengda
    Zheng, Kunhao
    Zhang, Ya
    Xie, Weidi
    COMPUTER VISION - ECCV 2022, PT XXXV, 2022, 13695 : 105 - 124