Test-Time Domain Adaptation by Learning Domain-Aware Batch Normalization

被引:0
|
作者
Wu, Yanan [1 ,2 ]
Chi, Zhixiang [3 ]
Wang, Yang [4 ]
Plataniotis, Konstantinos N. [3 ]
Feng, Songhe [1 ,2 ]
机构
[1] Beijing Jiaotong Univ, Minist Educ, Key Lab Big Data & Artificial Intelligence Transp, Beijing 100044, Peoples R China
[2] Beijing Jiaotong Univ, Sch Comp & Informat Technol, Beijing 100044, Peoples R China
[3] Univ Toronto, Edward S Rogers Sr ECE Dept, Toronto, ON M5S 3G8, Canada
[4] Concordia Univ, Dept Comp Sci & Software Engn, Montreal, PQ H3G 2J1, Canada
基金
加拿大自然科学与工程研究理事会; 北京市自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Test-time domain adaptation aims to adapt the model trained on source domains to unseen target domains using a few unlabeled images. Emerging research has shown that the label and domain information is separately embedded in the weight matrix and batch normalization (BN) layer. Previous works normally update the whole network naively without explicitly decoupling the knowledge between label and domain. As a result, it leads to knowledge interference and defective distribution adaptation. In this work, we propose to reduce such learning interference and elevate the domain knowledge learning by only manipulating the BN layer. However, the normalization step in BN is intrinsically unstable when the statistics are re-estimated from a few samples. We find that ambiguities can be greatly reduced when only updating the two affine parameters in BN while keeping the source domain statistics. To further enhance the domain knowledge extraction from unlabeled data, we construct an auxiliary branch with label-independent self-supervised learning (SSL) to provide supervision. Moreover, we propose a bi-level optimization based on meta-learning to enforce the alignment of two learning objectives of auxiliary and main branches. The goal is to use the auxiliary branch to adapt the domain and benefit main task for subsequent inference. Our method keeps the same computational cost at inference as the auxiliary branch can be thoroughly discarded after adaptation. Extensive experiments show that our method outperforms the prior works on five WILDS real-world domain shift datasets. Our method can also be integrated with methods with label-dependent optimization to further push the performance boundary. Our code is available at https://github.com/ynanwu/MABN.
引用
收藏
页码:15961 / 15969
页数:9
相关论文
共 50 条
  • [1] Unraveling Batch Normalization for Realistic Test-Time Adaptation
    Su, Zixian
    Guo, Jingwei
    Yao, Kai
    Yang, Xi
    Wang, Qiufeng
    Huang, Kaizhu
    [J]. THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 15136 - 15144
  • [2] Continual Test-Time Domain Adaptation
    Wang, Qin
    Fink, Olga
    Van Gool, Luc
    Dai, Dengxin
    [J]. 2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 7191 - 7201
  • [3] Improved Test-Time Adaptation for Domain Generalization
    Chen, Liang
    Zhang, Yong
    Song, Yibing
    Shan, Ying
    Liu, Lingqiao
    [J]. 2023 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2023, : 24172 - 24182
  • [4] Domain Alignment Meets Fully Test-Time Adaptation
    Thopalli, Kowshik
    Turaga, Pavan
    Thiagarajan, Jayaraman J.
    [J]. Proceedings of Machine Learning Research, 2022, 189 : 1006 - 1021
  • [5] Test-time Domain Adaptation for Monocular Depth Estimation
    Li, Zhi
    Sh, Shaoshuai
    Schiele, Bernt
    Dai, Dengxin
    [J]. 2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 4873 - 4879
  • [6] Feature Stylization and Domain-aware Contrastive Learning for Domain Generalization
    Jeon, Seogkyu
    Hong, Kibeom
    Lee, Pilhyeon
    Lee, Jewook
    Byun, Hyeran
    [J]. PROCEEDINGS OF THE 29TH ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2021, 2021, : 22 - 31
  • [7] Multiple Teacher Model for Continual Test-Time Domain Adaptation
    Wang, Ran
    Zuo, Hua
    Fang, Zhen
    Lu, Jie
    [J]. ADVANCES IN ARTIFICIAL INTELLIGENCE, AI 2023, PT I, 2024, 14471 : 304 - 314
  • [8] Exploring Safety Supervision for Continual Test-time Domain Adaptation
    Yang, Xu
    Gu, Yanan
    Wei, Kun
    Deng, Cheng
    [J]. PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 1649 - 1657
  • [9] Noise-Robust Continual Test-Time Domain Adaptation
    Yu, Zhiqi
    Li, Jingjing
    Du, Zhekai
    Li, Fengling
    Zhu, Lei
    Yang, Yang
    [J]. PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 2654 - 2662
  • [10] Domain-Specific Batch Normalization for Unsupervised Domain Adaptation
    Chang, Woong-Gi
    You, Tackgeun
    Seo, Seonguk
    Kwak, Suha
    Han, Bohyung
    [J]. 2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 7346 - 7354