Test-time Adaptation for Machine Translation Evaluation by Uncertainty Minimization

被引:0
|
作者
Zhan, Runzhe [1 ]
Liu, Xuebo [2 ]
Wong, Derek F. [2 ]
Zhang, Cuilian [1 ]
Chao, Lidia S. [1 ]
Zhang, Min
机构
[1] Univ Macau, Dept Comp & Informat Sci, NLP2CT Lab, Taipa, Macau, Peoples R China
[2] Harbin Inst Technol, Inst Comp & Intelligence, Shenzhen, Peoples R China
基金
中国国家自然科学基金;
关键词
D O I
暂无
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
The neural metrics recently received considerable attention from the research community in the automatic evaluation of machine translation. Unlike text-based metrics that have interpretable and consistent evaluation mechanisms for various data sources, the reliability of neural metrics in assessing out-of-distribution data remains a concern due to the disparity between training data and real-world data. This paper aims to address the inference bias of neural metrics through uncertainty minimization during test time, without requiring additional data. Our proposed method comprises three steps: uncertainty estimation, test-time adaptation, and inference. Specifically, the model employs the prediction uncertainty of the current data as a signal to update a small fraction of parameters during test time and subsequently refine the prediction through optimization. To validate our approach, we apply the proposed method to three representative models and conduct experiments on the WMT21 benchmarks. The results obtained from both in-domain and out-of-distribution evaluations consistently demonstrate improvements in correlation performance across different models. Furthermore, we provide evidence that the proposed method effectively reduces model uncertainty. The code is publicly available at https://github.com/NLP2CT/TaU.
引用
收藏
页码:807 / 820
页数:14
相关论文
共 50 条
  • [21] Navigating Continual Test-time Adaptation with Symbiosis Knowledge
    Yang, Xu
    Li, Mogi
    Yin, Jie
    Wei, Kun
    Deng, Cheng
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 5326 - 5334
  • [22] Parameter-free Online Test-time Adaptation
    Boudiaf, Malik
    Mueller, Romain
    Ben Ayed, Ismail
    Bertinetto, Luca
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 8334 - 8343
  • [23] Domain Alignment Meets Fully Test-Time Adaptation
    Thopalli, Kowshik
    Turaga, Pavan
    Thiagarajan, Jayaraman J.
    Proceedings of Machine Learning Research, 2022, 189 : 1006 - 1021
  • [24] Improved Self-Training for Test-Time Adaptation
    Ma, Jing
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2024, : 23701 - 23710
  • [25] Prototypical class-wise test-time adaptation
    Lee, Hojoon
    Lee, Seunghwan
    Jung, Inyoung
    Korea, Sungeun Hong
    PATTERN RECOGNITION LETTERS, 2025, 187 : 49 - 55
  • [26] Efficient Test-Time Model Adaptation without Forgetting
    Niu, Shuaicheng
    Wu, Jiaxiang
    Zhang, Yifan
    Chen, Yaofo
    Zheng, Shijian
    Zhao, Peilin
    Tan, Mingkui
    INTERNATIONAL CONFERENCE ON MACHINE LEARNING, VOL 162, 2022,
  • [27] Unraveling Batch Normalization for Realistic Test-Time Adaptation
    Su, Zixian
    Guo, Jingwei
    Yao, Kai
    Yang, Xi
    Wang, Qiufeng
    Huang, Kaizhu
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 13, 2024, : 15136 - 15144
  • [28] Demographic bias mitigation at test-time using uncertainty estimation and human-machine partnership
    Nair, Anoop Krishnan Upendran
    Rattani, Ajita
    MACHINE LEARNING WITH APPLICATIONS, 2025, 19
  • [29] Test-time Domain Adaptation for Monocular Depth Estimation
    Li, Zhi
    Sh, Shaoshuai
    Schiele, Bernt
    Dai, Dengxin
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 4873 - 4879
  • [30] VPA: Fully Test-Time Visual Prompt Adaptation
    Sun, Jiachen
    Ibrahim, Mark
    Hall, Melissa
    Evtimov, Ivan
    Mao, Z. Morley
    Ferrer, Cristian Canton
    Hazirbas, Caner
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5796 - 5806