An Information-Theoretic Analysis of the Impact of Task Similarity on Meta-Learning

被引:10
|
作者
Jose, Sharu Theresa [1 ]
Simeone, Osvaldo [1 ]
机构
[1] Kings Coll London, Dept Engn, Kings Commun Learning & Informat Proc KCLIP Lab, London, England
基金
欧洲研究理事会;
关键词
D O I
10.1109/ISIT45174.2021.9517767
中图分类号
TP301 [理论、方法];
学科分类号
081202 ;
摘要
Meta-learning aims at optimizing the hyperparameters of a model class or training algorithm from the observation of data from a number of related tasks. Following the setting of Baxter [1], the tasks are assumed to belong to the same task environment, which is defined by a distribution over the space of tasks and by per-task data distributions. The statistical properties of the task environment thus dictate the similarity of the tasks. The goal of the meta-learner is to ensure that the hyperparameters obtain a small loss when applied for training of a new task sampled from the task environment. The difference between the resulting average loss, known as meta-population loss, and the corresponding empirical lass measured on the available data from related tasks, known as meta-generalization gap, is a measure of the generalization capability of the meta-learner. In this paper, we present novel information-theoretic bounds on the average absolute value of the meta-generalization gap. Unlike prior work [2], our bounds explicitly capture the impact of task relatedness, the number of tasks, and the number of data samples per task on the meta-generalization gap. Task similarity is gauged via the Kullback-Leibler (KL) and Jensen-Shannon (JS) divergences. We illustrate the proposed bounds on the example of ridge regression with meta-learned bias.
引用
收藏
页码:1534 / 1539
页数:6
相关论文
共 50 条
  • [1] Information-Theoretic Measures for Meta-learning
    Segrera, Saddys
    Pinho, Joel
    Moreno, Maria N.
    HYBRID ARTIFICIAL INTELLIGENCE SYSTEMS, 2008, 5271 : 458 - 465
  • [2] Generalization Bounds for Meta-Learning: An Information-Theoretic Analysis
    Chen, Qi
    Shui, Changjian
    Marchand, Mario
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 34 (NEURIPS 2021), 2021, 34
  • [3] Information-Theoretic Analysis of Epistemic Uncertainty in Bayesian Meta-learning
    Jose, Sharu Theresa
    Park, Sangwoo
    Simeone, Osvaldo
    INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND STATISTICS, VOL 151, 2022, 151
  • [4] Information-Theoretic Generalization Bounds for Meta-Learning and Applications
    Jose, Sharu Theresa
    Simeone, Osvaldo
    ENTROPY, 2021, 23 (01) : 1 - 28
  • [5] Information-theoretic Task Selection for Meta-Reinforcement Learning
    Gutierrez, Ricardo Luna
    Leonetti, Matteo
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 33, NEURIPS 2020, 2020, 33
  • [6] An Information-Theoretic Approach for Multi-task Learning
    Yang, Pei
    Tan, Qi
    Xu, Hao
    Ding, Yehua
    ADVANCED DATA MINING AND APPLICATIONS, PROCEEDINGS, 2009, 5678 : 386 - 396
  • [7] AN INFORMATION-THEORETIC APPROACH TO TRANSFERABILITY IN TASK TRANSFER LEARNING
    Bao, Yajie
    Li, Yang
    Huang, Shao-Lun
    Zhang, Lin
    Zheng, Lizhong
    Zamir, Amir
    Guibas, Leonidas
    2019 IEEE INTERNATIONAL CONFERENCE ON IMAGE PROCESSING (ICIP), 2019, : 2309 - 2313
  • [8] Incremental learning of task sequences with information-theoretic metricse
    Pardowitz, Michael
    Zoellner, Raoul
    Dillmann, Rudiger
    EUROPEAN ROBOTICS SYMPOSIUM 2006, 2006, 22 : 51 - +
  • [9] Information-theoretic and set-theoretic similarity
    Cazzanti, Luca
    Gupta, Maya R.
    2006 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY, VOLS 1-6, PROCEEDINGS, 2006, : 1836 - +
  • [10] Information-theoretic analysis for transfer learning
    Wu, Xuetong
    Manton, Jonathan H.
    Aickelin, Uwe
    Zhu, Jingge
    2020 IEEE INTERNATIONAL SYMPOSIUM ON INFORMATION THEORY (ISIT), 2020, : 2819 - 2824