论文标题

转移元学习:信息理论界限和信息元风险最小化

Transfer Meta-Learning: Information-Theoretic Bounds and Information Meta-Risk Minimization

论文作者

Jose, Sharu Theresa, Simeone, Osvaldo, Durisi, Giuseppe

论文摘要

元学习可以通过观察来自许多相关任务的数据来自动渗透归纳偏差。电感偏差是由确定模型类别或训练算法(例如初始化或学习率)的方面的超参数编码的。元学习假设学习任务属于任务环境,并且在元训练和元测试过程中,任务是从相同的任务环境中汲取的。但是,这在实践中可能不正确。在本文中,我们介绍了转移元学习的问题,其中在元测试过程中,从目标任务环境中汲取了任务,这可能与元训练期间观察到的源任务环境有所不同。新的信息理论上限是在转移元化差距上获得的,该差距衡量元训练损失之间的差异,在元学习者中获得,以及来自目标任务环境中新的,随机选择的任务的元测试数据的平均损失。在平均转移元化差距上,第一界通过源和目标数据分布之间的KL差异捕获了源和目标任务环境之间的元环境变化。第二个,Pac-bayesian Bound和第三个单draw绑定,通过源和目标任务分布之间的对数可能性比率来解释这一转变。此外,还引入了两个转移元学习溶液。对于第一个,称为经验元风险最小化(EMRM),我们在平均最优差距上得出界限。第二个,称为信息元风险最小化(IMRM),是通过最小化pac-bayesian结合而获得的。 IMRM通过实验显示了潜在胜过EMRM。

Meta-learning automatically infers an inductive bias by observing data from a number of related tasks. The inductive bias is encoded by hyperparameters that determine aspects of the model class or training algorithm, such as initialization or learning rate. Meta-learning assumes that the learning tasks belong to a task environment, and that tasks are drawn from the same task environment both during meta-training and meta-testing. This, however, may not hold true in practice. In this paper, we introduce the problem of transfer meta-learning, in which tasks are drawn from a target task environment during meta-testing that may differ from the source task environment observed during meta-training. Novel information-theoretic upper bounds are obtained on the transfer meta-generalization gap, which measures the difference between the meta-training loss, available at the meta-learner, and the average loss on meta-test data from a new, randomly selected, task in the target task environment. The first bound, on the average transfer meta-generalization gap, captures the meta-environment shift between source and target task environments via the KL divergence between source and target data distributions. The second, PAC-Bayesian bound, and the third, single-draw bound, account for this shift via the log-likelihood ratio between source and target task distributions. Furthermore, two transfer meta-learning solutions are introduced. For the first, termed Empirical Meta-Risk Minimization (EMRM), we derive bounds on the average optimality gap. The second, referred to as Information Meta-Risk Minimization (IMRM), is obtained by minimizing the PAC-Bayesian bound. IMRM is shown via experiments to potentially outperform EMRM.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源