论文标题
定义不断学习的基准
Defining Benchmarks for Continual Few-Shot Learning
论文作者
论文摘要
由于推出了适当的基准,在过去几年中,几乎没有射击和持续学习都取得了长足的进步。话虽这么说,该领域仍必须为一套基准构建一套基准,以使高度理想地设置不断的几次学习,其中学习者被呈现了许多少数射击任务,一个又一个又一个又一个,然后要求在所有先前看到的任务的验证集上表现良好。持续的很少的学习学习具有少量的计算足迹,因此是有效研究和实验的绝佳环境。在本文中,我们首先定义了一个理论框架,用于考虑最新文献,然后提出一系列灵活的基准测试,以统一评估标准并允许从多个角度探索问题。作为基准的一部分,我们引入了一个称为Slimagenet64的紧凑型变体,该变体保留了所有原始的1000类,但仅包含每个一个(总计200k数据点)的200个实例,降低到64 x 64个像素。结果,我们使用许多流行的少数学习算法为提议的基准提供了基准,从而在连续和数据限制的设置中揭示了这些算法的先前未知优势和劣势。
Both few-shot and continual learning have seen substantial progress in the last years due to the introduction of proper benchmarks. That being said, the field has still to frame a suite of benchmarks for the highly desirable setting of continual few-shot learning, where the learner is presented a number of few-shot tasks, one after the other, and then asked to perform well on a validation set stemming from all previously seen tasks. Continual few-shot learning has a small computational footprint and is thus an excellent setting for efficient investigation and experimentation. In this paper we first define a theoretical framework for continual few-shot learning, taking into account recent literature, then we propose a range of flexible benchmarks that unify the evaluation criteria and allows exploring the problem from multiple perspectives. As part of the benchmark, we introduce a compact variant of ImageNet, called SlimageNet64, which retains all original 1000 classes but only contains 200 instances of each one (a total of 200K data-points) downscaled to 64 x 64 pixels. We provide baselines for the proposed benchmarks using a number of popular few-shot learning algorithms, as a result, exposing previously unknown strengths and weaknesses of those algorithms in continual and data-limited settings.