论文标题
利用事后看来锚定过去的知识持续学习
Using Hindsight to Anchor Past Knowledge in Continual Learning
论文作者
论文摘要
在持续学习中,学习者面对一系列数据流,其分布会随着时间而变化。众所周知,现代神经网络在这种情况下会受到影响,因为他们很快忘记了以前获得的知识。为了解决这种灾难性遗忘,许多持续学习的方法实现了不同类型的经验重播,重新学习了过去的数据中存储在一个称为情节记忆的小型缓冲区中。在这项工作中,我们以一个新的目标来补充经验重播,我们称之为锚定,学习者使用双重优化来更新其对当前任务的知识,同时保持对过去任务的某些锚点的预测。这些锚点是使用基于梯度的优化来学习的,以最大化遗忘,这是通过对过去任务的情节记忆进行微调模型来近似的。对几个有监督的学习基准进行的实验持续学习表明,我们的方法改善了准确性和忘记指标以及各种剧集记忆的标准经验重播。
In continual learning, the learner faces a stream of data whose distribution changes over time. Modern neural networks are known to suffer under this setting, as they quickly forget previously acquired knowledge. To address such catastrophic forgetting, many continual learning methods implement different types of experience replay, re-learning on past data stored in a small buffer known as episodic memory. In this work, we complement experience replay with a new objective that we call anchoring, where the learner uses bilevel optimization to update its knowledge on the current task, while keeping intact the predictions on some anchor points of past tasks. These anchor points are learned using gradient-based optimization to maximize forgetting, which is approximated by fine-tuning the currently trained model on the episodic memory of past tasks. Experiments on several supervised learning benchmarks for continual learning demonstrate that our approach improves the standard experience replay in terms of both accuracy and forgetting metrics and for various sizes of episodic memories.