论文标题
通过正交重量修饰的生成功能重播以持续学习
Generative Feature Replay with Orthogonal Weight Modification for Continual Learning
论文作者
论文摘要
智能代理人依次学习和记住多个任务的能力对于实现人工通用智能至关重要。已经提出了许多持续学习(CL)方法来克服灾难性的遗忘,这是由于神经网络的顺序学习中的非i.i.d数据所致。在本文中,我们专注于班级学习,这是一个具有挑战性的CL场景。在这种情况下,生成重播是一种有前途的策略,它为以前的任务生成和重播伪数据以减轻灾难性的遗忘。但是,很难连续训练生成模型以获得相对复杂的数据。基于最近提出的正交重量修改(OWM)算法,该算法在学习新任务时几乎可以保持先前学习的功能不变性,我们建议使用生成模型重放倒数第二层功能; 2)利用自我监督的辅助任务进一步提高功能的稳定性。几个数据集的经验结果表明,我们的方法始终对强大的OWM取得了重大改进,而常规生成重播总是会产生负面影响。同时,我们的方法击败了几个强大的基线,其中包括基于真实数据存储的基准。此外,我们进行实验以研究为什么我们的方法有效。
The ability of intelligent agents to learn and remember multiple tasks sequentially is crucial to achieving artificial general intelligence. Many continual learning (CL) methods have been proposed to overcome catastrophic forgetting which results from non i.i.d data in the sequential learning of neural networks. In this paper we focus on class incremental learning, a challenging CL scenario. For this scenario, generative replay is a promising strategy which generates and replays pseudo data for previous tasks to alleviate catastrophic forgetting. However, it is hard to train a generative model continually for relatively complex data. Based on recently proposed orthogonal weight modification (OWM) algorithm which can approximately keep previously learned feature invariant when learning new tasks, we propose to 1) replay penultimate layer feature with a generative model; 2) leverage a self-supervised auxiliary task to further enhance the stability of feature. Empirical results on several datasets show our method always achieves substantial improvement over powerful OWM while conventional generative replay always results in a negative effect. Meanwhile our method beats several strong baselines including one based on real data storage. In addition, we conduct experiments to study why our method is effective.