论文标题

辍学作为持续学习的隐性门控机制

Dropout as an Implicit Gating Mechanism For Continual Learning

论文作者

Mirzadeh, Seyed-Iman, Farajtabar, Mehrdad, Ghasemzadeh, Hassan

论文摘要

近年来,神经网络表现出了在各个领域完成复杂学习任务的出色能力。但是,当他们面对一系列学习任务时,他们遭受了“灾难性遗忘”问题的困扰,在学习新任务时,他们忘记了旧的任务。这个问题也与“稳定性困境”高度相关。网络的塑料越多,就越容易学习新任务,但是它越来越忘记了以前的任务。相反,稳定的网络无法像非常塑料网络一样快地学习新任务。但是,保留从先前任务中学到的知识更为可靠。已经提出了几种解决方案,以通过使神经网络参数更加稳定来克服遗忘问题,其中一些人提到了辍学在持续学习中的意义。但是,他们的关系尚未得到充分研究。在本文中,我们调查了这种关系,并表明一个带有辍学的网络学习了一种门控机制,以便对于不同的任务,网络的不同路径处于活动状态。我们的实验表明,这种隐式门控实现的稳定性在导致与其他涉及的持续学习算法相当或更好的绩效方面起着非常关键的作用,以克服灾难性的遗忘。

In recent years, neural networks have demonstrated an outstanding ability to achieve complex learning tasks across various domains. However, they suffer from the "catastrophic forgetting" problem when they face a sequence of learning tasks, where they forget the old ones as they learn new tasks. This problem is also highly related to the "stability-plasticity dilemma". The more plastic the network, the easier it can learn new tasks, but the faster it also forgets previous ones. Conversely, a stable network cannot learn new tasks as fast as a very plastic network. However, it is more reliable to preserve the knowledge it has learned from the previous tasks. Several solutions have been proposed to overcome the forgetting problem by making the neural network parameters more stable, and some of them have mentioned the significance of dropout in continual learning. However, their relationship has not been sufficiently studied yet. In this paper, we investigate this relationship and show that a stable network with dropout learns a gating mechanism such that for different tasks, different paths of the network are active. Our experiments show that the stability achieved by this implicit gating plays a very critical role in leading to performance comparable to or better than other involved continual learning algorithms to overcome catastrophic forgetting.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源