论文标题
实时学习的实用稀疏近似
A Practical Sparse Approximation for Real Time Recurrent Learning
论文作者
论文摘要
当前的培训方法复发性神经网络的方法基于时间的反向传播,这需要存储网络状态的完整历史记录,并禁止更新权重“在线”(在每个时间段之后)。实时经常性学习(RTRL)消除了对历史记录存储的需求,并允许在线重量更新,但以牺牲状态大小四重奏的计算成本为代价。这使除最小网络以外的所有网络(甚至是高度稀疏的网络)都使RTRL培训难以理解。 我们将稀疏的n步近似值(SNAP)引入RTRL影响矩阵,该矩阵仅将非零的条目保持在复发核的N步骤之内。 n = 1的快照并不比返回传播昂贵,我们发现它的表现大大优于其他RTRL近似值,而诸如无偏见的在线复发优化之类的相当成本。对于高度稀疏的网络,n = 2的快照仍然可拖动,并且在在线更新时,在学习速度方面可以超越时间反向传播。当n很大时,快照就等于rtrl。
Current methods for training recurrent neural networks are based on backpropagation through time, which requires storing a complete history of network states, and prohibits updating the weights `online' (after every timestep). Real Time Recurrent Learning (RTRL) eliminates the need for history storage and allows for online weight updates, but does so at the expense of computational costs that are quartic in the state size. This renders RTRL training intractable for all but the smallest networks, even ones that are made highly sparse. We introduce the Sparse n-step Approximation (SnAp) to the RTRL influence matrix, which only keeps entries that are nonzero within n steps of the recurrent core. SnAp with n=1 is no more expensive than backpropagation, and we find that it substantially outperforms other RTRL approximations with comparable costs such as Unbiased Online Recurrent Optimization. For highly sparse networks, SnAp with n=2 remains tractable and can outperform backpropagation through time in terms of learning speed when updates are done online. SnAp becomes equivalent to RTRL when n is large.