论文标题
丘脑皮层运动电路的见解,以对复杂序列进行更强的层次结构控制
Thalamocortical motor circuit insights for more robust hierarchical control of complex sequences
论文作者
论文摘要
我们研究了复发性神经网络的学习,这些神经网络产生了由可重复使用的“主题”串联组成的时间序列。在神经科学或机器人技术的背景下,这些基序将是产生复杂行为的运动原则。给定一组已知的主题,可以在不影响已知集的性能的情况下学习一个新的主题,然后在新序列中使用,而无需首先明确学习所有可能的过渡?两个要求启用了这一点:(i)学习新主题时的参数更新不会干扰先前获取的参数; (ii)从任何其他主题结束的网络状态开始时,即使该状态在培训期间不存在,也可以强劲地产生一个新的图案。我们通过研究具有特定架构的人工神经网络(ANN)来满足第一个要求,并试图通过训练它们来实现第二个要求,从而从随机初始状态中产生图案。我们发现,学习单基调的学习成功,但序列产生并不强大:观察到过渡失败。然后,我们将这些结果与一个模型进行了比较,该模型的架构和可分析吸引的动力学灵感来自电动机丘脑皮层电路,其中包括用于实现基序过渡的特定模块。可以调整此模型的突触权重,而无需在模拟网络输出上需要随机梯度下降(SGD),并且我们有渐近的确保过渡不会失败。实际上,在模拟中,我们在先前研究的ANN上达到了单位精度,并且没有过渡失败,并改善了测序鲁棒性。最后,我们表明,通过研究该模型的过渡子网获得的见解也可以改善先前研究的传统ANN中过渡的鲁棒性。
We study learning of recurrent neural networks that produce temporal sequences consisting of the concatenation of re-usable "motifs". In the context of neuroscience or robotics, these motifs would be the motor primitives from which complex behavior is generated. Given a known set of motifs, can a new motif be learned without affecting the performance of the known set and then used in new sequences without first explicitly learning every possible transition? Two requirements enable this: (i) parameter updates while learning a new motif do not interfere with the parameters used for the previously acquired ones; and (ii) a new motif can be robustly generated when starting from the network state reached at the end of any of the other motifs, even if that state was not present during training. We meet the first requirement by investigating artificial neural networks (ANNs) with specific architectures, and attempt to meet the second by training them to generate motifs from random initial states. We find that learning of single motifs succeeds but that sequence generation is not robust: transition failures are observed. We then compare these results with a model whose architecture and analytically-tractable dynamics are inspired by the motor thalamocortical circuit, and that includes a specific module used to implement motif transitions. The synaptic weights of this model can be adjusted without requiring stochastic gradient descent (SGD) on the simulated network outputs, and we have asymptotic guarantees that transitions will not fail. Indeed, in simulations, we achieve single-motif accuracy on par with the previously studied ANNs and have improved sequencing robustness with no transition failures. Finally, we show that insights obtained by studying the transition subnetwork of this model can also improve the robustness of transitioning in the traditional ANNs previously studied.