论文标题

SATA:刺激神经网络的稀疏感训练加速器

SATA: Sparsity-Aware Training Accelerator for Spiking Neural Networks

论文作者

Yin, Ruokai, Moitra, Abhishek, Bhattacharjee, Abhiroop, Kim, Youngeun, Panda, Priyadarshini

论文摘要

尖峰神经网络(SNN)由于其固有的高表象激活而引起了传统人工神经网络(ANN)的势能效率替代品。最近,与其他SNN训练算法相比,通过时间反向传播(BPTT)的SNN在图像识别任务上取得了更高的准确性结果。尽管从算法的角度取得了成功,但由于缺乏此SNN培训算法的硬件评估平台,因此先前的作品忽略了BPTT硬件能源开销的评估。此外,尽管SNN长期以来一直被视为能源有效的ANN,但缺少SNN和ANN的训练成本之间的定量比较。为了解决上述问题,在这项工作中,我们介绍了基于BPTT的SNNS培训加速器SATA(Sparsity-Awaine培训加速器)。拟议的SATA提供了一种简单且可重新配置的基于收缩期的加速器架构,这使得可以轻松分析基于BPTT的SNN培训算法的训练能量。通过利用稀疏性,SATA将其计算能效率提高了$ 5.58 \ times $ $,而无需使用稀疏性。根据SATA,我们显示了SNN培训能源效率的定量分析,并比较了SNN和ANN的培训成本。结果表明,与ANN相比,SNNS在基于Eyeriss的基于收缩期的架构上,消耗$ 1.27 \ times $多于稀有的能量。我们发现,这种高训练能源成本来自反向传播期间的时间重复卷积操作和数据运动。此外,为了推动未来的SNN培训算法设计,我们为不同的SNN特异性培训参数提供了几种有关能效的观察结果,并提出了SNN培训的能量估计框架。我们的框架代码可公开可用。

Spiking Neural Networks (SNNs) have gained huge attention as a potential energy-efficient alternative to conventional Artificial Neural Networks (ANNs) due to their inherent high-sparsity activation. Recently, SNNs with backpropagation through time (BPTT) have achieved a higher accuracy result on image recognition tasks than other SNN training algorithms. Despite the success from the algorithm perspective, prior works neglect the evaluation of the hardware energy overheads of BPTT due to the lack of a hardware evaluation platform for this SNN training algorithm. Moreover, although SNNs have long been seen as an energy-efficient counterpart of ANNs, a quantitative comparison between the training cost of SNNs and ANNs is missing. To address the aforementioned issues, in this work, we introduce SATA (Sparsity-Aware Training Accelerator), a BPTT-based training accelerator for SNNs. The proposed SATA provides a simple and re-configurable systolic-based accelerator architecture, which makes it easy to analyze the training energy for BPTT-based SNN training algorithms. By utilizing the sparsity, SATA increases its computation energy efficiency by $5.58 \times$ compared to the one without using sparsity. Based on SATA, we show quantitative analyses of the energy efficiency of SNN training and compare the training cost of SNNs and ANNs. The results show that, on Eyeriss-like systolic-based architecture, SNNs consume $1.27\times$ more total energy with sparsities when compared to ANNs. We find that such high training energy cost is from time-repetitive convolution operations and data movements during backpropagation. Moreover, to propel the future SNN training algorithm design, we provide several observations on energy efficiency for different SNN-specific training parameters and propose an energy estimation framework for SNN training. Code for our framework is made publicly available.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源