论文标题
培训具有降低量化水平的混合信号光卷积神经网络
Training of mixed-signal optical convolutional neural network with reduced quantization level
论文作者
论文摘要
采用模拟基质型加速器的混合信号人工神经网络(ANN)可以达到更高的速度并提高功率效率。尽管已知模拟计算容易受到噪声和设备缺陷的影响,但由于ANN的稳健性,各种模拟计算范式被认为是解决机器学习应用程序不断增长的计算需求的有希望的解决方案。这种鲁棒性已在低精确的定点ANN模型中探索,这些模型已证明成功地在数字计算机上压缩ANN模型大小。但是,这些有希望的结果和网络培训算法不能轻易迁移到模拟加速器。原因是数字计算机通常具有较高宽度的中间结果,尽管每个ANN层的输入和权重宽度低。虽然模拟中间结果的精度较低,但类似于量化水平降低的数字信号。在这里,我们报告了一种混合信号ANN的训练方法,其模拟信号,随机噪声和确定性误差(扭曲)中具有两种类型的错误。结果表明,用我们提出的方法训练的混合信号ANN可以达到等效分类精度,而噪声水平高达理想量化步长的50%。我们已经在基于衍射光学的混合信号光学卷积神经网络上证明了这种训练方法。
Mixed-signal artificial neural networks (ANNs) that employ analog matrix-multiplication accelerators can achieve higher speed and improved power efficiency. Though analog computing is known to be susceptible to noise and device imperfections, various analog computing paradigms have been considered as promising solutions to address the growing computing demand in machine learning applications, thanks to the robustness of ANNs. This robustness has been explored in low-precision, fixed-point ANN models, which have proven successful on compressing ANN model size on digital computers. However, these promising results and network training algorithms cannot be easily migrated to analog accelerators. The reason is that digital computers typically carry intermediate results with higher bit width, though the inputs and weights of each ANN layers are of low bit width; while the analog intermediate results have low precision, analogous to digital signals with a reduced quantization level. Here we report a training method for mixed-signal ANN with two types of errors in its analog signals, random noise, and deterministic errors (distortions). The results showed that mixed-signal ANNs trained with our proposed method can achieve an equivalent classification accuracy with noise level up to 50% of the ideal quantization step size. We have demonstrated this training method on a mixed-signal optical convolutional neural network based on diffractive optics.