论文标题
关于深神经网络压缩的调查:挑战,概述和解决方案
A Survey on Deep Neural Network Compression: Challenges, Overview, and Solutions
论文作者
论文摘要
由于其自动化功能提取能力,深度神经网络(DNN)已获得了前所未有的性能。在过去的十年中,这种高级性能导致DNN模型在不同的物联网(IoT)应用程序中的重大合并。但是,DNN模型的计算,能源和存储的巨大要求使其对资源限制物联网设备的部署持续不错。因此,近年来提出了几种压缩技术,以减少DNN模型的存储和计算要求。这些关于DNN压缩的技术利用了不同的观点来以最小的精度妥协压缩DNN。它鼓励我们对DNN压缩技术进行全面概述。在本文中,我们介绍了有关压缩DNN模型的现有文献的全面综述,该文献均减少了存储和计算要求。我们将现有方法分为五个广泛的类别,即基于压缩DNN模型的机制,即网络修剪,稀疏表示,精度,知识蒸馏和杂项。本文还讨论了与每种DNN压缩技术相关的挑战。最后,我们将在每个类别下的现有工作提供快速摘要,并在DNN压缩中的未来方向。
Deep Neural Network (DNN) has gained unprecedented performance due to its automated feature extraction capability. This high order performance leads to significant incorporation of DNN models in different Internet of Things (IoT) applications in the past decade. However, the colossal requirement of computation, energy, and storage of DNN models make their deployment prohibitive on resource constraint IoT devices. Therefore, several compression techniques were proposed in recent years for reducing the storage and computation requirements of the DNN model. These techniques on DNN compression have utilized a different perspective for compressing DNN with minimal accuracy compromise. It encourages us to make a comprehensive overview of the DNN compression techniques. In this paper, we present a comprehensive review of existing literature on compressing DNN model that reduces both storage and computation requirements. We divide the existing approaches into five broad categories, i.e., network pruning, sparse representation, bits precision, knowledge distillation, and miscellaneous, based upon the mechanism incorporated for compressing the DNN model. The paper also discussed the challenges associated with each category of DNN compression techniques. Finally, we provide a quick summary of existing work under each category with the future direction in DNN compression.