论文标题
快速进化神经架构搜索的采样培训和节点继承
Sampled Training and Node Inheritance for Fast Evolutionary Neural Architecture Search
论文作者
论文摘要
深度神经网络的性能在很大程度上取决于其体系结构,并且为自动化网络体系结构设计开发了各种神经体系结构搜索策略。最近,由于进化算法具有吸引力的全球优化能力,进化神经结构搜索(ENAS)受到了越来越多的关注。但是,ENA遭受了极高的计算成本,因为进化优化中通常需要大量的性能评估,而训练深度神经网络本身在计算上非常密集。为了解决这个问题,本文提出了一个基于定向无环图的快速enas的新进化框架,其中在每个小型培训数据上对父母进行随机采样和培训。此外,采用节点继承策略来产生后代个人,并且在没有培训的情况下直接评估其健身性。为了增强进化神经网络的特征处理能力,我们还编码了搜索空间中的通道注意机制。与26种最先进的同伴算法相比,我们评估了广泛使用的数据集上提出的算法。我们的实验结果表明,所提出的算法在计算上不仅要高效,而且在学习绩效方面也具有很高的竞争力。
The performance of a deep neural network is heavily dependent on its architecture and various neural architecture search strategies have been developed for automated network architecture design. Recently, evolutionary neural architecture search (ENAS) has received increasing attention due to the attractive global optimization capability of evolutionary algorithms. However, ENAS suffers from extremely high computation costs because a large number of performance evaluations is usually required in evolutionary optimization and training deep neural networks is itself computationally very intensive. To address this issue, this paper proposes a new evolutionary framework for fast ENAS based on directed acyclic graph, in which parents are randomly sampled and trained on each mini-batch of training data. In addition, a node inheritance strategy is adopted to generate offspring individuals and their fitness is directly evaluated without training. To enhance the feature processing capability of the evolved neural networks, we also encode a channel attention mechanism in the search space. We evaluate the proposed algorithm on the widely used datasets, in comparison with 26 state-of-the-art peer algorithms. Our experimental results show the proposed algorithm is not only computationally much more efficiently, but also highly competitive in learning performance.