论文标题
基于各种结构和自适应建议的多目标神经体系结构搜索
Multi-Objective Neural Architecture Search Based on Diverse Structures and Adaptive Recommendation
论文作者
论文摘要
卷积神经网络(CNN)的神经体系结构搜索(NAS)的搜索空间很大。为了降低搜索成本,大多数NAS算法都使用固定的外网级结构,并仅搜索可重复的单元结构。当使用足够的单元和通道时,这种固定架构的性能很好。但是,当体系结构变得更轻巧时,性能会大大降低。为了获得更好的轻质体系结构,需求更灵活,更多样化的神经体系结构,应为更大的搜索空间设计更有效的方法。在此激励的情况下,我们提出了Moarr算法,该算法利用现有的研究结果和历史信息来快速找到既轻量级又准确的体系结构。我们使用发现的高性能小区来构建网络体系结构。该方法增加了网络体系结构的多样性,同时还减少了单元结构设计的搜索空间。此外,我们设计了一种新型的多目标方法来有效地分析历史评估信息,以便有效地搜索具有高精度和小参数编号的帕累托最佳体系结构。实验结果表明,我们的MOARR可以在6 GPU小时内实现CIFAR-10上的强大且轻巧的模型(具有1.9%的错误率和230万参数),这比最先进的方法要好。探索的体系结构可传输到Imagenet,并以490万参数达到76.0%的TOP-1精度。
The search space of neural architecture search (NAS) for convolutional neural network (CNN) is huge. To reduce searching cost, most NAS algorithms use fixed outer network level structure, and search the repeatable cell structure only. Such kind of fixed architecture performs well when enough cells and channels are used. However, when the architecture becomes more lightweight, the performance decreases significantly. To obtain better lightweight architectures, more flexible and diversified neural architectures are in demand, and more efficient methods should be designed for larger search space. Motivated by this, we propose MoARR algorithm, which utilizes the existing research results and historical information to quickly find architectures that are both lightweight and accurate. We use the discovered high-performance cells to construct network architectures. This method increases the network architecture diversity while also reduces the search space of cell structure design. In addition, we designs a novel multi-objective method to effectively analyze the historical evaluation information, so as to efficiently search for the Pareto optimal architectures with high accuracy and small parameter number. Experimental results show that our MoARR can achieve a powerful and lightweight model (with 1.9% error rate and 2.3M parameters) on CIFAR-10 in 6 GPU hours, which is better than the state-of-the-arts. The explored architecture is transferable to ImageNet and achieves 76.0% top-1 accuracy with 4.9M parameters.