论文标题
联合神经建筑搜索
Federated Neural Architecture Search
论文作者
论文摘要
为了保留用户隐私在启用移动智能的同时,已经提出了技术来培训有关分散数据的深度神经网络。但是,对分散数据的培训使神经体系结构的设计非常困难。在设计和部署异构移动平台的不同神经体系结构时,这种困难将进一步扩大。在这项工作中,我们向分散的培训提出了自动的神经体系结构搜索,这是一种新的DNN培训范式,称为联合神经建筑搜索,即Federated Nas。为了应对有限的客户计算和通信资源的主要挑战,我们提出了FedNas,这是一个高度优化的框架,用于有效的联合NAS。 FedNAS充分利用了在建筑搜索过程中重新训练模型候选人不足的关键机会,并结合了三个关键的优化:平行候选人对部分客户进行培训,早期降低候选人的表现较低,并且动态循环数量。在大规模数据集和典型的CNN体系结构上进行了测试,FedNAS具有可比的模型精度作为最先进的NAS算法,该算法训练具有集中数据的模型,并且与Federated Nas的直接设计相比,最多将客户成本降低了两个幅度。
To preserve user privacy while enabling mobile intelligence, techniques have been proposed to train deep neural networks on decentralized data. However, training over decentralized data makes the design of neural architecture quite difficult as it already was. Such difficulty is further amplified when designing and deploying different neural architectures for heterogeneous mobile platforms. In this work, we propose an automatic neural architecture search into the decentralized training, as a new DNN training paradigm called Federated Neural Architecture Search, namely federated NAS. To deal with the primary challenge of limited on-client computational and communication resources, we present FedNAS, a highly optimized framework for efficient federated NAS. FedNAS fully exploits the key opportunity of insufficient model candidate re-training during the architecture search process, and incorporates three key optimizations: parallel candidates training on partial clients, early dropping candidates with inferior performance, and dynamic round numbers. Tested on large-scale datasets and typical CNN architectures, FedNAS achieves comparable model accuracy as state-of-the-art NAS algorithm that trains models with centralized data, and also reduces the client cost by up to two orders of magnitude compared to a straightforward design of federated NAS.