论文标题
合作与竞争:随着进化多代理增强学习的蜂拥而至
Cooperation and Competition: Flocking with Evolutionary Multi-Agent Reinforcement Learning
论文作者
论文摘要
在多代理系统中,蜂拥而至是一个非常具有挑战性的问题。传统的羊群方法还需要对环境的完整了解和控制模型。在本文中,我们建议在羊群任务中进化多代理增强学习(EMARL),这是一种混合算法,将合作和竞争与很少的先验知识相结合。至于合作,我们根据BOIDS模型设计了代理商对羊群任务的奖励。在竞争中,具有高健身的代理商是作为高级代理商设计的,而健身较低的人则被设计为初中,让初级代理商随时继承了高级代理人的参数。为了加强竞争,我们还设计了一种进化选择机制,该机制在羊群任务中显示出对信用分配的有效性。一系列具有挑战性和自我对比的基准测试的实验结果表明,EMARL明显优于完整的竞争或合作方法。
Flocking is a very challenging problem in a multi-agent system; traditional flocking methods also require complete knowledge of the environment and a precise model for control. In this paper, we propose Evolutionary Multi-Agent Reinforcement Learning (EMARL) in flocking tasks, a hybrid algorithm that combines cooperation and competition with little prior knowledge. As for cooperation, we design the agents' reward for flocking tasks according to the boids model. While for competition, agents with high fitness are designed as senior agents, and those with low fitness are designed as junior, letting junior agents inherit the parameters of senior agents stochastically. To intensify competition, we also design an evolutionary selection mechanism that shows effectiveness on credit assignment in flocking tasks. Experimental results in a range of challenging and self-contrast benchmarks demonstrate that EMARL significantly outperforms the full competition or cooperation methods.