论文标题

使用对抗网络中的多任务加强学习中的探索

Boosting Exploration in Multi-Task Reinforcement Learning using Adversarial Networks

论文作者

Kumar, Ramnath, Deleu, Tristan, Bengio, Yoshua

论文摘要

近年来,增强学习的进步(RL)非常出色。但是,传统培训方法的局限性变得越来越明显,尤其是在代理商面临新的,看不见的任务的元RL环境中。常规的培训方法容易在这种情况下失败,因为它们需要更多的逆境。我们提出的用于多任务加强学习(MT-RL)的对抗性培训制度解决了RL中常规培训方法的局限性,尤其是在代理商面临新任务的元RL环境中。对抗性组成部分挑战了代理商,迫使其在动态和不可预测的情况下提高其决策能力。该组件在不依赖手动干预或特定领域的知识的情况下运行,这使其成为一种高度的解决方案。在多个MT-RL环境中进行的实验表明,对抗训练会导致更好的探索和对环境的更深入的了解。 MT-RL的对抗性培训制度为RL代理人提供了有关培训和开发的新观点,对该领域是有价值的贡献。

Advancements in reinforcement learning (RL) have been remarkable in recent years. However, the limitations of traditional training methods have become increasingly evident, particularly in meta-RL settings where agents face new, unseen tasks. Conventional training approaches are susceptible to failure in such situations as they need more robustness to adversity. Our proposed adversarial training regime for Multi-Task Reinforcement Learning (MT-RL) addresses the limitations of conventional training methods in RL, especially in meta-RL environments where the agent faces new tasks. The adversarial component challenges the agent, forcing it to improve its decision-making abilities in dynamic and unpredictable situations. This component operates without relying on manual intervention or domain-specific knowledge, making it a highly versatile solution. Experiments conducted in multiple MT-RL environments demonstrate that adversarial training leads to better exploration and a deeper understanding of the environment. The adversarial training regime for MT-RL presents a new perspective on training and development for RL agents and is a valuable contribution to the field.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源