论文标题

培训对抗者在深控制政策中利用弱点

Training Adversarial Agents to Exploit Weaknesses in Deep Control Policies

论文作者

Kuutti, Sampo, Fallah, Saber, Bowden, Richard

论文摘要

深度学习已成为针对各种控制问题的越来越常见的技术,例如机器人手臂操纵,机器人导航和自动驾驶汽车。但是,使用深层神经网络学习控制政策的缺点是它们不透明的性质和验证其安全性的困难。随着用于获得最先进结果的网络变得越来越深和复杂,他们所学的规则以及如何运作变得更具挑战性的理解。这提出了一个问题,因为在关键安全应用中,必须确保控制政策的安全性达到高信心水平。在本文中,我们提出了一个基于对抗强化学习的自动化黑匣子测试框架。该技术使用对抗代理,其目标是降低正在测试的目标模型的性能。我们通过培训对抗性加固学习剂来测试有关自动驾驶汽车问题的方法,该方法旨在导致深层神经网络驱动的自动驾驶汽车碰撞。比较了两个接受自主驾驶训练的神经网络,并使用测试的结果比较其学习的控制策略的鲁棒性。我们表明,所提出的框架能够在两种控制策略中发现弱点,在在线测试中尚不明显,因此证明了与手动测试方法相比具有重大好处。

Deep learning has become an increasingly common technique for various control problems, such as robotic arm manipulation, robot navigation, and autonomous vehicles. However, the downside of using deep neural networks to learn control policies is their opaque nature and the difficulties of validating their safety. As the networks used to obtain state-of-the-art results become increasingly deep and complex, the rules they have learned and how they operate become more challenging to understand. This presents an issue, since in safety-critical applications the safety of the control policy must be ensured to a high confidence level. In this paper, we propose an automated black box testing framework based on adversarial reinforcement learning. The technique uses an adversarial agent, whose goal is to degrade the performance of the target model under test. We test the approach on an autonomous vehicle problem, by training an adversarial reinforcement learning agent, which aims to cause a deep neural network-driven autonomous vehicle to collide. Two neural networks trained for autonomous driving are compared, and the results from the testing are used to compare the robustness of their learned control policies. We show that the proposed framework is able to find weaknesses in both control policies that were not evident during online testing and therefore, demonstrate a significant benefit over manual testing methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源