论文标题
了解情节加强学习中中毒攻击的局限性
Understanding the Limits of Poisoning Attacks in Episodic Reinforcement Learning
论文作者
论文摘要
为了了解强化学习的安全威胁(RL)算法,本文研究中毒攻击以操纵\ emph {any}秩序 - 最佳学习算法对偶发性RL中的有针对性政策,并研究了两种自然中毒攻击的潜在损害,即,\ emph}和奖励}和 ^ e}和\ emph {我们发现攻击的效果至关重要地取决于奖励是有界还是无限的。在有限的奖励设置中,我们表明只有奖励操纵或只有动作操纵不能保证成功的攻击。但是,通过结合奖励和动作操纵,对手可以操纵任何订单最佳学习算法,以遵循任何有针对性的策略,并使用$ \tildeθ(\ sqrt {t})$总攻击成本,这是订单最佳的攻击成本,这是最佳的,而没有任何知识的MDP。相比之下,在无限的奖励设置中,我们表明奖励操纵攻击足以使对手成功操纵任何订单最佳的学习算法,以使用$ \ tilde {o}(\ sqrt {t} {t})遵循任何有针对性的政策。我们的结果揭示了有关中毒攻击无法实现或无法实现的有用见解,并将刺激有关强大RL算法设计的更多作品。
To understand the security threats to reinforcement learning (RL) algorithms, this paper studies poisoning attacks to manipulate \emph{any} order-optimal learning algorithm towards a targeted policy in episodic RL and examines the potential damage of two natural types of poisoning attacks, i.e., the manipulation of \emph{reward} and \emph{action}. We discover that the effect of attacks crucially depend on whether the rewards are bounded or unbounded. In bounded reward settings, we show that only reward manipulation or only action manipulation cannot guarantee a successful attack. However, by combining reward and action manipulation, the adversary can manipulate any order-optimal learning algorithm to follow any targeted policy with $\tildeΘ(\sqrt{T})$ total attack cost, which is order-optimal, without any knowledge of the underlying MDP. In contrast, in unbounded reward settings, we show that reward manipulation attacks are sufficient for an adversary to successfully manipulate any order-optimal learning algorithm to follow any targeted policy using $\tilde{O}(\sqrt{T})$ amount of contamination. Our results reveal useful insights about what can or cannot be achieved by poisoning attacks, and are set to spur more works on the design of robust RL algorithms.