论文标题
可证明的安全强化学习:概念分析,调查和基准测试
Provably Safe Reinforcement Learning: Conceptual Analysis, Survey, and Benchmarking
论文作者
论文摘要
确保增强学习(RL)算法的安全性对于释放其对许多现实世界任务的潜力至关重要。但是,香草RL和最安全的RL方法不能保证安全。近年来,已经提出了几种方法来为RL提供硬安全保证,这对于不安全行动可能带来灾难性后果的应用至关重要。然而,这些可证明的安全的RL方法尚无全面比较。因此,我们介绍了现有可证明的安全RL方法的分类,呈现了连续和离散的动作空间的概念基础,并从经验上进行基准的现有方法。我们根据它们如何适应动作来对方法进行分类:替换动作,动作投影和动作掩盖。我们对倒置的实验和四稳定稳定任务表明,尽管实现相对简单的实现,但动作更换是这些应用的最佳方法。此外,增加奖励罚款,每次进行安全验证时,都会在我们的实验中提高培训表现。最后,我们提供了根据安全规范,RL算法和动作空间类型选择可证明的安全RL方法的实用指南。
Ensuring the safety of reinforcement learning (RL) algorithms is crucial to unlock their potential for many real-world tasks. However, vanilla RL and most safe RL approaches do not guarantee safety. In recent years, several methods have been proposed to provide hard safety guarantees for RL, which is essential for applications where unsafe actions could have disastrous consequences. Nevertheless, there is no comprehensive comparison of these provably safe RL methods. Therefore, we introduce a categorization of existing provably safe RL methods, present the conceptual foundations for both continuous and discrete action spaces, and empirically benchmark existing methods. We categorize the methods based on how they adapt the action: action replacement, action projection, and action masking. Our experiments on an inverted pendulum and a quadrotor stabilization task indicate that action replacement is the best-performing approach for these applications despite its comparatively simple realization. Furthermore, adding a reward penalty, every time the safety verification is engaged, improved training performance in our experiments. Finally, we provide practical guidance on selecting provably safe RL approaches depending on the safety specification, RL algorithm, and type of action space.