论文标题

自动发现可解释的计划策略

Automatic Discovery of Interpretable Planning Strategies

论文作者

Skirzyński, Julian, Becker, Frederic, Lieder, Falk

论文摘要

在做出决定时,人们经常忽略关键信息,或者被无关的信息所影响。减轻这些偏见的一种常见方法是为决策者,尤其是医生等专业人士提供决策辅助工具,例如决策树和流程图。设计有效的决策艾滋病是一个困难的问题。我们建议,最近开发了强化学习方法来发现聪明的启发式方法以良好的决策,可以部分利用以帮助人类专家在此设计过程中。利用上述方法的剩余最大障碍之一是,他们所学的政策对人们来说是不透明的。为了解决这个问题,我们介绍了AI-Interpret:一种将特质策略转换为简单且可解释的描述的一般方法。我们的算法结合了模仿学习和程序归纳的最新进展,以及一种新的聚类方法,用于识别大量演示的子集,这些演示可以由简单,高性能的决策规则准确地描述。我们评估了我们的新算法,并采用它来翻译通过Metalevel增强学习发现的信息收购政策。大型行为实验的结果表明,随着流程图显着改善了三个不同类别的顺序决策问题类别的人的计划策略和决策,因此AI解释产生的决策规则很大。此外,另一个实验表明,这种方法比培训人们的绩效反馈更有效。最后,一系列消融研究证实,AI解释对于发现可解释的决策规则至关重要。我们得出的结论是,此处提出的方法和发现是利用自动策略发现以改善人类决策的重要一步。

When making decisions, people often overlook critical information or are overly swayed by irrelevant information. A common approach to mitigate these biases is to provide decision-makers, especially professionals such as medical doctors, with decision aids, such as decision trees and flowcharts. Designing effective decision aids is a difficult problem. We propose that recently developed reinforcement learning methods for discovering clever heuristics for good decision-making can be partially leveraged to assist human experts in this design process. One of the biggest remaining obstacles to leveraging the aforementioned methods is that the policies they learn are opaque to people. To solve this problem, we introduce AI-Interpret: a general method for transforming idiosyncratic policies into simple and interpretable descriptions. Our algorithm combines recent advances in imitation learning and program induction with a new clustering method for identifying a large subset of demonstrations that can be accurately described by a simple, high-performing decision rule. We evaluate our new algorithm and employ it to translate information-acquisition policies discovered through metalevel reinforcement learning. The results of large behavioral experiments showed that prividing the decision rules generated by AI-Interpret as flowcharts significantly improved people's planning strategies and decisions across three diferent classes of sequential decision problems. Moreover, another experiment revealed that this approach is significantly more effective than training people by giving them performance feedback. Finally, a series of ablation studies confirmed that AI-Interpret is critical to the discovery of interpretable decision rules. We conclude that the methods and findings presented herein are an important step towards leveraging automatic strategy discovery to improve human decision-making.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源