论文标题
非阻滞批次A*(技术报告)
Non-Blocking Batch A* (Technical Report)
论文作者
论文摘要
传统上,启发式搜索一直依赖于手工制作或编程派生的启发式方法。神经网络(NNS)是较新的强大工具,可用于从各州学习复杂的映射到成本到启发式。但是,他们缓慢的单个推理时间是一个很大的开销,可以在优化的启发式搜索实现中大大减少计划时间。最近的几项工作描述了利用NN的批处理计算的方法,以减少计划中的开销,同时保持(子)最优性的界限。但是,所有这些方法在建立批处理时以“阻止”方式使用了NN启发式方法,而忽略了通常可以使用的快速计算可接受的启发式方法(例如现有的经典启发式术)。我们介绍了一种非阻滞批次A*(NBBA*),这是一种有界的次优方法,它懒惰地分批计算NN启发式方法,同时允许通过非NN启发式的扩展。我们展示了与当前的阻止替代方案相比,这种微妙但重要的变化如何导致扩展大幅减少,并看到该性能与计算出的NN和快速非NN启发式的批处理差异有关。
Heuristic search has traditionally relied on hand-crafted or programmatically derived heuristics. Neural networks (NNs) are newer powerful tools which can be used to learn complex mappings from states to cost-to-go heuristics. However, their slow single inference time is a large overhead that can substantially slow down planning time in optimized heuristic search implementations. Several recent works have described ways to take advantage of NN's batch computations to decrease overhead in planning, while retaining bounds on (sub)optimality. However, all these methods have used the NN heuristic in a "blocking" manner while building up their batches, and have ignored possible fast-to-compute admissible heuristics (e.g. existing classically derived heuristics) that are usually available to use. We introduce Non-Blocking Batch A* (NBBA*), a bounded suboptimal method which lazily computes the NN heuristic in batches while allowing expansions informed by a non-NN heuristic. We show how this subtle but important change can lead to substantial reductions in expansions compared to the current blocking alternative, and see that the performance is related to the information difference between the batch computed NN and fast non-NN heuristic.