论文标题
VTGNET:一个基于视觉的轨迹生成网络,用于城市环境中的自动驾驶汽车
VTGNet: A Vision-based Trajectory Generation Network for Autonomous Vehicles in Urban Environments
论文作者
论文摘要
自主驾驶的传统方法具有从感知,计划和控制的许多构件,因此由于复杂的假设和相互依存关系而难以推广到各种场景。最近,出现了端到端驾驶方法,该方法的性能很好,并通过直接从导出提供的数据中学习来概括到新环境中。但是,许多有关此主题的现有方法忽略了以检查驾驶行动的信心以及从驾驶错误中恢复的能力。在本文中,我们基于模仿学习开发了一种不确定性感知的端到端轨迹生成方法。它可以从前视摄像头图像中提取时空特征,以了解场景的理解,然后在未来几秒钟生成无碰撞轨迹。实验结果表明,在各种天气和照明条件下,我们的网络可以在不同的城市环境中可靠地产生轨迹,例如在交叉路口转动并放慢速度以避免碰撞。此外,闭环驾驶测试表明,所提出的方法比最先进的(SOTA)端到端控制方法获得了更好的跨场所/平台驾驶结果,在这种方法中,我们的模型可以从中心异位和异位置误差中恢复,并捕获具有高不确定性估计的危险病例的80%。
Traditional methods for autonomous driving are implemented with many building blocks from perception, planning and control, making them difficult to generalize to varied scenarios due to complex assumptions and interdependencies. Recently, the end-to-end driving method has emerged, which performs well and generalizes to new environments by directly learning from export-provided data. However, many existing methods on this topic neglect to check the confidence of the driving actions and the ability to recover from driving mistakes. In this paper, we develop an uncertainty-aware end-to-end trajectory generation method based on imitation learning. It can extract spatiotemporal features from the front-view camera images for scene understanding, and then generate collision-free trajectories several seconds into the future. The experimental results suggest that under various weather and lighting conditions, our network can reliably generate trajectories in different urban environments, such as turning at intersections and slowing down for collision avoidance. Furthermore, closed-loop driving tests suggest that the proposed method achieves better cross-scene/platform driving results than the state-of-the-art (SOTA) end-to-end control method, where our model can recover from off-center and off-orientation errors and capture 80% of dangerous cases with high uncertainty estimations.