论文标题
推动视觉大满贯旋转的信封
Pushing the Envelope of Rotation Averaging for Visual SLAM
论文作者
论文摘要
作为运动(SFM)和同时定位和映射(SLAM)系统的结构的重要组成部分,在过去几年中进行了广泛的研究,并继续吸引令人振奋的研究注意力。尽管诸如捆绑调整之类的规范方法主要是在大多数最先进的SLAM系统中继承的,以估算和更新机器人导航中的轨迹,但SLAM系统中捆绑调整的实际实施在本质上受到了高计算复杂性,不可靠的融合,不可靠的融合和理想初始化的严格要求。在本文中,我们提高了这些局限性,并提出了一种新型的优化主链,用于视觉大满贯系统,我们利用旋转平均来提高常规单眼大满贯管道的准确性,效率和鲁棒性。在我们的方法中,我们首先将相机刚体变换中的旋转和翻译参数解次,并将高维的非矛盾非线性问题转换为较低尺寸的可拖动线性子问题,并表明可以使用适当的约束独立解决子问题。我们在姿势图优化中使用$ l_1 $ norm应用比例参数,以解决针对异常值的旋转稳定性。我们进一步验证了我们提出的方法的全球最优性,重新访问和解决初始化方案,纯粹的旋转场景处理和离群处理方法。我们证明,与公共基准相比,我们的方法可以以可比的准确性表现出高达10倍的速度。
As an essential part of structure from motion (SfM) and Simultaneous Localization and Mapping (SLAM) systems, motion averaging has been extensively studied in the past years and continues to attract surging research attention. While canonical approaches such as bundle adjustment are predominantly inherited in most of state-of-the-art SLAM systems to estimate and update the trajectory in the robot navigation, the practical implementation of bundle adjustment in SLAM systems is intrinsically limited by the high computational complexity, unreliable convergence and strict requirements of ideal initializations. In this paper, we lift these limitations and propose a novel optimization backbone for visual SLAM systems, where we leverage rotation averaging to improve the accuracy, efficiency and robustness of conventional monocular SLAM pipelines. In our approach, we first decouple the rotational and translational parameters in the camera rigid body transformation and convert the high-dimensional non-convex nonlinear problem into tractable linear subproblems in lower dimensions, and show that the subproblems can be solved independently with proper constraints. We apply the scale parameter with $l_1$-norm in the pose-graph optimization to address the rotation averaging robustness against outliers. We further validate the global optimality of our proposed approach, revisit and address the initialization schemes, pure rotational scene handling and outlier treatments. We demonstrate that our approach can exhibit up to 10x faster speed with comparable accuracy against the state of the art on public benchmarks.