论文标题
油漆和蒸馏:通过语义传递网络增强3D对象检测
Paint and Distill: Boosting 3D Object Detection with Semantic Passing Network
论文作者
论文摘要
LIDAR或相机传感器的3D对象检测任务对于自动驾驶至关重要。先锋尝试多模式融合的尝试补充了稀疏的激光雷达点云,其中包括图像的丰富语义纹理信息,以额外的网络设计和开销为代价。在这项工作中,我们提出了一个名为SPNET的新型语义传递框架,以通过丰富的上下文绘画的指导来提高现有基于激光雷达的3D检测模型的性能,而推断期间没有额外的计算成本。我们的关键设计是首先通过训练语义绘制的教师模型来利用地面真相标签中潜在的启发性语义知识,然后指导纯LIDAR网络通过不同粒度的知识传递模块来了解语义绘制的表示:班级通过,Pixel-Wise Wise通过和实例传递。实验结果表明,所提出的SPNET可以与大多数现有的3D检测框架无缝合作,AP增益为1〜5%,甚至在KITTI测试基准上实现了新的最新3D检测性能。代码可在以下网址提供:https://github.com/jb892/spnet。
3D object detection task from lidar or camera sensors is essential for autonomous driving. Pioneer attempts at multi-modality fusion complement the sparse lidar point clouds with rich semantic texture information from images at the cost of extra network designs and overhead. In this work, we propose a novel semantic passing framework, named SPNet, to boost the performance of existing lidar-based 3D detection models with the guidance of rich context painting, with no extra computation cost during inference. Our key design is to first exploit the potential instructive semantic knowledge within the ground-truth labels by training a semantic-painted teacher model and then guide the pure-lidar network to learn the semantic-painted representation via knowledge passing modules at different granularities: class-wise passing, pixel-wise passing and instance-wise passing. Experimental results show that the proposed SPNet can seamlessly cooperate with most existing 3D detection frameworks with 1~5% AP gain and even achieve new state-of-the-art 3D detection performance on the KITTI test benchmark. Code is available at: https://github.com/jb892/SPNet.