论文标题
UNIDA3D:统一域自适应3D语义分割管道
UniDA3D: Unified Domain Adaptive 3D Semantic Segmentation Pipeline
论文作者
论文摘要
最先进的3D语义细分模型在现成的公共基准测试中进行了培训,但是当将这些经过良好训练的模型部署到新领域时,它们将不可避免地面临识别精度下降的挑战。在本文中,我们引入了一个统一的域自适应3D语义分割管道(UNIDA3D),以增强弱的概括能力,并弥合域之间的点分布差距。与以前仅关注单个适应任务的研究不同,UNIDA3D可以通过设计统一的源和目标主动采样策略来解决3D分割领域中的几个适应任务,该采样策略从源和目标域中选择了有效模型适应的最大信息子集。此外,从多模式2D-3D数据集的兴起中受益,UNIDA3D通过开发跨模式功能交互模块来研究实现多模式采样策略的可能性,该模块可以提取具有代表性的图像和点功能,以实现双向图像点功能相互作用,以实现安全模型模型适应。在实验上,UNIDA3D在许多适应任务中有效,包括:1)无监督的域适应性,2)无监督的少数射击域适应; 3)主动域适应。他们的结果表明,通过轻松将UNIDA3D与现成的3D分割基准耦合,可以增强这些基线的域概括能力。
State-of-the-art 3D semantic segmentation models are trained on off-the-shelf public benchmarks, but they will inevitably face the challenge of recognition accuracy drop when these well-trained models are deployed to a new domain. In this paper, we introduce a Unified Domain Adaptive 3D semantic segmentation pipeline (UniDA3D) to enhance the weak generalization ability, and bridge the point distribution gap between domains. Different from previous studies that only focus on a single adaptation task, UniDA3D can tackle several adaptation tasks in 3D segmentation field, by designing a unified source-and-target active sampling strategy, which selects a maximally-informative subset from both source and target domains for effective model adaptation. Besides, benefiting from the rise of multi-modal 2D-3D datasets, UniDA3D investigates the possibility of achieving a multi-modal sampling strategy, by developing a cross-modality feature interaction module that can extract a representative pair of image and point features to achieve a bi-directional image-point feature interaction for safe model adaptation. Experimentally, UniDA3D is verified to be effective in many adaptation tasks including: 1) unsupervised domain adaptation, 2) unsupervised few-shot domain adaptation; 3) active domain adaptation. Their results demonstrate that, by easily coupling UniDA3D with off-the-shelf 3D segmentation baselines, domain generalization ability of these baselines can be enhanced.