论文标题
自我监督的真实访问现场一代
Self-Supervised Real-to-Sim Scene Generation
论文作者
论文摘要
综合数据正在作为对监督深度学习的可伸缩性问题的有希望的解决方案,尤其是当真实数据难以获取或难以注释时。但是,当域专家必须手动和艰苦地监督该过程时,合成数据生成本身可能会非常昂贵。此外,由于域间隙,经过合成数据训练的神经网络通常在实际数据上表现不佳。为了解决这些挑战,我们提出了SIM2SG,这是一种自我保护的自动场景生成技术,用于匹配真实数据的分布。重要的是,SIM2SG不需要实际数据集中的监督,因此使其适用于难以获得这样的注释的情况。 SIM2SG旨在通过匹配真实数据的内容以及匹配源和目标域中的功能来弥合内容和外观差距。由于标记的数据集的可用性有限,我们选择场景图(SG)生成作为下游任务。实验表明,在几个合成数据集以及现实世界中的KITTI数据集上,在定性和定量上降低域间隙方面表现出显着改善。
Synthetic data is emerging as a promising solution to the scalability issue of supervised deep learning, especially when real data are difficult to acquire or hard to annotate. Synthetic data generation, however, can itself be prohibitively expensive when domain experts have to manually and painstakingly oversee the process. Moreover, neural networks trained on synthetic data often do not perform well on real data because of the domain gap. To solve these challenges, we propose Sim2SG, a self-supervised automatic scene generation technique for matching the distribution of real data. Importantly, Sim2SG does not require supervision from the real-world dataset, thus making it applicable in situations for which such annotations are difficult to obtain. Sim2SG is designed to bridge both the content and appearance gaps, by matching the content of real data, and by matching the features in the source and target domains. We select scene graph (SG) generation as the downstream task, due to the limited availability of labeled datasets. Experiments demonstrate significant improvements over leading baselines in reducing the domain gap both qualitatively and quantitatively, on several synthetic datasets as well as the real-world KITTI dataset.