论文标题
Segdiscover:通过无监督的语义分割发现视觉概念
SegDiscover: Visual Concept Discovery via Unsupervised Semantic Segmentation
论文作者
论文摘要
长期以来,视觉概念发现对于提高神经网络的可解释性很重要,因为一群具有语义上有意义的概念将为我们提供展示可理解推理过程的机器学习模型的起点。以前的方法存在缺点:他们要么依赖于标有“有用”对象的人类偏见的标记支持集,要么无法识别单个图像中发生的多个概念。我们将概念发现任务重新构架为无监督的语义分割问题,并呈现Segdiscover,这是一个新颖的框架,从图像数据集中发现具有复杂场景的无监督的图像数据集中的语义有意义的视觉概念。我们的方法包含三个重要的部分:从原始图像中生成概念基础,通过在自我监督预审计的编码器的潜在空间中聚类来发现概念,以及通过神经网络平滑的概念改进。实验结果提供了证据表明,我们的方法可以在单个图像中发现多个概念,并且在复杂数据集(例如CityScapes和Coco-stuff)上胜过最先进的无监督方法。通过比较不同编码器获得的结果,我们的方法可以进一步用作神经网络解释工具。
Visual concept discovery has long been deemed important to improve interpretability of neural networks, because a bank of semantically meaningful concepts would provide us with a starting point for building machine learning models that exhibit intelligible reasoning process. Previous methods have disadvantages: either they rely on labelled support sets that incorporate human biases for objects that are "useful," or they fail to identify multiple concepts that occur within a single image. We reframe the concept discovery task as an unsupervised semantic segmentation problem, and present SegDiscover, a novel framework that discovers semantically meaningful visual concepts from imagery datasets with complex scenes without supervision. Our method contains three important pieces: generating concept primitives from raw images, discovering concepts by clustering in the latent space of a self-supervised pretrained encoder, and concept refinement via neural network smoothing. Experimental results provide evidence that our method can discover multiple concepts within a single image and outperforms state-of-the-art unsupervised methods on complex datasets such as Cityscapes and COCO-Stuff. Our method can be further used as a neural network explanation tool by comparing results obtained by different encoders.