论文标题
通过模型可解释性深入积极学习
Deep Active Learning by Model Interpretability
论文作者
论文摘要
然而,深度神经网络(DNN)最近在各种研究任务中取得了成功,这在很大程度上取决于大量标记的样本。这可能需要在现实世界中的大量注释成本。幸运的是,积极学习是一种以最低的注释成本来训练高性能模型的有前途的方法。在深度学习环境中,积极学习的关键问题是如何精确确定DNN样品的信息性。在本文中,受到DNN中的线性解释性的启发,我们将样本的线性分离区域引入了主动学习问题,并通过模型可解释性(DAMI)提出了一种新型的深层积极学习方法。为了保持整个未标记数据的最大代表性,Dami试图在DNN中通过零件线性解释性引入的不同线性可分离区域选择和标记样品。我们专注于建模用于对表格数据进行建模的多层感知(MLP)。具体而言,我们将MLP中的本地零件解释用作每个样本的表示,然后直接运行K-中心聚类以选择和标记样本。需要指出的是,DAMI的整个过程不需要任何超参数手动调节。为了验证我们方法的有效性,已经在几个表格数据集上进行了广泛的实验。实验结果表明,达米不断优于几个最先进的方法。
Recent successes of Deep Neural Networks (DNNs) in a variety of research tasks, however, heavily rely on the large amounts of labeled samples. This may require considerable annotation cost in real-world applications. Fortunately, active learning is a promising methodology to train high-performing model with minimal annotation cost. In the deep learning context, the critical question of active learning is how to precisely identify the informativeness of samples for DNN. In this paper, inspired by piece-wise linear interpretability in DNN, we introduce the linearly separable regions of samples to the problem of active learning, and propose a novel Deep Active learning approach by Model Interpretability (DAMI). To keep the maximal representativeness of the entire unlabeled data, DAMI tries to select and label samples on different linearly separable regions introduced by the piece-wise linear interpretability in DNN. We focus on modeling Multi-Layer Perception (MLP) for modeling tabular data. Specifically, we use the local piece-wise interpretation in MLP as the representation of each sample, and directly run K-Center clustering to select and label samples. To be noted, this whole process of DAMI does not require any hyper-parameters to tune manually. To verify the effectiveness of our approach, extensive experiments have been conducted on several tabular datasets. The experimental results demonstrate that DAMI constantly outperforms several state-of-the-art compared approaches.