论文标题

优化低注释预算的积极学习

Optimizing Active Learning for Low Annotation Budgets

论文作者

Aggarwal, Umang, Popescu, Adrian, Hudelot, Céline

论文摘要

当我们无法假设大量注释的数据时,主动学习是一个很好的策略。它包括学习一个模型,以少量注释的数据(注释预算)以及选择注释的最佳点以改善先前的模型并获得概括。在深度学习中,通常将主动学习作为迭代过程实施,在该过程中,通过微调更新了连续的深层模型,但仍然构成一些问题。首先,初始的注释图像必须足够大才能训练深层模型。这样的假设很强,尤其是当总注释预算减少时。我们使用受转移学习启发的方法来解决此问题。预训练的模型被用作特征提取器,并且在主动迭代期间仅学习浅层分类器。第二个问题是概率或早期模型的特征估计的有效性。通常仅基于最后一次学习的模型,使用采集函数选择样品进行注释。我们介绍了一种新颖的采集功能,该功能利用了Al过程的迭代性质,以更健壮的方式选择样品。在最后两个学习的模型预测之间有最大转向不确定性的样本受到青睐。添加了一个多元化步骤,以从分类空间的不同区域选择样本,从而在我们的方法中引入了代表性组成部分。评估是针对具有三个平衡和不平衡数据集的竞争方法进行的,并且表现优于它们。

When we can not assume a large amount of annotated data , active learning is a good strategy. It consists in learning a model on a small amount of annotated data (annotation budget) and in choosing the best set of points to annotate in order to improve the previous model and gain in generalization. In deep learning, active learning is usually implemented as an iterative process in which successive deep models are updated via fine tuning, but it still poses some issues. First, the initial batch of annotated images has to be sufficiently large to train a deep model. Such an assumption is strong, especially when the total annotation budget is reduced. We tackle this issue by using an approach inspired by transfer learning. A pre-trained model is used as a feature extractor and only shallow classifiers are learned during the active iterations. The second issue is the effectiveness of probability or feature estimates of early models for AL task. Samples are generally selected for annotation using acquisition functions based only on the last learned model. We introduce a novel acquisition function which exploits the iterative nature of AL process to select samples in a more robust fashion. Samples for which there is a maximum shift towards uncertainty between the last two learned models predictions are favored. A diversification step is added to select samples from different regions of the classification space and thus introduces a representativeness component in our approach. Evaluation is done against competitive methods with three balanced and imbalanced datasets and outperforms them.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源