论文标题
原型完成,具有原始知识,以进行几次学习
Prototype Completion with Primitive Knowledge for Few-Shot Learning
论文作者
论文摘要
很少有学习是一项具有挑战性的任务,该任务旨在学习一个很少有例子的新颖课程的分类器。基于训练前的元学习方法可以通过预训练提取器进行预训练,然后通过基于质心的元学习来有效地解决该问题。但是,结果表明,微调步骤可以非常微不足道。在本文中,1)我们找出关键原因,即,在预训练的特征空间中,基类已经形成了紧凑的簇,而新颖的类作为具有较大差异的组,这意味着对特征提取器进行微调的意义较小。 2)我们专注于估计元学习过程中更多代表性原型的,而不是对特征提取器进行微调。因此,我们提出了一个基于原型完成原型的元学习框架。该框架首先引入原始知识(即类级部分或属性注释),并将代表性属性特征作为先验提取。然后,我们设计了一个原型完成网络,以学习使用这些先验完成原型。为了避免由原始知识噪音或类别差异引起的原型完成误差,我们进一步开发了一种基于高斯的原型融合策略,该策略通过利用未标记的样本来结合基于均值的原型和完成的原型。广泛的实验表明我们的方法:(i)可以获得更准确的原型; (ii)就分类准确性而言,最先进的技术优于最新技术。我们的代码可在线提供。
Few-shot learning is a challenging task, which aims to learn a classifier for novel classes with few examples. Pre-training based meta-learning methods effectively tackle the problem by pre-training a feature extractor and then fine-tuning it through the nearest centroid based meta-learning. However, results show that the fine-tuning step makes very marginal improvements. In this paper, 1) we figure out the key reason, i.e., in the pre-trained feature space, the base classes already form compact clusters while novel classes spread as groups with large variances, which implies that fine-tuning the feature extractor is less meaningful; 2) instead of fine-tuning the feature extractor, we focus on estimating more representative prototypes during meta-learning. Consequently, we propose a novel prototype completion based meta-learning framework. This framework first introduces primitive knowledge (i.e., class-level part or attribute annotations) and extracts representative attribute features as priors. Then, we design a prototype completion network to learn to complete prototypes with these priors. To avoid the prototype completion error caused by primitive knowledge noises or class differences, we further develop a Gaussian based prototype fusion strategy that combines the mean-based and completed prototypes by exploiting the unlabeled samples. Extensive experiments show that our method: (i) can obtain more accurate prototypes; (ii) outperforms state-of-the-art techniques by 2% - 9% in terms of classification accuracy. Our code is available online.