论文标题
共匹匹:半监督学习的自适应阈值
FreeMatch: Self-adaptive Thresholding for Semi-supervised Learning
论文作者
论文摘要
由于基于伪标签和一致性正则化的各种方法带来的令人印象深刻的表现,半监督学习(SSL)取得了巨大的成功。但是,我们认为现有方法可能无法更有效地使用未标记的数据,因为它们要么使用预定义 /固定阈值或临时阈值调整方案,从而导致性能较低和收敛速度缓慢。我们首先分析一个激励示例,以获取有关理想阈值与模型学习状况之间关系的直觉。基于分析,我们提议根据模型的学习状况以自适应方式调整自适应方式的置信度阈值。我们进一步引入了自适应阶级的公平正规化惩罚,以鼓励在早期培训阶段进行不同预测的模型。广泛的实验表明了自由比赛的优势,尤其是当标记的数据极为罕见时。在最新的CIFAR-10上,共匹匹降低了5.78%,13.59%和1.28%的错误率降低,cifar-10的flexMatch flexMatch flexMatch,每班1个标签,每类1个标签,每个类标签为4个标签,并分别为每个类带100个标签的Imagenet。此外,共匹匹可以提高不平衡SSL的性能。这些代码可以在https://github.com/microsoft/semi-supervised-learning上找到。
Semi-supervised Learning (SSL) has witnessed great success owing to the impressive performances brought by various methods based on pseudo labeling and consistency regularization. However, we argue that existing methods might fail to utilize the unlabeled data more effectively since they either use a pre-defined / fixed threshold or an ad-hoc threshold adjusting scheme, resulting in inferior performance and slow convergence. We first analyze a motivating example to obtain intuitions on the relationship between the desirable threshold and model's learning status. Based on the analysis, we hence propose FreeMatch to adjust the confidence threshold in a self-adaptive manner according to the model's learning status. We further introduce a self-adaptive class fairness regularization penalty to encourage the model for diverse predictions during the early training stage. Extensive experiments indicate the superiority of FreeMatch especially when the labeled data are extremely rare. FreeMatch achieves 5.78%, 13.59%, and 1.28% error rate reduction over the latest state-of-the-art method FlexMatch on CIFAR-10 with 1 label per class, STL-10 with 4 labels per class, and ImageNet with 100 labels per class, respectively. Moreover, FreeMatch can also boost the performance of imbalanced SSL. The codes can be found at https://github.com/microsoft/Semi-supervised-learning.