论文标题
依赖实例的部分标签学习的进行性净化
Progressive Purification for Instance-Dependent Partial Label Learning
论文作者
论文摘要
部分标签学习(PLL)旨在从每个示例中训练多类分类器,每个示例都注释了一组候选标签,其中固定但未知的候选标签是正确的。在过去的几年中,基于在PLL中取得了许多理论上的进步,对候选标签的实例无关生成过程进行了广泛的研究。然而,候选标签始终在实践中依赖于实践,并且没有理论上的保证,即对实例依赖的PLL示例训练的模型可以收敛到理想的模型。在本文中,提出了一种理论上扎根的,实际上有效的方法,即依赖实例的部分标签学习的渐进性纯化。具体而言,POP更新学习模型并净化每个候选标签在每个时期逐渐设置的标签。从理论上讲,我们证明在模型可靠的情况下,POP可以适当地扩大该区域,并最终近似具有轻度假设的贝叶斯最佳分类器。从技术上讲,POP具有任意PLL损失的灵活性,并且可以在实例依赖性情况下提高先前PLL损失的性能。基准数据集和实际数据集的实验验证了所提出方法的有效性。
Partial label learning (PLL) aims to train multiclass classifiers from the examples each annotated with a set of candidate labels where a fixed but unknown candidate label is correct. In the last few years, the instance-independent generation process of candidate labels has been extensively studied, on the basis of which many theoretical advances have been made in PLL. Nevertheless, the candidate labels are always instance-dependent in practice and there is no theoretical guarantee that the model trained on the instance-dependent PLL examples can converge to an ideal one. In this paper, a theoretically grounded and practically effective approach named POP, i.e. PrOgressive Purification for instance-dependent partial label learning, is proposed. Specifically, POP updates the learning model and purifies each candidate label set progressively in every epoch. Theoretically, we prove that POP enlarges the region appropriately fast where the model is reliable, and eventually approximates the Bayes optimal classifier with mild assumptions. Technically, POP is flexible with arbitrary PLL losses and could improve the performance of the previous PLL losses in the instance-dependent case. Experiments on the benchmark datasets and the real-world datasets validate the effectiveness of the proposed method.