论文标题

来自人群注释的软标签汇总标签可改善分配变化下的不确定性估计

Aggregating Soft Labels from Crowd Annotations Improves Uncertainty Estimation Under Distribution Shift

论文作者

Wright, Dustin, Augenstein, Isabelle

论文摘要

为机器学习任务选择有效的培训信号很困难:专家注释很昂贵,众包注释可能不可靠。最近的工作表明,从人群注释中获取的标签上的分布中学习可以有效地估算性能和不确定性估计。但是,这主要是使用有限的软标记方法在内域设置中研究的。此外,没有一种方法被证明在跨任务中始终如一地表现良好,因此很难知道可以选择的先验。为了填补这些空白,本文提供了第一项大规模实证研究,该研究在室外环境中从人群中学习学习,系统地分析了4种语言和视觉任务的8种软标签方法。此外,我们建议通过简单的平均平均值汇总软标签,以便在整个任务之间达到一致的性能。我们证明,与从单个软标记方法中学习或对注释的多数投票相比,在大多数设置中,在大多数情况下都可以通过提高预测不确定性估计的分类器来产生分类器。我们还强调,在具有丰富或最小训练数据的制度中,选择软标签方法的选择不太重要,而对于高度主观的标签和中等量的训练数据,聚合可以对单个方法的不确定性估计得到显着改善。代码可以在https://github.com/copenlu/aggregating-crowd-annotations-ood上找到。

Selecting an effective training signal for machine learning tasks is difficult: expert annotations are expensive, and crowd-sourced annotations may not be reliable. Recent work has demonstrated that learning from a distribution over labels acquired from crowd annotations can be effective both for performance and uncertainty estimation. However, this has mainly been studied using a limited set of soft-labeling methods in an in-domain setting. Additionally, no one method has been shown to consistently perform well across tasks, making it difficult to know a priori which to choose. To fill these gaps, this paper provides the first large-scale empirical study on learning from crowd labels in the out-of-domain setting, systematically analyzing 8 soft-labeling methods on 4 language and vision tasks. Additionally, we propose to aggregate soft-labels via a simple average in order to achieve consistent performance across tasks. We demonstrate that this yields classifiers with improved predictive uncertainty estimation in most settings while maintaining consistent raw performance compared to learning from individual soft-labeling methods or taking a majority vote of the annotations. We additionally highlight that in regimes with abundant or minimal training data, the selection of soft labeling method is less important, while for highly subjective labels and moderate amounts of training data, aggregation yields significant improvements in uncertainty estimation over individual methods. Code can be found at https://github.com/copenlu/aggregating-crowd-annotations-ood.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源