论文标题
一项关于图像分类的半半,自我和无监督学习的调查
A survey on Semi-, Self- and Unsupervised Learning for Image Classification
论文作者
论文摘要
尽管深度学习策略在计算机视觉任务中取得了出色的成果,但仍有一个问题:当前的策略在很大程度上依赖大量标记的数据。在许多实际问题中,创建如此众多的标记培训数据是不可行的。因此,通常将未标记的数据纳入训练过程中,以更少的标签达到相等的结果。由于很多并发研究,很难跟踪最近的发展。在这项调查中,我们概述了图像分类中经常使用的想法和方法,标签较少。我们根据其性能和常用思想,而不是细粒的分类法对34种方法进行了详细的比较。在我们的分析中,我们确定了导致未来研究机会的三个主要趋势。 1。在理论上,最新的方法可扩展到现实世界的应用,但是不考虑阶级不平衡,稳健性或模糊标签等问题。 2。获得与所有标签使用的可比结果所需的监督程度正在减少,因此需要将方法扩展到具有可变数量的类别的设置。 3。所有方法都共享一些共同的想法,但我们确定了不共享许多想法的方法群。我们表明,将不同群集的想法结合起来可以提高性能。
While deep learning strategies achieve outstanding results in computer vision tasks, one issue remains: The current strategies rely heavily on a huge amount of labeled data. In many real-world problems, it is not feasible to create such an amount of labeled training data. Therefore, it is common to incorporate unlabeled data into the training process to reach equal results with fewer labels. Due to a lot of concurrent research, it is difficult to keep track of recent developments. In this survey, we provide an overview of often used ideas and methods in image classification with fewer labels. We compare 34 methods in detail based on their performance and their commonly used ideas rather than a fine-grained taxonomy. In our analysis, we identify three major trends that lead to future research opportunities. 1. State-of-the-art methods are scaleable to real-world applications in theory but issues like class imbalance, robustness, or fuzzy labels are not considered. 2. The degree of supervision which is needed to achieve comparable results to the usage of all labels is decreasing and therefore methods need to be extended to settings with a variable number of classes. 3. All methods share some common ideas but we identify clusters of methods that do not share many ideas. We show that combining ideas from different clusters can lead to better performance.