论文标题
密集预测任务的多任务学习:一项调查
Multi-Task Learning for Dense Prediction Tasks: A Survey
论文作者
论文摘要
随着深度学习的出现,许多密集的预测任务,即产生像素级预测的任务,都显着改善了性能。典型的方法是隔离学习这些任务,即为每个任务培训一个单独的神经网络。然而,最近的多任务学习(MTL)技术已显示出令人鼓舞的结果W.R.T.绩效,计算和/或内存足迹,通过通过学习的共享表示形式共同处理多个任务。在这项调查中,我们对计算机视觉中MTL的最新深度学习方法提供了全面的看法,明确强调了密集的预测任务。我们的贡献涉及以下内容。首先,我们从网络体系结构观察点考虑MTL。我们包括广泛的概述,并讨论最近流行的MTL模型的优势/缺点。其次,我们研究了各种优化方法来解决多个任务的联合学习。我们总结了这些作品的定性要素,并探索了它们的共同点和差异。最后,我们对各种密集的预测基准进行了广泛的实验评估,以检查不同方法的利弊,包括基于建筑和优化的策略。
With the advent of deep learning, many dense prediction tasks, i.e. tasks that produce pixel-level predictions, have seen significant performance improvements. The typical approach is to learn these tasks in isolation, that is, a separate neural network is trained for each individual task. Yet, recent multi-task learning (MTL) techniques have shown promising results w.r.t. performance, computations and/or memory footprint, by jointly tackling multiple tasks through a learned shared representation. In this survey, we provide a well-rounded view on state-of-the-art deep learning approaches for MTL in computer vision, explicitly emphasizing on dense prediction tasks. Our contributions concern the following. First, we consider MTL from a network architecture point-of-view. We include an extensive overview and discuss the advantages/disadvantages of recent popular MTL models. Second, we examine various optimization methods to tackle the joint learning of multiple tasks. We summarize the qualitative elements of these works and explore their commonalities and differences. Finally, we provide an extensive experimental evaluation across a variety of dense prediction benchmarks to examine the pros and cons of the different methods, including both architectural and optimization based strategies.