论文标题
半塞亚姆培训浅脸学习
Semi-Siamese Training for Shallow Face Learning
论文作者
论文摘要
大多数现有的公共面部数据集,例如MS-CELEB-1M和VGGFACE2,都提供了广度(大量ID)和深度(足够数量的样本)的丰富信息。但是,在许多实际情况下,面部识别情况,训练数据集的深度有限,即每个ID都有两个面部图像。 $ \ textit {我们将这种情况定义为浅面脸学习,并发现现有的培训方法有问题。} $与深面数据不同,浅面脸数据缺乏类内的多样性。因此,它可能导致特征维度的崩溃,因此,学到的网络很容易在倒塌的维度中遭受变性和过拟合的困扰。在本文中,我们旨在通过引入一种名为Semiiamese Training(SST)的新型培训方法来解决问题。一对半启用网络构成了正向传播结构,并通过更新画廊队列计算训练损失,对浅层培训数据进行有效的优化。我们的方法是没有超依赖性的,因此可以灵活地与现有的损失功能和网络体系结构集成在一起。对面部识别的各种基准的广泛实验表明,该方法显着改善了训练,不仅在浅面学习中,而且在常规的深面数据中也可以改善训练。
Most existing public face datasets, such as MS-Celeb-1M and VGGFace2, provide abundant information in both breadth (large number of IDs) and depth (sufficient number of samples) for training. However, in many real-world scenarios of face recognition, the training dataset is limited in depth, i.e. only two face images are available for each ID. $\textit{We define this situation as Shallow Face Learning, and find it problematic with existing training methods.}$ Unlike deep face data, the shallow face data lacks intra-class diversity. As such, it can lead to collapse of feature dimension and consequently the learned network can easily suffer from degeneration and over-fitting in the collapsed dimension. In this paper, we aim to address the problem by introducing a novel training method named Semi-Siamese Training (SST). A pair of Semi-Siamese networks constitute the forward propagation structure, and the training loss is computed with an updating gallery queue, conducting effective optimization on shallow training data. Our method is developed without extra-dependency, thus can be flexibly integrated with the existing loss functions and network architectures. Extensive experiments on various benchmarks of face recognition show the proposed method significantly improves the training, not only in shallow face learning, but also for conventional deep face data.