论文标题
重新审视培训策略和深度度量学习中的概括性能
Revisiting Training Strategies and Generalization Performance in Deep Metric Learning
论文作者
论文摘要
深度度量学习(DML)可以说是每年通过许多拟议的方法学习视觉相似性的最具影响力的研究线之一。尽管该领域受益于快速进步,但训练方案,架构和参数选择的差异使得无偏的比较变得困难。为了提供一致的参考点,我们重新审视了最广泛使用的DML目标函数,并研究了关键参数选择以及通常被忽略的迷你批次采样过程。在一致的比较下,DML目标显示出比文献所示的饱和度要高得多。进一步基于我们的分析,我们发现了嵌入空间密度与压缩与DML模型的概括性能之间的相关性。利用这些见解,我们提出了一个简单但有效的培训正则化,以可靠地提高各种标准基准数据集上基于排名的DML模型的性能。代码和可公开访问的Wandb-Repo可在https://github.com/confusezius/revisiting_deep_metric_learning_pytorch上找到。
Deep Metric Learning (DML) is arguably one of the most influential lines of research for learning visual similarities with many proposed approaches every year. Although the field benefits from the rapid progress, the divergence in training protocols, architectures, and parameter choices make an unbiased comparison difficult. To provide a consistent reference point, we revisit the most widely used DML objective functions and conduct a study of the crucial parameter choices as well as the commonly neglected mini-batch sampling process. Under consistent comparison, DML objectives show much higher saturation than indicated by literature. Further based on our analysis, we uncover a correlation between the embedding space density and compression to the generalization performance of DML models. Exploiting these insights, we propose a simple, yet effective, training regularization to reliably boost the performance of ranking-based DML models on various standard benchmark datasets. Code and a publicly accessible WandB-repo are available at https://github.com/Confusezius/Revisiting_Deep_Metric_Learning_PyTorch.