论文标题

统一的度量学习的统一相互信息视图:跨渗透与成对损失

A unifying mutual information view of metric learning: cross-entropy vs. pairwise losses

论文作者

Boudiaf, Malik, Rony, Jérôme, Ziko, Imtiaz Masud, Granger, Eric, Pedersoli, Marco, Piantanida, Pablo, Ayed, Ismail Ben

论文摘要

最近,深度度量学习(DML)的大量研究工作着重于设计复杂的成对距离损失,这些损失需要复杂的方案以简化优化,例如样品挖掘或配对加权。在DML中,分类的标准跨透镜损失在很大程度上被忽略了。从表面上看,跨凝结似乎与公制学习无关,因为它不明确涉及成对的距离。但是,我们提供了理论分析,将跨凝结与几个众所周知和最近的成对损失联系起来。我们的联系是从两个不同的角度绘制的:一个基于明确的优化见解;另一个关于标签和学识渊博特征之间相互信息的歧视性和生成性观点。首先,我们明确证明了跨渗透性是新成对损失的上限,该结构具有类似于各种成对损耗的结构:它最小化了阶层的距离,同时最大程度地提高了阶层间距离。结果,最小化跨渗透性可以看作是一种近似边界优化(或大量最小化)算法,以最大程度地减少该成对损耗。其次,我们表明,更一般而言,最小化跨膜片实际上等效于最大化相互信息,我们将几个众所周知的成对损失连接起来。此外,我们表明可以通过界关系明确地相互关联,各种标准的成对损失。我们的发现表明,跨凝性代表了最大化相互信息的代理 - 如成对损失所做的,而无需复杂的样品启发式方法。我们超过四个标准DML基准测试的实验强烈支持我们的发现。我们获得了最先进的结果,表现优于最近和复杂的DML方法。

Recently, substantial research efforts in Deep Metric Learning (DML) focused on designing complex pairwise-distance losses, which require convoluted schemes to ease optimization, such as sample mining or pair weighting. The standard cross-entropy loss for classification has been largely overlooked in DML. On the surface, the cross-entropy may seem unrelated and irrelevant to metric learning as it does not explicitly involve pairwise distances. However, we provide a theoretical analysis that links the cross-entropy to several well-known and recent pairwise losses. Our connections are drawn from two different perspectives: one based on an explicit optimization insight; the other on discriminative and generative views of the mutual information between the labels and the learned features. First, we explicitly demonstrate that the cross-entropy is an upper bound on a new pairwise loss, which has a structure similar to various pairwise losses: it minimizes intra-class distances while maximizing inter-class distances. As a result, minimizing the cross-entropy can be seen as an approximate bound-optimization (or Majorize-Minimize) algorithm for minimizing this pairwise loss. Second, we show that, more generally, minimizing the cross-entropy is actually equivalent to maximizing the mutual information, to which we connect several well-known pairwise losses. Furthermore, we show that various standard pairwise losses can be explicitly related to one another via bound relationships. Our findings indicate that the cross-entropy represents a proxy for maximizing the mutual information -- as pairwise losses do -- without the need for convoluted sample-mining heuristics. Our experiments over four standard DML benchmarks strongly support our findings. We obtain state-of-the-art results, outperforming recent and complex DML methods.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源