论文标题

Demi:互相信息的歧视性估计器

DEMI: Discriminative Estimator of Mutual Information

论文作者

Liao, Ruizhi, Moyer, Daniel, Golland, Polina, Wells, William M.

论文摘要

在连续随机变量之间估计相互信息通常是棘手的,并且对于高维数据而言极具挑战性。最近的进展已利用神经网络来优化相互信息的变异下限。尽管对这个困难问题表现出了希望,但在理论上和经验上证明,变异方法具有严重的统计局限性:1)当基础相互信息低或高时,许多方法都难以产生准确的估计; 2)由此产生的估计器可能会遭受较高的差异。我们的方法是基于培训分类器,该分类器提供了从关节分布而不是从其边际分布的乘积中得出数据样本对的可能性。此外,我们建立了共同信息与分类器在测试集中产生的平均对数赔率估计之间的直接连接,从而导致了简单准确的互相估计器。从理论上讲,我们的方法和其他变分方法在达到最佳时是等效的,而我们的方法避开了变异结合。经验结果表明,在表示学习的背景下,我们的方法的准确性和估计器的优势。我们的演示可从https://github.com/rayruizhiliao/demi_mi_estimator获得。

Estimating mutual information between continuous random variables is often intractable and extremely challenging for high-dimensional data. Recent progress has leveraged neural networks to optimize variational lower bounds on mutual information. Although showing promise for this difficult problem, the variational methods have been theoretically and empirically proven to have serious statistical limitations: 1) many methods struggle to produce accurate estimates when the underlying mutual information is either low or high; 2) the resulting estimators may suffer from high variance. Our approach is based on training a classifier that provides the probability that a data sample pair is drawn from the joint distribution rather than from the product of its marginal distributions. Moreover, we establish a direct connection between mutual information and the average log odds estimate produced by the classifier on a test set, leading to a simple and accurate estimator of mutual information. We show theoretically that our method and other variational approaches are equivalent when they achieve their optimum, while our method sidesteps the variational bound. Empirical results demonstrate high accuracy of our approach and the advantages of our estimator in the context of representation learning. Our demo is available at https://github.com/RayRuizhiLiao/demi_mi_estimator.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源