论文标题
重新审视神经协作过滤与矩阵分解
Neural Collaborative Filtering vs. Matrix Factorization Revisited
论文作者
论文摘要
十多年来,基于嵌入的模型一直是协作过滤的最新技术。传统上,DOT产物或高阶等价物已被用来结合两个或多个嵌入,例如,最著名的是矩阵分解。近年来,有人建议将DOT产品替换为学到的相似性,例如使用多层感知器(MLP)。这种方法通常称为神经协作过滤(NCF)。在这项工作中,我们重新审视了NCF论文的实验,该纸张使用MLP推广了相似之处。首先,我们表明,通过适当的超参数选择,一种简单的点产品大大优于所提出的相似之处。其次,尽管MLP理论上可以近似任何功能,但我们表明,学习MLP的DOT产品是不平凡的。最后,我们讨论应用基于MLP的相似性时出现的实际问题,并表明MLP太昂贵了,无法用于生产环境中的项目推荐,而DOT产品则可以应用非常有效的检索算法。我们得出的结论是,应将MLP谨慎用作嵌入组合仪,并且点产品可能是更好的默认选择。
Embedding based models have been the state of the art in collaborative filtering for over a decade. Traditionally, the dot product or higher order equivalents have been used to combine two or more embeddings, e.g., most notably in matrix factorization. In recent years, it was suggested to replace the dot product with a learned similarity e.g. using a multilayer perceptron (MLP). This approach is often referred to as neural collaborative filtering (NCF). In this work, we revisit the experiments of the NCF paper that popularized learned similarities using MLPs. First, we show that with a proper hyperparameter selection, a simple dot product substantially outperforms the proposed learned similarities. Second, while a MLP can in theory approximate any function, we show that it is non-trivial to learn a dot product with an MLP. Finally, we discuss practical issues that arise when applying MLP based similarities and show that MLPs are too costly to use for item recommendation in production environments while dot products allow to apply very efficient retrieval algorithms. We conclude that MLPs should be used with care as embedding combiner and that dot products might be a better default choice.