论文标题
剥夺建议的隐性反馈
Denoising Implicit Feedback for Recommendation
论文作者
论文摘要
隐式反馈的无处不在使它们成为构建在线推荐系统的默认选择。尽管大量隐式反馈减轻了数据稀疏问题,但缺点是它们在反映用户的实际满意度方面并不那么干净。例如,在电子商务中,很大一部分点击并未转化为购买,许多购买最终会得到负面评论。因此,重要的是说明建议培训的隐性反馈中不可避免的噪声。但是,关于推荐的工作很少,考虑到隐性反馈的嘈杂性质。 在这项工作中,我们探讨了将隐性反馈进行推荐培训的核心主题。我们发现嘈杂隐式反馈的严重负面影响,即拟合嘈杂的数据可阻止推荐人学习实际的用户偏好。我们的目标是确定和修剪嘈杂的互动,以提高推荐培训的质量。通过观察正常推荐训练的过程,我们发现嘈杂的反馈通常在早期阶段具有较大的损失值。受这一观察的启发,我们提出了一种名为Adaptive denoising Training(ADT)的新培训策略,该培训在训练过程中适应了嘈杂的互动。具体而言,我们设计了两个自适应损失配方的范式:截断的损失,在每次迭代中都丢弃具有动态阈值的大损失样本;并重新享受损失,可适应大量样本的重量。我们将两个范式实例化在广泛使用的二进制跨透明拷贝丢失上,并对三个代表推荐人进行测试拟议的ADT策略。对三个基准测试的广泛实验表明,ADT显着提高了针对正常训练的推荐质量。
The ubiquity of implicit feedback makes them the default choice to build online recommender systems. While the large volume of implicit feedback alleviates the data sparsity issue, the downside is that they are not as clean in reflecting the actual satisfaction of users. For example, in E-commerce, a large portion of clicks do not translate to purchases, and many purchases end up with negative reviews. As such, it is of critical importance to account for the inevitable noises in implicit feedback for recommender training. However, little work on recommendation has taken the noisy nature of implicit feedback into consideration. In this work, we explore the central theme of denoising implicit feedback for recommender training. We find serious negative impacts of noisy implicit feedback,i.e., fitting the noisy data prevents the recommender from learning the actual user preference. Our target is to identify and prune noisy interactions, so as to improve the quality of recommender training. By observing the process of normal recommender training, we find that noisy feedback typically has large loss values in the early stages. Inspired by this observation, we propose a new training strategy namedAdaptive Denoising Training(ADT), which adaptively prunes noisy interactions during training. Specifically, we devise two paradigms for adaptive loss formulation: Truncated Loss that discards the large-loss samples with a dynamic threshold in each iteration; and reweighted Loss that adaptively lowers the weight of large-loss samples. We instantiate the two paradigms on the widely used binary cross-entropy loss and test the proposed ADT strategies on three representative recommenders. Extensive experiments on three benchmarks demonstrate that ADT significantly improves the quality of recommendation over normal training.