论文标题
通过音频功能对音乐推荐的上下文个性化重新排序
Contextual Personalized Re-Ranking of Music Recommendations through Audio Features
论文作者
论文摘要
用户能够通过Spotify,Pandora和Deezer等音乐流媒体服务访问数百万首歌。访问如此大的目录,创造了相关歌曲建议的需求。但是,用户偏好本质上是高度主观的,并且根据环境(例如,早上适合的音乐在晚上不那么合适)。此外,一个用户在给定上下文中可能更喜欢的音乐可能与其他用户在同一上下文中更喜欢的音乐(即,被认为早上好音乐在用户之间有所不同)。准确地表示这些偏好对于创建准确有效的歌曲建议至关重要。用户对歌曲的喜好可以基于高级音频功能,例如节奏和价。因此,在本文中,我们基于在特定上下文条件下用户偏好的音频特征表示,提出了一种上下文重新排列算法。我们使用#NowPlaying-RS数据集评估了重新排列算法的性能,该数据集存在于Twitter上散发出的用户侦听事件,并充满了歌曲音频功能。我们根据这些音频功能比较了一个全局(所有用户的上下文)和个性化(每个用户的上下文)模型。全局模型根据所有用户的偏好创建每个上下文条件的音频特征表示。与全球模型不同,个性化模型创建了上下文条件的特定用户音频特征表示,并在333个不同的用户中进行了测量。我们表明,使用精度和平均平均精度指标进行评估时,个性化模型优于全局模型。
Users are able to access millions of songs through music streaming services like Spotify, Pandora, and Deezer. Access to such large catalogs, created a need for relevant song recommendations. However, user preferences are highly subjective in nature and change according to context (e.g., music that is suitable in the morning is not as suitable in the evening). Moreover, the music one user may prefer in a given context may be different from what another user prefers in the same context (i.e., what is considered good morning music differs across users). Accurately representing these preferences is essential to creating accurate and effective song recommendations. User preferences for songs can be based on high level audio features, such as tempo and valence. In this paper, we therefore propose a contextual re-ranking algorithm, based on audio feature representations of user preferences in specific contextual conditions. We evaluate the performance of our re-ranking algorithm using the #NowPlaying-RS dataset, which exists of user listening events crawled from Twitter and is enriched with song audio features. We compare a global (context for all users) and personalized (context for each user) model based on these audio features. The global model creates an audio feature representation of each contextual condition based on the preferences of all users. Unlike the global model, the personalized model creates user-specific audio feature representations of contextual conditions, and is measured across 333 distinct users. We show that the personalized model outperforms the global model when evaluated using the precision and mean average precision metrics.