论文标题
建议在建议中进行辍学方法和实验验证的调查
A Survey on Dropout Methods and Experimental Verification in Recommendation
论文作者
论文摘要
过度拟合是机器学习中的一个常见问题,这意味着该模型在测试数据中表现较差,同时非常适合培训数据。在应对过度拟合的各种应对方法中,辍学是一种代表性的方式之一。从随机掉落神经元到掉落神经结构,辍学在改善模型性能方面取得了巨大的成功。尽管在过去几年中设计和广泛应用了各种辍学方法,但尚未对其有效性,应用程序场景和贡献进行全面总结,并在迄今为止进行了经验比较。现在是进行全面调查的合适时机。 在本文中,我们根据执行辍学操作的阶段,系统地检查以前的辍学方法并将其分为三个主要类别。具体而言,涉及在AI顶级会议或期刊上发表的七十多种辍学方法(例如,TKDE,KDD,ThewebConf,Sigir)。设计的分类法易于理解,并且能够包括新的辍学方法。然后,我们进一步讨论他们的应用程序方案,连接和贡献。为了验证不同辍学方法的有效性,对具有丰富异质信息的建议方案进行了广泛的实验。最后,我们提出了一些关于辍学的开放问题和潜在的研究指示,值得进一步探索。
Overfitting is a common problem in machine learning, which means the model too closely fits the training data while performing poorly in the test data. Among various methods of coping with overfitting, dropout is one of the representative ways. From randomly dropping neurons to dropping neural structures, dropout has achieved great success in improving model performances. Although various dropout methods have been designed and widely applied in past years, their effectiveness, application scenarios, and contributions have not been comprehensively summarized and empirically compared by far. It is the right time to make a comprehensive survey. In this paper, we systematically review previous dropout methods and classify them into three major categories according to the stage where dropout operation is performed. Specifically, more than seventy dropout methods published in top AI conferences or journals (e.g., TKDE, KDD, TheWebConf, SIGIR) are involved. The designed taxonomy is easy to understand and capable of including new dropout methods. Then, we further discuss their application scenarios, connections, and contributions. To verify the effectiveness of distinct dropout methods, extensive experiments are conducted on recommendation scenarios with abundant heterogeneous information. Finally, we propose some open problems and potential research directions about dropout that worth to be further explored.