论文标题

跳动抽样:非平稳环境的简单正则图学习

Hop Sampling: A Simple Regularized Graph Learning for Non-Stationary Environments

论文作者

Park, Young-Jin, Shin, Kyuyong, Kim, Kyung-Min

论文摘要

图表学习在广泛的应用中广受欢迎,例如社交网络分析,计算生物学和推荐系统。但是,由于许多学术研究的积极结果不同,由于非平稳环境,在现实世界应用中应用图形神经网络(GNN)仍然具有挑战性。流数据的基本分布意外变化,从而导致不同的图形结构(又称概念漂移)。因此,必须设计出强大的图形学习技术,以使模型不会过度拟合训练图。在这项工作中,我们提出了一种直接的正则化方法,可以有效地防止GNN过度捕获。跳跃采样随机选择了传播步骤的数量,而不是修复它,并鼓励模型学习所有中间传播层的有意义的节点表示形式,并体验在训练集中不在训练集中的各种合理图表。特别是,我们描述了建议系统中方法的用例,这是现实世界中非固定案例的代表性示例。我们在大规模的现实世界数据集上评估了跳动采样,并在线票息pusement of Line Wallet Tab中进行了在线A/B/N测试。实验结果表明,提出的方案提高了GNN的预测准确性。我们观察到的Hop采样可为NDCG和MAP提供7.97%和16.93%的改进,而在线服务中的非注册GNN模型则提供了改善。此外,使用Hop采样的模型减轻了GNN中的过度厚度问题,从而使模型和更多元化的表示形式更深入。

Graph representation learning is gaining popularity in a wide range of applications, such as social networks analysis, computational biology, and recommender systems. However, different with positive results from many academic studies, applying graph neural networks (GNNs) in a real-world application is still challenging due to non-stationary environments. The underlying distribution of streaming data changes unexpectedly, resulting in different graph structures (a.k.a., concept drift). Therefore, it is essential to devise a robust graph learning technique so that the model does not overfit to the training graphs. In this work, we present Hop Sampling, a straightforward regularization method that can effectively prevent GNNs from overfishing. The hop sampling randomly selects the number of propagation steps rather than fixing it, and by doing so, it encourages the model to learn meaningful node representation for all intermediate propagation layers and to experience a variety of plausible graphs that are not in the training set. Particularly, we describe the use case of our method in recommender systems, a representative example of the real-world non-stationary case. We evaluated hop sampling on a large-scale real-world LINE dataset and conducted an online A/B/n test in LINE Coupon recommender systems of LINE Wallet Tab. Experimental results demonstrate that the proposed scheme improves the prediction accuracy of GNNs. We observed hop sampling provides 7.97% and 16.93% improvements for NDCG and MAP compared to non-regularized GNN models in our online service. Furthermore, models using hop sampling alleviate the oversmoothing issue in GNNs enabling a deeper model as well as more diversified representation.

扫码加入交流群

加入微信交流群

微信交流群二维码

扫码加入学术交流群,获取更多资源