论文标题
SCE:可扩展的网络嵌入最稀疏的切割
SCE: Scalable Network Embedding from Sparsest Cut
论文作者
论文摘要
大规模网络嵌入是要以无监督的方式学习每个节点的潜在表示,从而捕获基础图的固有属性和结构信息。在这一领域,许多流行的方法受到自然语言处理的跳过模型的影响。他们中的大多数使用对比目标来训练编码器,该编码器迫使类似对的嵌入到接近,而负样品的嵌入距离很远。这种对比性学习方法成功的关键是如何绘制正面样本和负面样本。尽管直接随机抽样产生的负样本通常令人满意,但绘制积极示例的方法仍然是一个热门话题。 在本文中,我们建议仅使用负面样本进行培训的无监督网络嵌入。我们的方法基于一个受众所周知的最稀疏剪切问题启发的新对比目标。为了解决潜在的优化问题,我们引入了一个拉普拉斯平滑技巧,该技巧将图形卷积运算符用作平滑节点表示的低通滤波器。结果模型由GCN型结构作为编码器和简单的损耗函数组成。值得注意的是,我们的模型不使用阳性样本,而只能用于培训的负样本,这不仅使实施和调整变得更加容易,而且还大大减少了训练时间。 最后,对现实世界数据集进行了广泛的实验研究。与图形,G2G和DGI等强基线相比,结果清楚地证明了我们新模型的准确性和可扩展性的优势。
Large-scale network embedding is to learn a latent representation for each node in an unsupervised manner, which captures inherent properties and structural information of the underlying graph. In this field, many popular approaches are influenced by the skip-gram model from natural language processing. Most of them use a contrastive objective to train an encoder which forces the embeddings of similar pairs to be close and embeddings of negative samples to be far. A key of success to such contrastive learning methods is how to draw positive and negative samples. While negative samples that are generated by straightforward random sampling are often satisfying, methods for drawing positive examples remains a hot topic. In this paper, we propose SCE for unsupervised network embedding only using negative samples for training. Our method is based on a new contrastive objective inspired by the well-known sparsest cut problem. To solve the underlying optimization problem, we introduce a Laplacian smoothing trick, which uses graph convolutional operators as low-pass filters for smoothing node representations. The resulting model consists of a GCN-type structure as the encoder and a simple loss function. Notably, our model does not use positive samples but only negative samples for training, which not only makes the implementation and tuning much easier, but also reduces the training time significantly. Finally, extensive experimental studies on real world data sets are conducted. The results clearly demonstrate the advantages of our new model in both accuracy and scalability compared to strong baselines such as GraphSAGE, G2G and DGI.