论文标题
通过分解的确定点过程,图形卷积神经网络具有不同的负样品
Graph Convolutional Neural Networks with Diverse Negative Samples via Decomposed Determinant Point Processes
论文作者
论文摘要
图形卷积网络(GCN)通过从节点及其拓扑中提取高级特征,在图表学习方面取得了巨大的成功。由于GCN通常遵循通信机制,因此每个节点汇总了其一阶邻居的信息以更新其表示形式。结果,它们之间具有边缘之间的节点的表示应呈正相关,因此可以被视为阳性样品。但是,整个图中有更多的非纽布节点,这些节点为表示更新提供了多种有用的信息。两个非染色节点通常具有不同的表示,可以看作是负样本。除节点表示外,图的结构信息对于学习也至关重要。在本文中,我们在确定点过程(DPP)中使用了质量多样性分解来获得不同的负样本。在定义所有非邻个节点的不同子集上的分布时,我们同时将图形结构信息和节点表示。由于DPP采样过程需要矩阵本特征值分解,因此我们提出了一种新的最短标准碱方法来提高计算效率。最后,我们将所获得的负样品纳入图卷积操作中。这些想法在节点分类任务的实验中进行经验评估。这些实验表明,新提出的方法不仅可以改善标准表示学习的整体性能,而且可以显着减轻过度平滑的问题。
Graph convolutional networks (GCNs) have achieved great success in graph representation learning by extracting high-level features from nodes and their topology. Since GCNs generally follow a message-passing mechanism, each node aggregates information from its first-order neighbour to update its representation. As a result, the representations of nodes with edges between them should be positively correlated and thus can be considered positive samples. However, there are more non-neighbour nodes in the whole graph, which provide diverse and useful information for the representation update. Two non-adjacent nodes usually have different representations, which can be seen as negative samples. Besides the node representations, the structural information of the graph is also crucial for learning. In this paper, we used quality-diversity decomposition in determinant point processes (DPP) to obtain diverse negative samples. When defining a distribution on diverse subsets of all non-neighbouring nodes, we incorporate both graph structure information and node representations. Since the DPP sampling process requires matrix eigenvalue decomposition, we propose a new shortest-path-base method to improve computational efficiency. Finally, we incorporate the obtained negative samples into the graph convolution operation. The ideas are evaluated empirically in experiments on node classification tasks. These experiments show that the newly proposed methods not only improve the overall performance of standard representation learning but also significantly alleviate over-smoothing problems.