高级检索

    面向非同配图的非对称自监督学习方法

    Asymmetric Self-Supervised Learning on Non-Homophilic Graphs

    • 摘要: 自监督学习已逐渐成为解决传统图神经网络模型因为过度依赖标签而导致模型泛化性能差的一种新的学习范式,该方法利用数据的固有结构和属性来生成监督信息,而不依赖于标记数据. 然而,大多数现有的自监督学习方法的前提假设是图具有同配性,不能较好地推广到异配性强的图,即连接的节点具有不同的类别和不同的特征. 研究非同配图的自监督学习,不依赖图的同配性假设,设计了一种非对称自监督学习框架MIRROR,通过捕获节点1阶邻域信息和自适应选择高阶邻域信息来学习节点的自监督信息. 根据预测邻域上下文信息和估计的高阶互信息进行联合优化. 模型在多个同配图数据集和非同配图数据集上进行了大量实验,与最新的基线相比都取得了较优的效果,在多个下游任务上的优越性也表明了提出的框架具有较好的泛化性能.

       

      Abstract: Self-supervised learning has emerged as a promising approach in addressing the issues of label dependency and poor generalization performance in traditional graph neural networks (GNNs). This method leverages the inherent structure and properties of the data to generate supervisory information without relying on extensive labeled datasets. However, most existing self-supervised learning methods hold the assumption that the graphs are homophilic and consequently fail to generalize well to graphs with high heterophily, where connected nodes may have different class labels and dissimilar features. In this paper, we study the problem by developing an asymmetric self-supervised learning on non-homophilic graphs (MIRROR) framework, which is easy to implement and does not rely on random graph augmentations and homophily assumptions. Inspired by the designs of existing heterophilic graph neural networks, MIRROR learns the node representations by capturing one-hop local neighborhood information and informative distant neighbors. Such two properties are obtained through carefully designed pretext tasks that are optimized based on predicted neighborhood context and estimated high-order mutual information. Extensive experiments on various types of graph benchmarks demonstrate that our proposed framework can achieve better performance compared with competitive baselines on non-homophilic graphs. The superiority in multiple downstream tasks also demonstrates the generalization capability of MIRROR.

       

    /

    返回文章
    返回