Diffusion-Jump GNNs: Homophiliation via Learnable Metric Filters

29 Jun 2023  ·  Ahmed Begga, Francisco Escolano, Miguel Angel Lozano, Edwin R. Hancock ·

High-order Graph Neural Networks (HO-GNNs) have been developed to infer consistent latent spaces in the heterophilic regime, where the label distribution is not correlated with the graph structure. However, most of the existing HO-GNNs are hop-based, i.e., they rely on the powers of the transition matrix. As a result, these architectures are not fully reactive to the classification loss and the achieved structural filters have static supports. In other words, neither the filters' supports nor their coefficients can be learned with these networks. They are confined, instead, to learn combinations of filters. To address the above concerns, we propose Diffusion-jump GNNs a method relying on asymptotic diffusion distances that operates on jumps. A diffusion-pump generates pairwise distances whose projections determine both the support and coefficients of each structural filter. These filters are called jumps because they explore a wide range of scales in order to find bonds between scattered nodes with the same label. Actually, the full process is controlled by the classification loss. Both the jumps and the diffusion distances react to classification errors (i.e. they are learnable). Homophiliation, i.e., the process of learning piecewise smooth latent spaces in the heterophilic regime, is formulated as a Dirichlet problem: the known labels determine the border nodes and the diffusion-pump ensures a minimal deviation of the semi-supervised grouping from a canonical unsupervised grouping. This triggers the update of both the diffusion distances and, consequently, the jumps in order to minimize the classification error. The Dirichlet formulation has several advantages. It leads to the definition of structural heterophily, a novel measure beyond edge heterophily. It also allows us to investigate links with (learnable) diffusion distances, absorbing random walks and stochastic diffusion.

PDF Abstract

Results from the Paper


Task Dataset Model Metric Name Metric Value Global Rank Result Benchmark
Node Classification Actor DJ-GNN Accuracy 36.93 ± 0.84 # 20
Node Classification arXiv-year DJ-GNN Accuracy 49.21±0.20 # 10
Node Classification Chameleon DJ-GNN Accuracy 80.48±1.46 # 1
Node Classification Cornell DJ-GNN Accuracy 87.03±1.62 # 4
Node Classification Penn94 DJ-GNN Accuracy 84.84±0.34 # 8
Node Classification Squirrel DJ-GNN Accuracy 73.48±1.59 # 5
Node Classification Texas DJ-GNN Accuracy 92.43±3.15 # 2
Node Classification Wisconsin DJ-GNN Accuracy 92.54±3.70 # 2

Methods