View a PDF of the paper titled Supercharging Graph Transformers with Advective Diffusion, by Qitian Wu and 3 other authors
View PDF
HTML (experimental)
Abstract:The capability of generalization is a cornerstone for the success of modern learning systems. For non-Euclidean data, e.g., graphs, that particularly involves topological structures, one important aspect neglected by prior studies is how machine learning models generalize under topological shifts. This paper proposes AdvDIFFormer, a physics-inspired graph Transformer model designed to address this challenge. The model is derived from advective diffusion equations which describe a class of continuous message passing process with observed and latent topological structures. We show that AdvDIFFormer has provable capability for controlling generalization error with topological shifts, which in contrast cannot be guaranteed by graph diffusion models. Empirically, the model demonstrates superiority in various predictive tasks across information networks, molecular screening and protein interactions.
Submission history
From: Qitian Wu [view email]
[v1]
Tue, 10 Oct 2023 08:40:47 UTC (761 KB)
[v2]
Sun, 4 May 2025 05:18:43 UTC (2,456 KB)