Exchangeability of GNN Representations with Applications to Graph Retrieval
Kartik Nair, Indradyumna Roy, Soumen Chakrabarti, Anirban Dasgupta, Abir De
It shows that graph representations are exchangeable random variables which can help in LSH in graphs
Abstract
In this work, we discover a probabilistic symmetry, called as exchangeability in graph neural networks (GNNs). Specifically, we show that the trained node embedding computed using a large family of graph neural networks, learned under standard optimization tools, are exchangeable random variables. This implies that the probability density of the node embeddings remains invariant with respect to a permutation applied on their dimension axis. This results in identical distribution across the elements of the graph representations. Such a property enables approximation of transportation-based graph similarities by Euclidean similarities between order statistics. Leveraging this reduction, we propose a unified locality-sensitive hashing (LSH) framework that supports diverse relevance measures, including subgraph matching and graph edit distance. Experiments show that our method helps to do LSH more effectively than baselines.
Graph embeddings exhibit exchangeability property, enabling efficient graph retrieval via transport-based similarity approximation with locality-sensitive hashing.
- Discovery of exchangeability property in trained GNN node embeddings across broad family of architectures
- Concentration bound for reducing transport problems on node embeddings
- GRAPHHASH framework for approximate graph retrieval using general transport-based distances
- exchangeability analysis
- locality-sensitive hashing
- order statistics
- transport distances
Authors did not state explicit limitations.
Explore other consequences of exchangeability on learning and training dynamics
from the paperExtend framework to similarities over richer classes of similarity functions for 3D molecular structures
from the paper
Author keywords
- GNN
- Locality sensitive hashing
Related orals
One for Two: A Unified Framework for Imbalanced Graph Classification via Dynamic Balanced Prototype
Unified framework for imbalanced graph classification using dynamic balanced prototypes and prototype load-balancing optimization.
Compactness and Consistency: A Conjoint Framework for Deep Graph Clustering
CoCo framework captures compactness and consistency in graph neural network representations for improved deep graph clustering.
Multi-Domain Riemannian Graph Gluing for Building Graph Foundation Models
GraphGlue uses Riemannian geometry to merge multi-domain graphs into unified manifolds, enabling knowledge transfer across graph domains.
Learning with Dual-level Noisy Correspondence for Multi-modal Entity Alignment
Proposes framework to handle noisy entity-attribute and inter-graph correspondences in multi-modal entity alignment.
Scaling Laws and Spectra of Shallow Neural Networks in the Feature Learning Regime
Analyzes scaling laws for shallow networks with feature learning via sparse estimation and matrix compression theory.