Expressive GNNs should map subtrees to the node embeddings injectively.
Key observation: Subtrees of the same depth can be recursively characterized from the leaf nodes to the root nodes.
If each step of GNN’s aggregation can fully retain the neighboring information, the generated node embeddings can distinguish different rooted subtrees.
In other words, most expressive GNN would use an injective neighbor aggregation function at each step.
Computational graph = Rooted subtree
Key observation: Expressive power of GNNs can be characterized by that of neighbor aggregation functions they use.
A more expressive aggregation function leads to a more expressive a GNN.
Element-wise mean pooling + Linear + ReLU non-linearity
GCN’s aggregation function cannot distinguish different multi-sets with the same color proportion. - Theorem [Xu et al. ICLR 2019] → not injective
GraphSAGE (max-pool), Hamilton et al., NeurIPS 2017
MLP + element-wise max-pooling
GraphSAGE’s aggregation function cannot distinguish different multi-sets with the same set of distinct colors. - Theorem [Xu et al. ICLR 2019] → not injective
f produces one-hot encodings of colors. Summation of the one-hot encodings retains all the information about the input multi-set.
Universal Approximation Theorem [Hornik et al., 1989]
1-hidden-layer MLP with sufficiently-large hidden dimensionality and appropriate non-linearity σ(⋅)(including ReLU and sigmoid) can approximate any continuous function to an arbitrary accuracy.
MLPΦ(∑x∈SMLPf(x))
the full model of GIN by relating it to WL graph kernel (traditional way of obtaining graph-level features).
Any injective function over the tuple, (c(k)(v),{c(k)(u)}u∈N(v))