f(⋅): a neural network like differentiable function λ: a weight decay coefficient X: a matrix of node feature vectors Δ=D−A: the unnormalized graph Laplacian of an undirected graph. G=(V,E): undirected graph vi∈N: Nodes (vi,vj)∈E: edges A∈RN×N: an adjacency matrix (binary or weighted) Dii=∑jAij: a degree matrix
The normalized graph Laplacian, L=IN−D−21AD−21=UΛU⊤
gθ⋆x=UgθU⊤x
x,θ: scalar values gθ=diag(θ): a filter; assumed an eigenvalues Λ of L, i.e. gθ(Λ) U: the matrix of eigenvectors of the normalized graph Laplacian Λ: a diagonal matrix of its eigenvalues U⊤x: the graph Fourier transform of x.
A multiplicatioin with the eigenvector matrix U is O(N2). Thus, using the eigendecomposition of L for computational efficiency.
To circumvent this problem, it was suggested in Hammond et al. (2011) that gθ(Λ) can be well-approximated by a truncated expansion in terms of Chebyshev polynomialsTk(x) up to Kth order:
gθ′(Λ)≈∑k=0Kθk′Tk(Λ~)
Λ~=λmax2Λ−IN λmax: the largest eigenvalue of L θ′∈RK: a vector of Chebyshev coefficients
gθ′⋆x≈∑k=0Kθk′Tk(L~)x
L~=λmax2L−IN L~ can easily be verified by noticing that (UΛU⊤)k=UΛkU⊤.
This expression is now K-localized since it is a Kth-order polynomial in the Laplacian, i.e. it depends only on nodes that are at maximum K steps away from the central node (Kth-order neighborhood). (O(∣E∣), the linear time complexity)
Chebyshev polynomials를 활용하여, 이차에서 일차로 시간 복잡도가 감소해 computational efficiency를 유지할 수 있다. 또한 a flexibility of the explicit parameterization 유지할 수 있다.
gθ′⋆x≈θ0′x+θ1′(L−IN)x=θ0′x−θ1′D−21AD−21x
The filter parameters, θ0′,θ1′, can be shared over the whole graph. Filters effectively convolve the Kth-order neighborhood of a node. It can be beneficial to constrain the number of parameters further to address overfitting and to minimize the number of operations (such as matrix multiplications) per layer.
Filter parameters를 하나로 보면, θ=θ0′=−θ1′, 아래와 같다.
gθ′⋆x≈θ(IN+D−21AD−21)x
위의 식을 따라 층이 깊어지면 IN으로 인해, 일부 node들은 exploding이 발생하고, 그외 node들은 상대적으로 vanishing을 겪을 수 있다. 이것을 개선하고자, 저자들은 Identity matrix도 normalize하는, re-normalization trick을 도입하였다.
Identity matrix도 normalize 함으로, 전체적으로 보다 안정한 structure flow를 가져올 수 있게 된다.
Z=D~−21A~D~−21XΘ
X∈RN×C: a signal, C is the dimension of features of a input Θ∈RC×F: filter parameters, F is the dimension of filter parameters Z∈RN×F: a convolved signal (activation)
This equation’s time complexity is O(∣E∣FC).
structure data인 adjacency matrix를 활용한다.
If using the re-normalized trick of the GCN paper, as follow:
The Moore-Penrose inverse, also known as the pseudoinverse, is a generalization of the matrix inverse that can be applied to non-square matrices and singular matrices.
The Moore-Penrose inverse is defined as a unique matrix that satisfies four properties:
In the context of graph theory, the Laplacian matrix is a square matrix that captures the topology of a graph.
First, it encodes the structure of the graph in a compact, matrix form that can be easily manipulated using linear algebra.
For example, the Laplacian matrix can be used to compute various graph properties, such as the number of connected components, the size of the largest connected component, and the graph conductance.
Second, the Laplacian matrix can be used as a tool for graph-based data analysis and machine learning.
For example, it can be used to define graph Laplacian regularization terms in optimization problems, which encourage smoothness of functions defined on the graph. It can also be used to define graph Laplacian eigenmaps, which are low-dimensional embeddings of the nodes in the graph that preserve the graph structure.
The Laplacian matrix is a real and symmetric n x n matrix, and it has n nonnegative eigenvalues, denoted by λ1,λ2,...,λn. The eigenvectors of the Laplacian matrix are n linearly independent vectors v1,v2,...,vn, which satisfy the eigenvalue equation Lvi=λivi for i=1,2,...,n.
In mathematics, the spectrum of a graph refers to the set of eigenvalues of its adjacency matrix, Laplacian matrix, or other related matrices. The eigenvalues of these matrices provide information about the structure and properties of the graph, such as its connectivity, number of connected components, and expansion properties.
The spectrum of a graph is a fundamental concept in graph theory and has important applications in various fields, including computer science, physics, and social network analysis. For example, the spectral properties of a graph can be used to analyze its robustness to attacks, the diffusion of information or influence through the network, and the performance of graph-based machine learning algorithms.
It is a symmetric and positive semidefinite matrix, which means that it has real eigenvalues and orthogonal eigenvectors.
The eigenvalues of the symmetric normalized Laplacian matrix can be used to analyze the spectral properties of the graph.
The eigenvectors can be used for dimensionality reduction and clustering.
The degree matrix D is a diagonal matrix with the degree of each node on the diagonal. However, it can be viewed as a special case of the Laplacian matrix, where the adjacency matrix is the identity matrix. In this case, the Laplacian matrix is L=D−I, and the eigenvalues and eigenvectors of L coincide with those of D.
The adjacency matrix A is a matrix that encodes the topology of the graph.
Lsym:=(D+)1/2L(D+)1/2=I−(D+)1/2A(D+)1/2
Li,jsym:=⎩⎨⎧1−deg(vi)deg(vj)10if i=j anddeg(vi)=0if i=j and vi is adjacenct to vjotherwise.