Graph properties from eigenvalues of adjacency/Laplacian matrices.
Multi-session curriculum - substantial prior knowledge and complex material. Use mastery gates and deliberate practice.
Graphs look discrete—just nodes and edges—but many of their most important properties become visible only when you translate the graph into a matrix and study its eigenvalues. Spectral graph theory is the toolkit that makes that translation precise: connectivity, expansion, random-walk mixing, and good graph partitions all leave signatures in the spectrum of the adjacency matrix and (especially) the graph Laplacian.
Encode a graph with matrices like the adjacency matrix A and Laplacian L = D − A (or normalized variants). Because these matrices are symmetric for undirected graphs, their eigenvalues are real and their eigenvectors form an orthonormal basis. The smallest Laplacian eigenvalues describe connected components and “how connected” the graph is (algebraic connectivity), while specific eigenvectors (Fiedler vector) guide spectral partitioning. For expansion, the normalized Laplacian spectrum relates to conductance via Cheeger-type inequalities. Random walks connect through the random-walk Laplacian and Markov chain stationary distributions.
This node assumes you already know basic eigenvalues/eigenvectors () and graph representations. For spectral graph theory, a few extra prerequisites matter a lot; this section makes them explicit and flags where each appears.
1) Symmetric eigendecomposition (for real symmetric matrices)
2) Quadratic forms and positive semidefinite (PSD) matrices
3) Rayleigh quotient and the min–max principle
1) Undirected vs directed, unweighted vs weighted graphs
2) Connected components
3) Cuts and volumes
1) Markov chains on graphs: transition matrix
2) Stationary distribution
If any of these are shaky, it’s worth a quick review before continuing—spectral graph theory is powerful, but it’s also picky about definitions (especially which Laplacian and which conductance).
Spectral graph theory studies graphs through the eigenvalues and eigenvectors (the spectrum) of matrices that encode graph structure.
At a high level, the workflow is:
1) Start with a graph (possibly weighted).
2) Build a matrix representation like the adjacency matrix or a Laplacian.
3) Analyze eigenvalues/eigenvectors of that matrix.
4) Translate spectral facts back into graph properties: connectivity, clustering, expansion, random-walk behavior, and more.
Why this is even plausible: eigenvectors provide “global coordinates” on the vertices. If a graph has two clusters with sparse connections between them, there tends to exist a vector x that is roughly constant on each cluster but differs between clusters. When you apply a graph matrix to x, the result changes little if edges mostly connect equal values. This idea becomes a quadratic form: edges penalize differences.
For a graph with vertices (labeled 1…n),
For undirected graphs, is symmetric.
Adjacency spectra are great for: regular graphs, counting walks, and global “density/structure” signals. But for cuts, flows, and random walks, Laplacians are often more natural.
is diagonal with (weighted degree if weighted).
This is the central object for many connectivity and cut problems.
There are two common Laplacians beyond :
1) Symmetric normalized Laplacian
This is symmetric (for undirected graphs with positive degrees).
2) Random-walk Laplacian
where is the random-walk transition matrix. is generally not symmetric, but it is similar to and thus shares eigenvalues.
The spectrum of a matrix is the multiset of its eigenvalues:
Eigenvectors matter too, especially for algorithms: the eigenvector associated with (or ) often reveals a good partition.
A key mindset shift: spectral information is not just a “summary.” Many graph optimization problems (like finding a minimum cut under balance constraints) are hard combinatorial problems. Spectral methods relax them into continuous problems solvable by eigenvectors.
The combinatorial Laplacian is more than a definition—it encodes a geometry on the graph.
Take any vector x ∈ ℝⁿ assigning a scalar to each vertex . Consider the quadratic form:
Expand it step by step (undirected, weighted case; assume ):
\begin{align*}
\mathbf{x}^\top L\mathbf{x}
&= \mathbf{x}^\top (D-A)\mathbf{x}\\
&= \mathbf{x}^\top D\mathbf{x} - \mathbf{x}^\top A\mathbf{x}\\
&= \sum_i d_i x_i^2 - \sum_{i,j} w_{ij} x_i x_j.
\end{align*}
Now rewrite the second term using symmetry:
\begin{align*}
\sum_{i,j} w_{ij} x_i x_j
&= \sum_{i<j} w_{ij}(x_i x_j + x_j x_i) + \sum_i w_{ii}x_i^2\\
&= 2\sum_{i<j} w_{ij}x_i x_j \quad (\text{usually } w_{ii}=0).
\end{align*}
Also note:
So
\begin{align*}
\mathbf{x}^\top L\mathbf{x}
&= \sum_{i,j} w_{ij} x_i^2 - \sum_{i,j} w_{ij} x_i x_j\\
&= \frac{1}{2}\sum_{i,j} w_{ij}(x_i^2 - 2x_i x_j + x_j^2)\\
&= \frac{1}{2}\sum_{i,j} w_{ij}(x_i - x_j)^2.
\end{align*}
Interpretation: is the total “edge disagreement energy.” It’s small when neighboring vertices have similar values.
From
we immediately get: is PSD, hence all eigenvalues satisfy .
If x is constant, say for all , then every difference , so
Thus 0 is always an eigenvalue with eigenvector 1.
If the graph has connected components, you can build linearly independent vectors that are constant on one component and 0 elsewhere. Each has zero energy, so each lies in the nullspace of .
Theorem: The multiplicity of eigenvalue 0 of equals the number of connected components.
This is one of the cleanest examples of “graph property ↔ eigenvalues.”
The second-smallest eigenvalue (for a connected graph) is called the algebraic connectivity.
Variationally,
This is where Rayleigh quotient/min–max enters: is the best (smallest) achievable energy among vectors orthogonal to constants.
Combinatorial treats every vertex value equally in the denominator . But for irregular graphs, high-degree vertices can dominate behavior.
The symmetric normalized Laplacian has quadratic form
This makes “differences” comparable relative to degrees. Many expansion and random-walk results (including Cheeger-type inequalities) are stated for .
A practical rule:
A central algorithmic use of spectral graph theory is to find a “good” partition of the vertices: two groups with few edges between them but each group not too small.
For a subset (nonempty, not all vertices), define:
A widely used balanced-separation score is conductance:
And the graph conductance is
This definition is the one that matches the standard Cheeger inequality for the normalized Laplacian.
If you use a raw cut size alone, you can get trivial answers: isolate a single low-degree vertex. Conductance normalizes by volume, which corresponds to probability mass under the random-walk stationary distribution .
Suppose you want a cut. A discrete way is an indicator vector f where
But optimizing over such discrete vectors is hard. Spectral methods relax the problem: allow f to be real-valued, solve a continuous minimization, then “round” back to a set by thresholding.
This is where eigenvectors enter: the minimizing relaxed vector is an eigenvector.
Let be eigenvalues of .
For the conductance definition above and :
Important clarifications:
Let u₂ be an eigenvector for of (or equivalently for the second-largest eigenvalue of ). The spectral partitioning recipe:
1) Compute u₂.
2) Sort vertices by their coordinate in u₂ (or by depending on implementation).
3) Consider prefix sets consisting of the first vertices in that order.
4) Compute conductance for each; choose the best.
This “sweep” is the rounding step that turns a continuous eigenvector into a discrete cut.
If the graph has two clusters weakly connected:
Thresholding separates the two sign/level regions.
Adjacency eigenvectors can also show community structure, especially in stochastic block models. But Laplacians are generally more robust for irregular degrees because they directly encode “difference across edges” and integrate naturally with random-walk normalization.
A useful comparison:
| Goal | Typical matrix | Why |
|---|---|---|
| Connectivity / components | or | Nullspace reveals components |
| Balanced sparse cut / expansion | Cheeger inequality + conductance | |
| Random walk mixing | , , | Markov chain eigenvalues control convergence |
| Regular graphs structure | Counts walks; eigenvalues relate to expansion in regular case |
Takeaway: when you see conductance/Markov chains, think normalized objects.
Spectral graph theory isn’t just about partitions. The same matrices govern diffusion, random walks, and learning algorithms.
For an undirected weighted graph, define
Then is the probability of moving from to in one step.
The stationary distribution is
This is why volumes (sums of degrees) appear in conductance: they’re the natural “mass” under the walk.
Even though is not symmetric, it is similar to a symmetric matrix:
So has real eigenvalues (for undirected graphs) and they match those of .
Also,
So if are eigenvalues of , then eigenvalues of are .
Let be eigenvalues of (for connected graphs, ).
The spectral gap is often (for lazy walks you avoid negative eigenvalues issues). Since , the second normalized Laplacian eigenvalue is exactly that gap.
Intuition:
This is the random-walk mirror of Cheeger: bottlenecks ↔ small ↔ slow mixing.
In continuous-time diffusion on graphs, a common model is
Solution uses the matrix exponential:
Eigenvectors of are the diffusion “modes”:
This perspective explains why low-frequency eigenvectors are used for embedding and clustering: they capture coarse geometry.
You’ll see several related matrices in practice:
| Object | Formula | Notes |
|---|---|---|
| Adjacency | Symmetric for undirected; counts walks via | |
| Combinatorial Laplacian | PSD; components ↔ zero eigenvalues | |
| Symmetric normalized Laplacian | Spectrum in ; Cheeger inequality | |
| Random-walk matrix | Markov chain; eigenvalues relate to mixing | |
| Random-walk Laplacian | Similar to |
Spectral graph theory is a hub: once you can move between graphs ↔ matrices ↔ eigen-objects, many results become variations on the same theme—energy minimization under constraints.
Let G be the unweighted path 1—2—3. Build A, D, L = D − A, then compute eigenvalues to see how connectivity appears spectrally.
Adjacency matrix:
A = [[0,1,0],
[1,0,1],
[0,1,0]]
Degrees: d₁=1, d₂=2, d₃=1 so
D = diag(1,2,1).
Laplacian:
L = D − A = [[ 1,-1, 0],
[-1, 2,-1],
[ 0,-1, 1]].
Check the always-true fact L1=0:
L[1,1,1]ᵀ = [1−1+0, −1+2−1, 0−1+1]ᵀ = [0,0,0]ᵀ.
So λ₁=0.
Compute remaining eigenvalues by characteristic polynomial det(L−λI)=0.
L−λI = [[1−λ, -1, 0],
[ -1, 2−λ, -1],
[ 0, -1, 1−λ]].
Determinant expansion (use the first row):
det(L−λI) = (1−λ) det([[2−λ, -1],[-1, 1−λ]]) − (−1) det([[-1, -1],[0, 1−λ]]).
Compute the 2×2 determinants:
det([[2−λ, -1],[-1, 1−λ]]) = (2−λ)(1−λ) − 1.
det([[-1, -1],[0, 1−λ]]) = (−1)(1−λ) − 0 = −(1−λ).
So:
det(L−λI) = (1−λ)[(2−λ)(1−λ) − 1] + (1)[−(1−λ)]
= (1−λ)[(2−λ)(1−λ) − 2].
Expand (2−λ)(1−λ)=2−3λ+λ².
Then [(2−λ)(1−λ) − 2] = (2−3λ+λ²) − 2 = λ² − 3λ.
Thus det(L−λI) = (1−λ)(λ²−3λ) = (1−λ)λ(λ−3).
Eigenvalues are {0, 1, 3}.
Because the graph is connected, 0 has multiplicity 1 (one component). The second eigenvalue λ₂=1 reflects nontrivial connectivity; if we made the middle edge very weak, λ₂ would drop toward 0.
Insight: Even in a tiny graph, λ₁=0 encodes components, and λ₂ (algebraic connectivity) quantifies how hard it is to separate the graph. The eigenvalues come from an energy perspective: vectors that vary across the single bottleneck edge pay disagreement energy.
You have computed the second eigenvector u₂ of the symmetric normalized Laplacian 𝓛 for an undirected graph. Show how to obtain a cut and how conductance is evaluated during the sweep.
Assume degrees are all positive so 𝓛 is well-defined: 𝓛 = I − D^{-1/2} A D^{-1/2}.
Compute (or are given) the eigenvector u₂ corresponding to ν₂, the second-smallest eigenvalue of 𝓛.
Convert to a degree-aware ordering statistic. A common choice is:
score(i) = u₂,i / √d_i.
(This aligns with the fact that 𝓛’s geometry weights by degree.)
Sort vertices so that score(v₁) ≤ score(v₂) ≤ ... ≤ score(v_n).
For each k = 1,...,n−1 define S_k = {v₁,...,v_k}.
For each S_k compute:
cut(S_k, \bar S_k) = ∑_{i∈S_k, j∉S_k} w_{ij}.
vol(S_k) = ∑_{i∈S_k} d_i.
vol(\bar S_k) = vol(V) − vol(S_k).
Compute conductance:
φ(S_k) = cut(S_k, \bar S_k) / min{vol(S_k), vol(\bar S_k)}.
Return the k with minimum φ(S_k). This is the sweep cut produced by the eigenvector.
Cheeger (normalized Laplacian setting) guarantees: if ν₂ is small, there exists some k whose φ(S_k) is O(√ν₂), more precisely φ(G) ≤ √(2ν₂) and the sweep procedure can achieve comparable bounds.
Insight: The eigenvector gives a 1D embedding where nearby coordinates tend to be densely connected. Sweeping thresholds searches for the best place to ‘cut’ that line, translating a continuous relaxation back to a discrete set with provable guarantees (for 𝓛 and conductance defined via volumes).
Prove two foundational facts for an undirected weighted graph: (1) L is PSD, and (2) the nullspace corresponds to connected components.
Start from L = D − A with symmetric weights w_{ij}=w_{ji}.
For any vector x, expand the quadratic form:
xᵀ L x = xᵀ D x − xᵀ A x.
Rewrite each term:
xᵀ D x = ∑_i d_i x_i² = ∑_{i,j} w_{ij} x_i².
xᵀ A x = ∑_{i,j} w_{ij} x_i x_j.
Combine:
xᵀ L x = ∑_{i,j} w_{ij}(x_i² − x_i x_j)
= (1/2)∑_{i,j} w_{ij}(x_i² − 2x_i x_j + x_j²)
= (1/2)∑_{i,j} w_{ij}(x_i − x_j)² ≥ 0.
Therefore L is PSD and all its eigenvalues are nonnegative.
Now characterize when xᵀ L x = 0. Since it’s a sum of nonnegative terms, it equals 0 iff for every edge (i,j) with w_{ij}>0, we have x_i = x_j.
Thus, on each connected component, x_i must be constant (because along any path values must agree). Different components may have different constants.
So the nullspace consists exactly of vectors that are constant on each connected component. If there are k components, the nullspace dimension is k, meaning eigenvalue 0 has multiplicity k.
Insight: The Laplacian is a discrete smoothness operator. Zero energy means perfectly smooth—no variation across any edge—so the only allowed variation is between disconnected components. This is the clean algebraic reason eigenvalue-0 multiplicity equals the number of components.
Undirected graphs map to symmetric matrices (A, L, 𝓛), so eigenvalues are real and eigenvectors form an orthonormal basis—this is the foundation for spectral reasoning.
The Laplacian quadratic form is an edge disagreement energy: xᵀ L x = (1/2)∑_{i,j} w_{ij}(x_i − x_j)², making L PSD with nonnegative eigenvalues.
Eigenvalue 0 of L always exists (L1=0), and its multiplicity equals the number of connected components.
The second-smallest Laplacian eigenvalue (λ₂ for L or ν₂ for 𝓛) measures how well-connected the graph is; small values signal near-disconnectivity and potential good cuts.
For expansion/balanced cuts you typically want the normalized Laplacian 𝓛 and conductance defined using volumes; this is the setting of the standard Cheeger inequality.
Cheeger inequality (normalized Laplacian): ν₂/2 ≤ φ(G) ≤ √(2ν₂), linking spectral gap to best conductance cut (up to constants).
The second eigenvector (Fiedler vector) enables spectral partitioning: sort vertices by eigenvector coordinate and sweep thresholds to find a low-conductance cut.
Random walks connect through P=D^{-1}A; its eigenvalues relate to 𝓛 via ν_i = 1 − μ_i, tying spectral gaps to mixing behavior.
Mixing Laplacian variants without noticing: statements about conductance/Cheeger usually refer to the normalized Laplacian 𝓛, not the combinatorial L.
Using an ambiguous conductance definition (edge-count vs volume-normalized). Always specify φ(S)=cut(S,\bar S)/min{vol(S),vol(\bar S)} when quoting Cheeger bounds.
Forgetting assumptions needed for symmetry/real spectra: adjacency and Laplacian are symmetric only for undirected graphs with symmetric weights.
Interpreting λ₂ or ν₂ without conditioning on connectedness: if the graph is disconnected, λ₂=0 and you should reason in terms of component structure first.
Let G be two disconnected triangles (two copies of K₃ with no edges between them). What is the multiplicity of eigenvalue 0 of the combinatorial Laplacian L? Explain without computing the full spectrum.
Hint: Use the theorem relating components to the nullspace of L.
The graph has k=2 connected components (each triangle is connected, and there are two of them). The multiplicity of eigenvalue 0 of L equals the number of connected components, so 0 has multiplicity 2.
For an undirected weighted graph, show directly that L = D − A is PSD by deriving xᵀLx = (1/2)∑_{i,j} w_{ij}(x_i − x_j)². State clearly where you use symmetry of weights.
Hint: Expand xᵀ(D−A)x and then symmetrize the double sum.
Expand: xᵀLx = ∑_i d_i x_i² − ∑_{i,j} w_{ij}x_i x_j. Replace d_i with ∑_j w_{ij} to get ∑_{i,j} w_{ij}x_i² − ∑_{i,j} w_{ij}x_i x_j = ∑_{i,j} w_{ij}(x_i² − x_i x_j). Now use symmetry w_{ij}=w_{ji} to average terms (i,j) and (j,i):
∑_{i,j} w_{ij}(x_i² − x_i x_j) = (1/2)∑_{i,j} w_{ij}(x_i² − 2x_i x_j + x_j²) = (1/2)∑_{i,j} w_{ij}(x_i − x_j)² ≥ 0. Hence L is PSD.
You are given ν₂, the second-smallest eigenvalue of the normalized Laplacian 𝓛, equals 0.02. Using Cheeger’s inequality (in the 𝓛 + conductance-by-volume setting), give an interval of possible values for φ(G).
Hint: Use ν₂/2 ≤ φ(G) ≤ √(2ν₂).
Cheeger (normalized Laplacian) gives
Lower bound: φ(G) ≥ ν₂/2 = 0.02/2 = 0.01.
Upper bound: φ(G) ≤ √(2ν₂) = √(0.04) = 0.2.
So 0.01 ≤ φ(G) ≤ 0.2.