随机算法 (Spring 2013)/Expander Graphs and Mixing: Difference between revisions

From TCS Wiki
Jump to navigation Jump to search
imported>Etone
No edit summary
imported>Etone
No edit summary
 
Line 461: Line 461:
}}
}}
}}
}}
= Spectral approach for symmetric chain =
We consider the '''symmetric Markov chains''' defined on the state space <math>\Omega</math>, where the transition matrix <math>P</math> is symmetric.
We have the following powerful spectral theorem for symmetric matrices.
{{Theorem|Theorem (Spectral theorem)|
:Let <math>S</math> be a symmetric <math>n\times n</math> matrix, whose eigenvalues are <math>\lambda_1\ge\lambda_2\ge\cdots\ge\lambda_n</math>. There exist eigenvectors <math>v_1,v_2,\ldots,v_n</math> such that <math>v_iS=\lambda_iv_i</math> for all <math>i=1,\ldots, n</math> and <math>v_1,v_2,\ldots,v_n</math> are '''orthonormal'''.
}}
A set of orthonormal vectors <math>v_1,v_2,\ldots,v_n</math> satisfy that
* for any <math>i\neq j</math>, <math>v_i</math> is orthogonal to <math>v_j</math>, which means that <math>v_i^Tv_j=0</math>, denoted <math>v_i\bot v_j</math>;
* all <math>v_i</math> have unit length, i.e., <math>\|v_i\|_2=1</math>.
Since the eigenvectors <math>v_1,v_2,\ldots,v_n</math> are orthonormal, we can use them as orthogonal basis, so that any vector <math>x\in\mathbb{R}^n</math> can be expressed as <math>x=\sum_{i=1}^nc_iv_i</math> where <math>c_i=q^Tv_i</math>, therefore
:<math>xS=\sum_{i=1}^nc_iv_iS=\sum_{i=1}^nc_i\lambda_iv_i.</math>
So multiplying by <math>S</math> corresponds to multiplying the length of <math>x</math> along the direction of every eigenvector by a factor of the corresponding eigenvalue.
-----
Back to the symmetric Markov chain. Let <math>\Omega</math> be a finite state space of size <math>N</math>, and <math>P</math> be a symmetric transition matrix, whose eigenvalues are <math>\lambda_1\ge\lambda_2\ge\ldots\ge \lambda_N</math>. The followings hold for a symmetric transition matrix <math>P</math>:
*Due to the spectral theorem,  there exist orthonormal eigenvectors <math>v_1,v_2,\ldots,v_N</math> such that <math>v_iP=\lambda_iv_i</math> for <math>i=1,2,\ldots,N</math> and any distribution <math>q</math> over <math>\Omega</math> can be expressed as <math>q=\sum_{i=1}^Nc_iv_i</math> where <math>c_i=q^Tv_i</math>.
*A symmetric <math>P</math> must be double stochastic, thus the stationary distribution <math>\pi</math> is the uniform distribution.
Recall that due to Perron-Frobenius theorem, <math>\lambda_1=1</math>. And <math>\boldsymbol{1}P=\boldsymbol{1}</math> since <math>P</math> is double stochastic, thus <math>v_1=\frac{\boldsymbol{1}}{\|\boldsymbol{1}\|_2}=\left(\frac{1}{\sqrt{N}},\ldots,\frac{1}{\sqrt{N}}\right)</math>.
When <math>q</math> is a distribution, i.e., <math>q</math> is a nonnegative vector and <math>\|q\|_1=1</math>, it holds that <math>c_1=q^Tv_1=\frac{1}{\sqrt{N}}</math> and
<math>c_1v_1=\left(\frac{1}{N},\ldots,\frac{1}{N}\right)=\pi</math>, thus
:<math>
q=\sum_{i=1}^Nc_iv_i=\pi+\sum_{i=2}^Nc_iv_i,
</math>
and the distribution at time <math>t</math> when the initial distribution is <math>q</math>, is given by
:<math>
qP^t=\pi P^t+\sum_{i=2}^Nc_iv_iP^t=\pi+\sum_{i=2}^Nc_i\lambda_i^tv_i.
</math>
It is easy to see that this distribution converges to <math>\pi</math> when the absolute values of <math>\lambda_2,\ldots,\lambda_N</math> are all less than 1. And the rate at which it converges to <math>\pi</math>, namely the mixing rate, is determined by the quantity <math>\lambda_{\max}=\max\{|\lambda_2|,|\lambda_N|\}\,</math>, which is the largest absolute eigenvalues other than <math>\lambda_1</math>.
{{Theorem|Theorem|
:Let <math>P</math> be the transition matrix for a symmetric Markov chain on state space <math>\Omega</math> where <math>|\Omega|=N</math>. Let <math>\lambda_1\ge\lambda_2\ge\cdots\ge\lambda_N</math> be the spectrum of <math>P</math> and <math>\lambda_{\max}=\max\{|\lambda_2|,|\lambda_N|\}\,</math>. The mixing rate of the Markov chain is
::<math>\tau(\epsilon)\le\frac{\frac{1}{2}\ln N+\ln\frac{1}{2\epsilon}}{1-\lambda_{\text{max}}}</math>.
}}
{{Proof|
As analysed above, if <math>P</math> is symmetric, it has orthonormal eigenvectors <math>v_1,\ldots,v_N</math> such that any distribution <math>q</math> over <math>\Omega</math> can be expressed as
:<math>q=\sum_{i=1}^Nc_iv_i=\pi+\sum_{i=2}^Nc_iv_i</math>
with <math>c_i=q^Tv_i</math>, and
:<math>
qP^t=\pi+\sum_{i=2}^Nc_i\lambda_i^tv_i.
</math>
Thus,
:<math>
\begin{align}
\|qP^t-\pi\|_1
&=
\left\|\sum_{i=2}^Nc_i\lambda_i^tv_i\right\|_1\
&\le
\sqrt{N}\left\|\sum_{i=2}^Nc_i\lambda_i^tv_i\right\|_2 &\quad\text{(Cauchy-Schwarz)}\
&=
\sqrt{N}\sqrt{\sum_{i=2}^Nc_i^2\lambda_i^{2t}}\
&\le
\sqrt{N}\lambda_{\max}^t\sqrt{\sum_{i=2}^Nc_i^2}\
&=
\sqrt{N}\lambda_{\max}^t\|q\|_2\
&\le
\sqrt{N}\lambda_{\max}^t.
\end{align}
</math>
The last inequality is due to a universal relation <math>\|q\|_2\le\|q\|_1</math> and the fact that <math>q</math> is a distribution.
Then for any <math>x\in\Omega</math>, denoted by <math>\boldsymbol{1}_x</math> the indicator vector for <math>x</math> such that <math>\boldsymbol{1}_x(x)=1</math> and <math>\boldsymbol{1}_x(y)=0</math> for <math>y\neq x</math>, we have
:<math>
\begin{align}
\Delta_x(t)
&=\left\|\boldsymbol{1}_x P^t-\pi\right\|_{TV}=\frac{1}{2}\left\|\boldsymbol{1}_x P^t-\pi\right\|_1\
&\le\frac{\sqrt{N}}{2}\lambda_{\max}^t\le\frac{\sqrt{N}}{2}\mathrm{e}^{-t(1-\lambda_{\max})}.
\end{align}
</math>
Therefore, we have
:<math>
\tau_x(\epsilon)
=\min\{t\mid\Delta_x(t)\le\epsilon\}
\le\frac{\frac{1}{2}\ln N+\ln\frac{1}{2\epsilon}}{1-\lambda_{\max}}
</math>
for any <math>x\in\Omega</math>, thus the bound holds for <math>\tau(\epsilon)=\max_{x}\tau_x(\epsilon)</math>.
}}
== Rapid mixing of expander walk ==
Let <math>G(V,E)</math> be a <math>d</math>-regular graph on <math>n</math> vertices. Let <math>A</math> be its adjacency matrix. The transition matrix of the lazy random walk on <math>G</math> is given by <math>P=\frac{1}{2}\left(I+\frac{1}{d}A\right)</math>. Specifically,
:<math>P(u,v)=\begin{cases}
\frac{1}{2} & \text{if }u=v,\
\frac{1}{2d} & \text{if }uv\in E,\
0 & \text{otherwise.}
\end{cases}</math>
Obviously <math>P</math> is symmetric.
Let <math>\lambda_1\ge\lambda_2\ge\cdots\ge\lambda_n</math> be the eigenvalues of <math>A</math>, and <math>\nu_1\ge\nu_2\ge\cdots\ge\nu_n</math> be the eigenvalues of <math>P</math>. It is easy to verify that
:<math>\nu_i=\frac{1}{2}\left(1+\frac{\lambda_i}{d}\right)</math>.
We know that <math>-d\le\lambda_i\le d</math>, thus <math>0\le\nu_i\le1</math>. Therefore, <math>\nu_{\max}=\max\{|\nu_2|,|\nu_n|\}=\nu_2\,</math>. Due to the above analysis of symmetric Markov chain,
:<math>\tau(\epsilon)\le\frac{\frac{1}{2}(\ln n+\ln\frac{1}{2\epsilon})}{1-\nu_{\max}}=\frac{\frac{1}{2}(\ln n+\ln\frac{1}{2\epsilon})}{1-\nu_2}=\frac{d(\ln n+\ln\frac{1}{2\epsilon})}{d-\lambda_2}</math>.
Thus we prove the following theorem for lazy random walk on <math>d</math>-regular graphs.
{{Theorem|Theorem|
:Let <math>G(V,E)</math> be a <math>d</math>-regular graph on <math>n</math> vertices, with spectrum <math>\lambda_1\ge\lambda_2\ge\cdots\ge\lambda_n</math>. The mixing rate of lazy random walk on <math>G</math> is
::<math>\tau(\epsilon)\le\frac{d(\ln n+\ln\frac{1}{2\epsilon})}{d-\lambda_2}</math>.
}}
Due to Cheeger's inequality, the spectral gap is bounded by expansion ratio as
:<math>d-\lambda_2\ge\frac{\phi^2}{2d},</math>
where <math>\phi=\phi(G)</math> is the expansion ratio of the graph <math>G</math>. Therefore, we have the following corollary which bounds the mixing time by graph expansion.
{{Theorem|Corollary|
:Let <math>G(V,E)</math> be a <math>d</math>-regular graph on <math>n</math> vertices, whose expansion ratio is <math>\phi=\phi(G)</math>. The mixing rate of lazy random walk on <math>G</math> is
::<math>\tau(\epsilon)\le\frac{2d^2(\ln n+\ln\frac{1}{2\epsilon})}{\phi^2}</math>.
:In particular, the mixing time is
::<math>\tau_{\text{mix}}=\tau(1/2\mathrm{e})\le\frac{2d^2(\ln n+2)}{\phi^2}</math>.
}}
For expander graphs, both <math>d</math> and <math>\phi</math> are constants. The mixing time of lazy random walk is <math>\tau_{\text{mix}}=O(\ln n)</math> so the random walk is rapidly mixing.

Latest revision as of 15:28, 8 June 2013

Expander Graphs

According to wikipedia:

"Expander graphs have found extensive applications in computer science, in designing algorithms, error correcting codes, extractors, pseudorandom generators, sorting networks and robust computer networks. They have also been used in proofs of many important results in computational complexity theory, such as SL=L and the PCP theorem. In cryptography too, expander graphs are used to construct hash functions."

We will not explore everything about expander graphs, but will focus on the performances of random walks on expander graphs.

Consider an undirected (multi-)graph G(V,E), where the parallel edges between two vertices are allowed.

Some notations:

  • For S,TV, let E(S,T)={uvEuS,vT}.
  • The Edge Boundary of a set SV, denoted S, is S=E(S,S¯).
Definition (Graph expansion)
The expansion ratio of an undirected graph G on n vertices, is defined as
ϕ(G)=min|S|n2SV|S||S|.

Expander graphs are d-regular (multi)graphs with d=O(1) and ϕ(G)=Ω(1).

This definition states the following properties of expander graphs:

  • Expander graphs are sparse graphs. This is because the number of edges is dn/2=O(n).
  • Despite the sparsity, expander graphs have good connectivity. This is supported by the expansion ratio.
  • This one is implicit: expander graph is a family of graphs {Gn}, where n is the number of vertices. The asymptotic order O(1) and Ω(1) in the definition is relative to the number of vertices n, which grows to infinity.

The following fact is directly implied by the definition.

An expander graph has diameter O(logn).

The proof is left for an exercise.

For a vertex set S, the size of the edge boundary |S| can be seen as the "perimeter" of S, and |S| can be seen as the "volume" of S. The expansion property can be interpreted as a combinatorial version of isoperimetric inequality.

Vertex expansion
We can alternatively define the vertex expansion. For a vertex set SV, its vertex boundary, denoted δS is defined as that
δS={uSuvE and vS},
and the vertex expansion of a graph G is ψ(G)=min|S|n2SV|δS||S|.

Existence of expander graph

We will show the existence of expander graphs by the probabilistic method. In order to do so, we need to generate random d-regular graphs.

Suppose that d is even. We can generate a random d-regular graph G(V,E) as follows:

  • Let V be the vertex set. Uniformly and independently choose d2 cycles of V.
  • For each vertex v, for every cycle, assuming that the two neighbors of v in that cycle is w and u, add two edges wv and uv to E.

The resulting G(V,E) is a multigraph. That is, it may have multiple edges between two vertices. We will show that G(V,E) is an expander graph with high probability. Formally, for some constant d and constant α,

Pr[ϕ(G)α]=1o(1).

By the probabilistic method, this shows that there exist expander graphs. In fact, the above probability bound shows something much stronger: it shows that almost every regular graph is an expander.

Recall that ϕ(G)=minS:|S|n2|S||S|. We call such SV that |S||S|<α a "bad S". Then ϕ(G)<α if and only if there exists a bad S of size at most n2. Therefore,

Pr[ϕ(G)<α]=Pr[minS:|S|n2|S||S|<α]=k=1n2Pr[bad S of size k]k=1n2S(Vk)Pr[S is bad]

Let RS be the set of vertices in S which has neighbors in S¯, and let r=|R|. It is obvious that |S|r, thus, for a bad S, r<αk. Therefore, there are at most r=1αk(kr) possible choices such R. For any fixed choice of R, the probability that an edge picked by a vertex in SR connects to a vertex in S is at most k/n, and there are d(kr) such edges. For any fixed S of size k and R of size r, the probability that all neighbors of all vertices in SR are in S is at most (kn)d(kr). Due to the union bound, for any fixed S of size k,

Pr[S is bad]r=1αk(kr)(kn)d(kr)αk(kαk)(kn)dk(1α)

Therefore,

Pr[ϕ(G)<α]k=1n2S(Vk)Pr[S is bad]k=1n2(nk)αk(kαk)(kn)dk(1α)k=1n2(enk)kαk(ekαk)αk(kn)dk(1α)(Stirling formula (nk)(enk)k)k=1n2exp(O(k))(kn)k(d(1α)1).

The last line is o(1) when d21α. Therefore, G is an expander graph with expansion ratio α with high probability for suitable choices of constant d and constant α.

Computing graph expansion

Computation of graph expansion seems hard, because the definition involves the minimum over exponentially many subsets of vertices. In fact, the problem of deciding whether a graph is an expander is co-NP-complete. For a non-expander G, the vertex set SV which has low expansion ratio is a proof of the fact that G is not an expander, which can be verified in poly-time. However, there is no efficient algorithm for computing the ϕ(G) unless NP=P.

The expansion ratio of a graph is closely related to the sparsest cut of the graph, which is the dual problem of the multicommodity flow problem, both NP-complete. Studies of these two problems revolutionized the area of approximation algorithms.

We will see right now that although it is hard to compute the expansion ratio exactly, the expansion ratio can be approximated by some efficiently computable algebraic identity of the graph.

Spectral Graph Theory

Graph spectrum

The adjacency matrix of an n-vertex graph G, denoted A=A(G), is an n×n matrix where A(u,v) is the number of edges in G between vertex u and vertex v. Because A is a symmetric matrix with real entries, due to the Perron-Frobenius theorem, it has real eigenvalues λ1λ2λn, which associate with an orthonormal system of eigenvectors v1,v2,,vn with Avi=λivi. We call the eigenvalues of A the spectrum of the graph G.

The spectrum of a graph contains a lot of information about the graph. For example, supposed that G is d-regular, the following lemma holds.

Lemma
  1. |λi|d for all 1in.
  2. λ1=d and the corresponding eigenvector is (1n,1n,,1n).
  3. G is connected if and only if λ1>λ2.
  4. If G is bipartite then λ1=λn.
Proof.
Let A be the adjacency matrix of G, with entries aij. It is obvious that jaij=d for any j.
  • (1) Suppose that Ax=λx,x0, and let xi be an entry of x with the largest absolute value. Since (Ax)i=λxi, we have
jaijxj=λxi,
and so
|λ||xi|=|jaijxj|jaij|xj|jaij|xi|d|xi|.
Thus |λ|d.
  • (2) is easy to check.
  • (3) Let x be the nonzero vector for which Ax=dx, and let xi be an entry of x with the largest absolute value. Since (Ax)i=dxi, we have
jaijxj=dxi.
Since jaij=d and by the maximality of xi, it follows that xj=xi for all j that aij>0. Thus, xi=xj if i and j are adjacent, which implies that xi=xj if i and j are connected. For connected G, all vertices are connected, thus all xi are equal. This shows that if G is connected, the eigenvalue d=λ1 has multiplicity 1, thus λ1>λ2.
If otherwise, G is disconnected, then for two different components, we have Ax=dx and Ay=dy, where the entries of x and y are nonzero only for the vertices in their components components. Then A(αx+βy)=d(αx+βy). Thus, the multiplicity of d is greater than 1, so λ1=λ2.
  • (4) If G if bipartite, then the vertex set can be partitioned into two disjoint nonempty sets V1 and V2 such that all edges have one endpoint in each of V1 and V2. Algebraically, this means that the adjacency matrix can be organized into the form
PTAP=[0BBT0]
where P is a permutation matrix, which has no change on the eigenvalues.
If x is an eigenvector corresponding to the eigenvalue λ, then x which is obtained from x by changing the sign of the entries corresponding to vertices in V2, is an eigenvector corresponding to the eigenvalue λ. It follows that the spectrum of a bipartite graph is symmetric with respect to 0.

Cheeger's Inequality

One of the most exciting results in spectral graph theory is the following theorem which relate the graph expansion to the spectral gap.

Theorem (Cheeger's inequality)
Let G be a d-regular graph with spectrum λ1λ2λn. Then
dλ22ϕ(G)2d(dλ2).

The theorem was first stated for Riemannian manifolds, and was proved by Cheeger and Buser (for different directions of the inequalities). The discrete case is proved independently by Dodziuk and Alon-Milman.

For a d-regular graph, the quantity (dλ2) is called the spectral gap. The name is due to the fact that it is the gap between the first and the second largest eigenvalues of a graph.

If we write α=1λ2d (sometimes it is called the normalized spectral gap), the Cheeger's inequality is turned into a nicer form:

α2ϕd2α or equivalently 12(ϕd)2α2(ϕd).

Optimization Characterization of Eigenvalues

Theorem (Rayleigh-Ritz theorem)
Let A be a symmetric n×n matrix. Let λ1λ2λn be the eigen values of A and v1,v2,,vn be the corresponding eigenvectors. Then
λ1=maxxRnxTAxxTx and λ2=maxxv1xTAxxTx.
Proof.

Without loss of generality, we may assume that v1,v2,,vn are orthonormal eigen-basis. Then it holds that

v1TAv1v1Tv1=λ1v1Tv1=λ1,

thus we have maxxRnxTAxxTxλ1.

Let xRn be an arbitrary vector and let y=xxTx=xx be its normalization. Since v1,v2,,vn are orthonormal basis, y can be expressed as y=i=1ncivi. Then

xTAxxTx=yTAy=(i=1ncivi)TA(i=1ncivi)=(i=1ncivi)T(i=1nλicivi)=i=1nλici2λ1i=1nci2=λ1y=λ1.

Therefore, maxxRnxTAxxTxλ1. Altogether we have maxxRnxTAxxTx=λ1

It is similar to prove maxxv1xTAxxTx=λ2. In the first part take x=v2 to show that maxxv1xTAxxTxλ2; and in the second part take an arbitrary xv1 and y=xx. Notice that yv1, thus y=i=1ncivi with c1=0.

The Rayleigh-Ritz Theorem is a special case of a fundamental theorem in linear algebra, called the Courant-Fischer theorem, which characterizes the eigenvalues of a symmetric matrix by a series of optimizations:

Theorem (Courant-Fischer theorem)
Let A be a symmetric matrix with eigenvalues λ1λ2λn. Then
λk=maxv1,v2,,vnkRnminxv1,v2,,vnkxRn,x0xTAxxTx=minv1,v2,,vk1Rnmaxxv1,v2,,vk1xRn,x0xTAxxTx.

Graph Laplacian

Let G(V,E) be a d-regular graph of n vertices and let A be its adjacency matrix. We define L=dIA to be the Laplacian of the graph G. Take xRV as a distribution over vertices, its Laplacian quadratic form xTLx measures the "smoothness" of x over the graph topology, just as what the Laplacian operator does to the differentiable functions.

Laplacian Property
For any vector xRn, it holds that
xTLx=uvE(xuxv)2.
Proof.
xTLx=u,vVxu(dIA)uvxv=u(dxu2uvExuxv)=uVuvE(xu2xuxv).

On the other hand,

uvE(xuxv)2=uvE(xu22xuxv+xv2)=uvE((xu2xuxv)+(xv2xvxu))=uVuvE(xu2xuxv).

Applying the Rayleigh-Ritz theorem to the Laplacian matrix of the graph, we have the following "variational characterization" of the spectral gap dλ2.

Theorem (Variational Characterization)
Let G(V,E) be a d-regular graph of n vertices. Suppose that its adjacency matrix is A, whose eigenvalues are λ1λ2λn. Let L=dIA be the Laplacian matrix. Then
dλ2=minx1xTLxxTx=minx1uvE(xuxv)2vVxv2.
Proof.

For d-regular graph, we know that λ1=d and 1A=d1, thus 1 is the eigenvector of λ1. Due to Rayleigh-Ritz Theorem, it holds that λ2=maxx1xTAxxTx. Then

minx1xTLxxTx=minx1xT(dIA)xxTx=minx1dxTxxTAxxTx=minx1(dxTAxxTx)=dmaxx1xTAxxTx=dλ2.

We know it holds for the graph Laplacian that xTLx=uvE(xuxv)2. So the variational characterization of the second eigenvalue of graph is proved.

Proof of Cheeger's Inequality

We will first give an informal explanation why Cheeger's inequality holds.

Recall that the expansion is defined as

ϕ(G)=min|S|n2SV|S||S|.

Let χS be the characteristic vector of the set S such that

χS(v)={1vS,0vS.

It is easy to see that

uvE(χS(u)χS(v))2vVχS(v)2=|S||S|.

Thus, the expansion can be expressed algebraically as

ϕ(G)=min|S|n2SVuvE(χS(u)χS(v))2vVχS(v)2=minx1n2x{0,1}nuvE(xuxv)2vVxv2.

On the other hand, due to the variational characterization of the spectral gap, we have

dλ2=minx1uvE(xuxv)2vVxv2.

We can easily observe the similarity between the two formulas. Both the expansion ration ϕ(G) and the spectral gap dλ2 can be characterized by optimizations of the same objective function uvE(xuxv)2vVxv2 over different domains (for the spectral gap, the optimization is over all x1; and for the expansion ratio, it is over all such vectors x{0,1}n with at most n/2 many 1-entries).


Notations

Throughout the proof, we assume that G(V,E) is the d-regular graph of n vertices, A is the adjacency matrix, whose eigenvalues are λ1λ2λn, and L=(dIA) is the graph Laplacian.

Large spectral gap implies high expansion

Cheeger's inequality (lower bound)
ϕ(G)dλ22.
Proof.

Let SV, |S|n2, be the vertex set achieving the optimal expansion ratio ϕ(G)=min|S|n2SV|S||S|=|S||S|, and xRn be a vector defined as

xv={1/|S|if vS,1/|S¯|if vS¯.

Clearly, x1=vS1|S|vS¯1|S¯|=0, thus x1.

Due to the variational characterization of the second eigenvalue,

dλ2uvE(xuxv)2vVxv2=uS,vS¯,uvE(1/|S|+1/|S¯|)21/|S|+1/|S¯|=(1|S|+1|S¯|)|S|2|S||S|(since |S|n2)=2ϕ(G).

High expansion implies large spectral gap

We next prove the upper bound direction of the Cheeger's inequality:

Cheeger's inequality (upper bound)
ϕ(G)2d(dλ2).

This direction is harder than the lower bound direction. But it is mathematically more interesting and also more useful to us for analyzing the mixing time of random walks.

We prove the following equivalent inequality:

ϕ22ddλ2.

Let x satisfy that

  • Ax=λ2x, i.e., it is a eigenvector for λ2;
  • |{vVxv>0}|n2, i.e., x has at most n/2 positive entries. (We can always choose x to be x if this is not satisfied.)

And let nonnegative vector y be defined as

yv={xvxv>0,0otherwise.

We then prove the following inequalities:

  1. yTLyyTydλ2;
  2. ϕ22dyTLyyTy.

The theorem is then a simple consequence by combining these two inequalities.

We prove the first inequality:

Lemma
yTLyyTydλ2.
Proof.

If xu0, then

(Ly)u=((dIA)y)u=dyuvA(u,v)yv=dxuv:xv0A(u,v)xvdxuvA(u,v)xv=((dIA)x)u=(dλ2)xu.

Then

yTLy=uyu(Ly)u=u:xu0yu(Ly)u=u:xu0xu(Ly)u(dλ2)u:xu0xu2=(dλ2)uyu2=(dλ2)yTy,

which proves the lemma.

We then prove the second inequality:

Lemma
ϕ22dyTLyyTy.
Proof.

To prove this, we introduce a new quantity uvE|yu2yv2|yTy and shows that

ϕuvE|yu2yv2|yTy2dyTLyyTy.

This will give us the desired inequality ϕ22dyTLyyTy.

Lemma
uvE|yu2yv2|yTy2dyTLyyTy.
Proof.

By the Cauchy-Schwarz Inequality,

uvE|yu2yv2|=uvE|yuyv||yu+yv|uvE(yuyv)2uvE(yu+yv)2.

By the Laplacian property, the first term uvE(yuyv)2=yTLy. By the Inequality of Arithmetic and Geometric Means, the second term

uvE(yu+yv)22uvE(yu2+yv2)=2duVyu2=2dyTy.

Combining them together, we have

uvE|yu2yv2|2dyTLyyTy.
Lemma
ϕuvE|yu2yv2|yTy.
Proof.

Suppose that y has t nonzero entries. We know that tn/2 due to the definition of y. We enumerate the vertices u1,u2,,unV such that

yu1yu2yut>yut+1==yun=0.

Then

uvE|yu2yv2|=uiujEi<j(yui2yuj2)=uiujEi<jk=ij1(yuk2yuk+12)=i=1nj>iA(ui,uj)k=ij1(yuk2yuk+12)=i=1nj>ik=ij1A(ui,uj)(yuk2yuk+12).

We have the following universal equation for sums:

i=1nj>ik=ij1A(ui,uj)(yuk2yuk+12)=k=1nikj>kA(ui,uj)(yuk2yuk+12)=k=1t(yuk2yuk+12)ikj>kA(ui,uj)

Notice that ikj>kA(ui,uj)=|{u1,,uk}|, which is at most ϕk since ktn/2. Therefore, combining these together, we have

uvE|yu2yv2|=k=1t(yuk2yuk+12)ikj>kA(ui,uj)=k=1t(yuk2yuk+12)|{u1,,uk}|ϕk=1t(yuk2yuk+12)k=ϕk=1tyuk2=ϕyTy.

Spectral approach for symmetric chain

We consider the symmetric Markov chains defined on the state space Ω, where the transition matrix P is symmetric.

We have the following powerful spectral theorem for symmetric matrices.

Theorem (Spectral theorem)
Let S be a symmetric n×n matrix, whose eigenvalues are λ1λ2λn. There exist eigenvectors v1,v2,,vn such that viS=λivi for all i=1,,n and v1,v2,,vn are orthonormal.

A set of orthonormal vectors v1,v2,,vn satisfy that

  • for any ij, vi is orthogonal to vj, which means that viTvj=0, denoted vivj;
  • all vi have unit length, i.e., vi2=1.

Since the eigenvectors v1,v2,,vn are orthonormal, we can use them as orthogonal basis, so that any vector xRn can be expressed as x=i=1ncivi where ci=qTvi, therefore

xS=i=1nciviS=i=1nciλivi.

So multiplying by S corresponds to multiplying the length of x along the direction of every eigenvector by a factor of the corresponding eigenvalue.


Back to the symmetric Markov chain. Let Ω be a finite state space of size N, and P be a symmetric transition matrix, whose eigenvalues are λ1λ2λN. The followings hold for a symmetric transition matrix P:

  • Due to the spectral theorem, there exist orthonormal eigenvectors v1,v2,,vN such that viP=λivi for i=1,2,,N and any distribution q over Ω can be expressed as q=i=1Ncivi where ci=qTvi.
  • A symmetric P must be double stochastic, thus the stationary distribution π is the uniform distribution.

Recall that due to Perron-Frobenius theorem, λ1=1. And 1P=1 since P is double stochastic, thus v1=112=(1N,,1N).

When q is a distribution, i.e., q is a nonnegative vector and q1=1, it holds that c1=qTv1=1N and c1v1=(1N,,1N)=π, thus

q=i=1Ncivi=π+i=2Ncivi,

and the distribution at time t when the initial distribution is q, is given by

qPt=πPt+i=2NciviPt=π+i=2Nciλitvi.

It is easy to see that this distribution converges to π when the absolute values of λ2,,λN are all less than 1. And the rate at which it converges to π, namely the mixing rate, is determined by the quantity λmax=max{|λ2|,|λN|}, which is the largest absolute eigenvalues other than λ1.

Theorem
Let P be the transition matrix for a symmetric Markov chain on state space Ω where |Ω|=N. Let λ1λ2λN be the spectrum of P and λmax=max{|λ2|,|λN|}. The mixing rate of the Markov chain is
τ(ϵ)12lnN+ln12ϵ1λmax.
Proof.

As analysed above, if P is symmetric, it has orthonormal eigenvectors v1,,vN such that any distribution q over Ω can be expressed as

q=i=1Ncivi=π+i=2Ncivi

with ci=qTvi, and

qPt=π+i=2Nciλitvi.

Thus,

qPtπ1=i=2Nciλitvi1Ni=2Nciλitvi2(Cauchy-Schwarz)=Ni=2Nci2λi2tNλmaxti=2Nci2=Nλmaxtq2Nλmaxt.

The last inequality is due to a universal relation q2q1 and the fact that q is a distribution.

Then for any xΩ, denoted by 1x the indicator vector for x such that 1x(x)=1 and 1x(y)=0 for yx, we have

Δx(t)=1xPtπTV=121xPtπ1N2λmaxtN2et(1λmax).

Therefore, we have

τx(ϵ)=min{tΔx(t)ϵ}12lnN+ln12ϵ1λmax

for any xΩ, thus the bound holds for τ(ϵ)=maxxτx(ϵ).

Rapid mixing of expander walk

Let G(V,E) be a d-regular graph on n vertices. Let A be its adjacency matrix. The transition matrix of the lazy random walk on G is given by P=12(I+1dA). Specifically,

P(u,v)={12if u=v,12dif uvE,0otherwise.

Obviously P is symmetric.

Let λ1λ2λn be the eigenvalues of A, and ν1ν2νn be the eigenvalues of P. It is easy to verify that

νi=12(1+λid).

We know that dλid, thus 0νi1. Therefore, νmax=max{|ν2|,|νn|}=ν2. Due to the above analysis of symmetric Markov chain,

τ(ϵ)12(lnn+ln12ϵ)1νmax=12(lnn+ln12ϵ)1ν2=d(lnn+ln12ϵ)dλ2.

Thus we prove the following theorem for lazy random walk on d-regular graphs.

Theorem
Let G(V,E) be a d-regular graph on n vertices, with spectrum λ1λ2λn. The mixing rate of lazy random walk on G is
τ(ϵ)d(lnn+ln12ϵ)dλ2.

Due to Cheeger's inequality, the spectral gap is bounded by expansion ratio as

dλ2ϕ22d,

where ϕ=ϕ(G) is the expansion ratio of the graph G. Therefore, we have the following corollary which bounds the mixing time by graph expansion.

Corollary
Let G(V,E) be a d-regular graph on n vertices, whose expansion ratio is ϕ=ϕ(G). The mixing rate of lazy random walk on G is
τ(ϵ)2d2(lnn+ln12ϵ)ϕ2.
In particular, the mixing time is
τmix=τ(1/2e)2d2(lnn+2)ϕ2.

For expander graphs, both d and ϕ are constants. The mixing time of lazy random walk is τmix=O(lnn) so the random walk is rapidly mixing.