This document summarizes Ramezan Paravi's PhD research on estimation and sampling in slow mixing Markov processes. It discusses challenges in estimating transition probabilities and stationary probabilities from finite samples of a slow mixing Markov process before it reaches stationarity. It presents theoretical results showing that transition probabilities of frequently occurring strings can be accurately estimated from the sample. It also shows that stationary probabilities of these strings can be estimated with confidence bounds that are data-dependent and do not require knowledge of the underlying source process. The document concludes by mentioning an algorithm for community detection in slow mixing networks.
1. ESTIMATION AND sampling IN SLOW MIXING
MARKOV PROCESSES
Ramezan Paravi | Ph.D. Candidate
EE Department, UH Manoa
Advisor: Dr. Santhanam
August 2015
2. 2 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview
Markov Sources in slow mixing regime:
Empirical counts do not reflect stationary
Analysis of samples before mixing happens
Part I: Statistical properties of finite samples
Part II: Modeling using slow mixing Markov process
Andrey Markov
A
B C
1
3
2
31
2
3
4
1
4
1
2
3. 3 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
Any structure?
4. 3 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
Any structure?
Reorganizing
5. 3 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
Any structure?
Reorganizing
• Social Networks
• Biological Networks
• Recommender Systems
6. 4 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
1
2
3
4 5
6
7
8
9
10
11
1213
1415
Random walk on graph
7. 4 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
1
2
3
4 5
6
7
8
9
10
11
1213
1415
Uniform random walk.
Explore state space fast.
8. 4 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
1
2
3
4 5
6
7
8
9
10
11
1213
1415
Non-uniform random walk.
Polarized state space.
Walks starting here will be trapped.
9. 5 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
p(1|111)
p(1|011)
p(1|101)
p(1|001)
1
0
1
0
1
0
1
Given sample Y1, . . . ,Yn from unknown
binary Markov source p
Transition probabilities?
Stationary probabilities?
What we do:
fixed sample, best answer
What we are not doing:
e.a.s. results
10. 6 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Complications
Two major sources of difficulties:
Long memory
Slow mixing
May not estimate accurately/completely given n samples
Rather, want best possible answer with sample
11. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
12. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
13. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
14. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q2: If B ∼ i.i.d, what can be said?
15. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q2: If B ∼ i.i.d, what can be said?
Ans: P(1) P(0), (only 3 zeros in the sample)
16. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q2: If B ∼ i.i.d, what can be said?
Ans: P(1) P(0), (only 3 zeros in the sample)
Q3: If A ∼ Markov memory-1, what about transition probs?
17. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q2: If B ∼ i.i.d, what can be said?
Ans: P(1) P(0), (only 3 zeros in the sample)
Q3: If A ∼ Markov memory-1, what about transition probs?
Ans: P(1|1) is high
18. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q4: If B ∼ i.i.d, then what about P(0)?
19. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q4: If B ∼ i.i.d, then what about P(0)?
Ans: More 1’s than 0’s. With high confidence, P(0) small.
20. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q4: If B ∼ i.i.d, then what about P(0)?
Ans: More 1’s than 0’s. With high confidence, P(0) small.
Q5: If A ∼ Markov memory-1, what about P(0)?
21. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q4: If B ∼ i.i.d, then what about P(0)?
Ans: More 1’s than 0’s. With high confidence, P(0) small.
Q5: If A ∼ Markov memory-1, what about P(0)?
Ans: Can not judge with finite sample.
22. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q6: If B ∼ i.i.d and see more bits, then what?
23. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q6: If B ∼ i.i.d and see more bits, then what?
Ans: Likely lots of 1’s, few 0’s. P(0) ≈ #0’s
len is small.
24. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q6: If B ∼ i.i.d and see more bits, then what?
Ans: Likely lots of 1’s, few 0’s. P(0) ≈ #0’s
len is small.
Q7: If A ∼ Markov memory-1 and see more bits, can we say #0’s
len is small?
25. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q6: If B ∼ i.i.d and see more bits, then what?
Ans: Likely lots of 1’s, few 0’s. P(0) ≈ #0’s
len is small.
Q7: If A ∼ Markov memory-1 and see more bits, can we say #0’s
len is small?
Ans: NO. P(0) could be arbitrarily large!!
26. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
27. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
28. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
Subsequence following w is i.i.d
(# wa)
(# w) ≈ P(a|w)
29. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
Subsequence following w is i.i.d
(# wa)
(# w) ≈ P(a|w)
Harder: If memory unknown, but the source has mixed:
30. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
Subsequence following w is i.i.d
(# wa)
(# w) ≈ P(a|w)
Harder: If memory unknown, but the source has mixed:
Both #w and #wa reflect P(w) and P(wa)
Still (# wa)
(# w) ≈ P(a|w)
31. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
Subsequence following w is i.i.d
(# wa)
(# w) ≈ P(a|w)
Harder: If memory unknown, but the source has mixed:
Both #w and #wa reflect P(w) and P(wa)
Still (# wa)
(# w) ≈ P(a|w)
Difficult: If memory unknown and the source has not mixed:
32. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
Subsequence following w is i.i.d
(# wa)
(# w) ≈ P(a|w)
Harder: If memory unknown, but the source has mixed:
Both #w and #wa reflect P(w) and P(wa)
Still (# wa)
(# w) ≈ P(a|w)
Difficult: If memory unknown and the source has not mixed:
Non trivial
33. 9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Given a length-n sample from binary model with dying dependencies
Amount of information two bits provide about each other,
conditioned on middle, diminishes farther they are.
34. 9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Given a length-n sample from binary model with dying dependencies
Amount of information two bits provide about each other,
conditioned on middle, diminishes farther they are.
Identify from data which parameters can be estimated
Set ˜G of “good” strings w of length Θ(log n)
Only those frequently occurred in the sample
35. 9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Given a length-n sample from binary model with dying dependencies
Amount of information two bits provide about each other,
conditioned on middle, diminishes farther they are.
Identify from data which parameters can be estimated
Set ˜G of “good” strings w of length Θ(log n)
Only those frequently occurred in the sample
Provide (confidence and) accuracy bounds
36. 9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Given a length-n sample from binary model with dying dependencies
Amount of information two bits provide about each other,
conditioned on middle, diminishes farther they are.
Identify from data which parameters can be estimated
Set ˜G of “good” strings w of length Θ(log n)
Only those frequently occurred in the sample
Provide (confidence and) accuracy bounds
Transition probabilities of w ∈ ˜G
Universal Compression + combinatorial arguments
(# wa)
(# w) ≈ P(a|w)
37. 10 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Surprises:
!! even if (#wa) or (#w) ≈ stationary
! bound may not hold for shorter substrings of w
Stationary Transition
38. 11 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Provide (confidence and) accuracy bounds
39. 11 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Provide (confidence and) accuracy bounds
Stationary probabilities of w ∈ ˜G
Doob Martingale + Coupling Markov chains + Azuma
#w
˜n ≈ P(w)
P( ˜G)
for w ∈ ˜G,
˜n = number of times strings in ˜G appear in the sample
40. 12 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
All bounds are entirely data dependent
Confidence/accuracy obtained w.h.p. from sample
Knowledge of the source is not required
41. 13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Modeling using slow mixing
Provide randomized algorithm for community detection
42. 13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Modeling using slow mixing
Provide randomized algorithm for community detection
43. 13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Modeling using slow mixing
Provide randomized algorithm for community detection
Build slow mixing random walks on graphs
44. 13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Modeling using slow mixing
Provide randomized algorithm for community detection
Build slow mixing random walks on graphs
Mixing properties of the walk → reveal community structure
Framework: Coupling From the Past
45. 13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Modeling using slow mixing
Provide randomized algorithm for community detection
Build slow mixing random walks on graphs
Mixing properties of the walk → reveal community structure
Framework: Coupling From the Past
Simulation results on benchmark networks
46. 14 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Part I:
Estimation in Slow Mixing
Markov Processes
47. 15 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Arbitrary Mass
All strings in a finite sample, can have arbitrarily small mass.
1 −
m
1
0
p(0) = m
m+1
p(1) = 1
m+1
p(1) can be arbitrarily small for m large
enough.
If o(1/n), starting from 1, see a sequence
of O(1/ ) 1’s whp.
p(any seq. of 1’s) ≤ 1
m+1, can be arbitrarily
small.
49. 17 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Long Memory
1/2
1/2
1/2
1/2
1 −
2
1/2
1/2
-------------
- - - -
- - - -
- - - -
- - - -
depth k
↑1
↓0
If k ω(log n), starting from 1 we
will not see a sequence of k − 1
consecutive 0s whp in a sample of
size n.
50. 17 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Long Memory
1/2
1/2
1/2
1/2
1 −
2
1/2
1/2
-------------
- - - -
- - - -
- - - -
- - - -
depth k
↑1
↓0
If k ω(log n), starting from 1 we
will not see a sequence of k − 1
consecutive 0s whp in a sample of
size n.
Therefore, all bits generated wp 1
2.
Cannot distinguish from i.i.d.
Bernoulli (1/2) whp
51. 18 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Slow Mixing
p(0) = p(1) = 1
2
1 − 1
0
p (1) = 2
3, p (0) = 1
3
1 −
2
1
0
Starting from 1, whp both generate sequence of O(1/ ) 1s. If o(1/n), whp
cannot distinguish using length-n sample
52. 18 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Slow Mixing
p(0) = p(1) = 1
2
1 − 1
0
p (1) = 2
3, p (0) = 1
3
1 −
2
1
0
Starting from 1, whp both generate sequence of O(1/ ) 1s. If o(1/n), whp
cannot distinguish using length-n sample
Caution!
Slow mixing only hurts estimation, not compression!
good compression for memory-1 sources, slow mixing or not
53. 19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Prior Work
Any consistent estimator converges pointwise (NOT uniformly) over the class
of stationary and ergodic Markov models.
Extensive work on
54. 19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Prior Work
Any consistent estimator converges pointwise (NOT uniformly) over the class
of stationary and ergodic Markov models.
Extensive work on consistency,
55. 19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Prior Work
Any consistent estimator converges pointwise (NOT uniformly) over the class
of stationary and ergodic Markov models.
Extensive work on consistency, e.a.s. results,
56. 19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Prior Work
Any consistent estimator converges pointwise (NOT uniformly) over the class
of stationary and ergodic Markov models.
Extensive work on consistency, e.a.s. results, finite-sample but model
dependent results
[Buhlmann, Csiszár, alves, Gariviér, Leonardi, Marton, Maum-Deschamps,
Morvai, Rissanen, Schmitt,Shields, Talata, Weiss, Wyner]
Our Philosophy
Can we look at a length-n sample and identify what, if anything, can be
estimated accurately?
57. 20 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Dependencies
Problem futile if the dependencies are arbitrary.
We assume dependencies die down.
d(4)
Formalize with d : N → R+.
Siblings s, s (nodes of same color) satisfy
p(1|s)
p(1|s )
− 1 ≤ d(4).
Md = {srcs satisfying above for all siblings}.
58. 21 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Dependencies Die Down
Information-theoretic interpretation
I(Y0; Yi+1|Y i
1 ) ≤ log(1 + d(i)).
b1 b2
b2b1
Not related to mixing properties of the source.
No bound on memory of the source.
Need d(i) summable over i, equivalently δj = i≥j d(i) → 0
59. 22 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Aggregation
Unknown set of states T
kn = Θ(log n)
60. 22 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Aggregation
Unknown set of states T
Consider a coarser model (aggregation)
Ask p(1|w), where |w| = kn
kn = Θ(log n)
61. 22 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Aggregation
Unknown set of states T
Consider a coarser model (aggregation)
Ask p(1|w), where |w| = kn
Set kn = Θ(log n)
Makes sense to ask length-3 aggregations for
memory-2 source (source itself)
kn = Θ(log n)
62. 23 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Goal
Unknown source p in Md
For any length-kn string w (known kn = Θ(log n))
• Estimate transition p(·|w)
• Estimate stationary p(w)
kn = Θ(log n)
63. 23 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Goal
Unknown source p in Md
For any length-kn string w (known kn = Θ(log n))
• Estimate transition p(·|w)
• Estimate stationary p(w)
p
p
Unkown Src in Md
Space of memory-kn Srcs
Caution
The problem is not the same as estimating a memory-kn source.
We never see samples from the aggregated model.
64. 24 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Naive Estimators
Example
Suppose Y 0
−∞ = · · · 00 and Y n
1 = 11010010011.
Depth-2 aggregated parameters, e,g., transition prb from w = 10.
· · · 00, 110 10 10 10 0
Subsequence following 10 is 1110
65. 24 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Naive Estimators
Example
Suppose Y 0
−∞ = · · · 00 and Y n
1 = 11010010011.
Depth-2 aggregated parameters, e,g., transition prb from w = 10.
· · · 00, 110 10 10 10 0
Subsequence following 10 is 1110
This is not iid in general because sample is from true model, not aggregated model!
66. 24 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Naive Estimators
Example
Suppose Y 0
−∞ = · · · 00 and Y n
1 = 11010010011.
Depth-2 aggregated parameters, e,g., transition prb from w = 10.
· · · 00, 110 10 10 10 0
Subsequence following 10 is 1110
This is not iid in general because sample is from true model, not aggregated model!
Naive (“1110 i.i.d.”): ˆp(1|w) = 3
4 and ˆp(0|w) = 1
4.
67. 25 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Deviation Bounds for Conditional Probabilities
Theorem
Confidence ≥ 1 − 1
22kn+1 log n
(conditioned on any past Y 0
−∞), for all w ∈ {0, 1}kn
simultaneously
(#w·)
(#w)
− p(·|w)
1
≤ 2
(ln 2)(2kn+1 log n + nδkn )
Nw
.
Again, δkn = i≥kn
d(i).
The more occurrence, the stronger bound.
If d(i) decreases exponentially as γi and Nw = nβ, rhs diminishes as
O( 1√
nβ−1−γ log γ
).
68. 26 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Proof Idea
The result is built on following facts:
• Source belongs to Md, i.e., dependencies die down.
• Compression result on Md reminiscent of method of types.
• Arguments relating strong compression results to variational distance between
estimators.
69. 27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Surprises
Proof order
Stationary Transition
70. 27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Surprises
Proof order
Stationary Transition
Bound may hold for w
! even without empirical frequencies ≈ stationary probabilities
71. 27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Surprises
Proof order
Stationary Transition
Bound may hold for w
! even without empirical frequencies ≈ stationary probabilities
!! but not for its suffixes
72. 27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Surprises
Proof order
Stationary Transition
Bound may hold for w
! even without empirical frequencies ≈ stationary probabilities
!! but not for its suffixes
Possible p(1|100 zeros), but p(1|0)
73. 28 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Good States
Good “states” ˜G are length-kn strings appearing frequently enough
˜G = w : count(w) ≥ max{nδkn log
1
δkn
, 2kn+1
log2
n}
Concentration bound is at least as fast as 1√
log n
.
If d(i) decreases exponentially as γi, concentration bound is poly in n.
74. 29 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Stationary Probabilities
Sensitive function of conditional probabilities
How interpret counts of w ∈ ˜G in the sample?
75. 30 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Deviation Bounds for Stationary Probabilities
Theorem
For any t 0, Y 0
−∞ and w ∈ ˜G,
P
(#w)
˜n
−
p(w)
p( ˜G)
≥ t|Y 0
−∞ ≤ 2 exp −
(˜nt − B)2
2˜nB2
where ˜n is the sum count of states in ˜G, B depends on n and how quickly
dependencies d(i) die off.
76. 30 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Deviation Bounds for Stationary Probabilities
Theorem
For any t 0, Y 0
−∞ and w ∈ ˜G,
P
(#w)
˜n
−
p(w)
p( ˜G)
≥ t|Y 0
−∞ ≤ 2 exp −
(˜nt − B)2
2˜nB2
where ˜n is the sum count of states in ˜G, B depends on n and how quickly
dependencies d(i) die off.
B = O(log n) if d(i) = γi
For bound to be non-vacuously true, need d to be “twice summable”, or
δi = j≥i d(j) to be summable
77. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
↑1
↓0
78. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
↑1
↓0
˜G = {01, 10}
1
0
79. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
1 0
↑
τ0
↑1
↓0
w = 10 1
1
110
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Z0
1 1 0
80. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
0 1
↑
τ1
↑1
↓0
w = 01
0
0
001
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Z1
0 0 1
81. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
1 0
↑
τ2
↑1
↓0
w = 10 1
1
110
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Z2
1 1 0
82. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
0 1
↑
τ3
↑1
↓0
w = 01
0
0
001
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Z3
0 0 1
83. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
0 1
↑
τ˜n
↑1
↓0
w = 01
0
0
001
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Z˜n
0 0 1
84. 32 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Proof Outline
1 Construct a natural Doob martingale
Vm = E(# 01|Z0, Z1, · · · , Zm), m = 0, · · · , ˜n
85. 32 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Proof Outline
1 Construct a natural Doob martingale
Vm = E(# 01|Z0, Z1, · · · , Zm), m = 0, · · · , ˜n
2 Bound |Vm − Vm−1| using Coupling argument
86. 32 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Proof Outline
1 Construct a natural Doob martingale
Vm = E(# 01|Z0, Z1, · · · , Zm), m = 0, · · · , ˜n
2 Bound |Vm − Vm−1| using Coupling argument
3 Azuma’s inequality closes the bound
90. 33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling Technique
X0
Y0
X1
Y1
X2
Y2
X3
Y3
X4
Y4
· · ·
· · ·
Xn−1
Yn−1
Xn
Yn
Jointly evolve according to ω
91. 33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling Technique
X0
Y0
X1
Y1
X2
Y2
X3
Y3
X4
Y4
· · ·
· · ·
Xn−1
Yn−1
Xn
Yn
Jointly evolve according to ω
Encourage evolution st Xt = Yt for all t after (random) τ steps
P(τ i) = ω(Xi = Yi)
E(τ) =
i≥1
P(τ i) =
i≥1
ω(Xi=Yi)
Sample size Eτ ⇒ empirical ≈ stationary [Aldous]
92. 34 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Proof Outline (contd)
• Run two coupled copies Zj and Zj of the restricted process
• Long and unknown memory ⇒ They do not coalesce the usual way
• Instead, approximate coalescence ⇒ Longer together, harder to separate out
• Reason: Dependencies die down
|Vm − Vm−1| ≤
n
j=m+1
ω(Zj≈Zj )
93. 35 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Part II:
Sampling From Slow Mixing
Markov Processes
94. 36 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling From the Past (CFTP)
Markov Chain Over S
Stationary Distribution is π
Goal: Sample from exactly π(.)
Idea (Propp and Wilson):
Expose states to shared source of randomness
Simulate chains backward in time
Wait until all chains merge
A
B C
1
3
2
3
1
2
3
4
1
4
1
2
π(A) = 9
19, π(B) = 4
19, π(C) = 6
19
95. 37 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling
A
B C
1
3
2
3
1
2
3
4
1
4
1
2
A B C
A
C
B
A
A
B
[0, 1]
1
3
96. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
A B C
C
B
A
C
B
A
97. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
A B C
C
B
A
C
B
A
C
B
A
98. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
A B C
C
B
A
C
B
A
C
B
A
C
B
A
99. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
A B C
C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
100. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
A B C
C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
101. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
C
Coalescence
Output ∼ π(.)C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
106. 41 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Approaches
Spectral Clustering
• Eigenvectors of Laplacian of similarity graph
Simple implementation
Laplacian could be ill-conditioned
# of clusters need to be known in advance
[Ng, Jordan, White, Smyth, Weiss, Shi, Malik, Kannan, Vempala, Vetta, Meila, · · · ]
Semi-Definite Programming (SDP)
• LP relaxation
Asymptotically optimal for Stochastic Block Models
Implicit assumption on generative model
# of clusters need to be known in advance
[Abbe, Sandon, Hajeck, Bandeira, Hall, Decelle, Mossel, Neeman, Sly, Rao, Chen, Wu, Xu, · · · ]
107. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Similarity graph.
108. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Similarity graph.
Define random walk.
109. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Non-uniform
Probability of following a link ∝ # of common neighbors
Similarity graph.
Define random walk.
Couple random walks
starting from different nodes
110. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Similarity graph.
Define random walk.
Couple random walks
starting from different nodes
Adapt CFTP algorithm
111. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Similarity graph.
Define random walk.
Couple random walks
starting from different nodes
Adapt CFTP algorithm
Do not care about exact
sample
112. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Similarity graph.
Define random walk.
Couple random walks
starting from different nodes
Adapt CFTP algorithm
Do not care about exact
sample
Identify clusters “before
coalescence happens”
113. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Restricted walk within a cluster mixes faster.
Similarity graph.
Define random walk.
Couple random walks
starting from different nodes
Adapt CFTP algorithm
Do not care about exact
sample
Identify clusters “before
coalescence happens”
114. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
115. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Initially: All singletons
116. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Partial coalescence
Small Clusters formation
117. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Critical times
Clusters merge
118. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Critical times
Clusters merge
119. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Critical times
Clusters merge
120. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Full coalescence
121. 44 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Partial Coalescence
time
S
G
Gc
-10 -9 -8 -7 -6 -5 -4 -3 -2 -1 0
Paths starting from G ⊂ S
122. 45 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Remarks
Algorithm can stop at critical times
Still yield a cluster
Choose cluster with optimal cost
Purely based on random walk
Use of mixing properties
Circumvent issues with ill-conditioned matrices in spectral based approaches
No prior assumption on generative model
No prior assumption on number of clusters
123. 46 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Benchmark Networks
Stochastic Block Models (SBM)
Size of communities equal
Average degree equal for all nodes
LFR Models
More realistic
# of communities and sizes admit power law.
Real world networks
124. 47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Stochastic Block Model
125. 47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Stochastic Block Model
Ber(p)
126. 47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Stochastic Block Model
Ber(p)
Ber(q)
127. 47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Stochastic Block Model
Ber(p)
Ber(q)
p q
134. 53 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
American College Football
115 teams.
Divided into conferences.
137. 56 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Future Directions
Open: Theoretical guarantees for recovery in SBMs
Open: Stationary is more difficult (needs twice summability of d) than
transition (just summability of d)
Theoretical foundation of the proposed algorithm
Extension to broader set of community detection problems
139. 58 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Acknowledgement
My Advisor: Prasad
Committee Members
Dr. Rui Zhang
Office Mates:
• Maryam and Meysam
My Friends:
• Masoud, Saeed, Navid, Elyas, Harir Chee, Reza, Ehsaneh, Ali, Ashkan, Hamed,
Ehsan, Seyed, Alireza,...
My Family:
• My parents, my sister Nasrin and my brother Naser