SlideShare a Scribd company logo
1 of 139
Download to read offline
ESTIMATION AND sampling IN SLOW MIXING
MARKOV PROCESSES
Ramezan Paravi | Ph.D. Candidate
EE Department, UH Manoa
Advisor: Dr. Santhanam
August 2015
2 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview
Markov Sources in slow mixing regime:
Empirical counts do not reflect stationary
Analysis of samples before mixing happens
Part I: Statistical properties of finite samples
Part II: Modeling using slow mixing Markov process
Andrey Markov
A
B C
1
3
2
31
2
3
4
1
4
1
2
3 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
Any structure?
3 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
Any structure?
Reorganizing
3 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
Any structure?
Reorganizing
• Social Networks
• Biological Networks
• Recommender Systems
4 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
1
2
3
4 5
6
7
8
9
10
11
1213
1415
Random walk on graph
4 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
1
2
3
4 5
6
7
8
9
10
11
1213
1415
Uniform random walk.
Explore state space fast.
4 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Motivation
1
2
3
4 5
6
7
8
9
10
11
1213
1415
Non-uniform random walk.
Polarized state space.
Walks starting here will be trapped.
5 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
p(1|111)
p(1|011)
p(1|101)
p(1|001)
1
0
1
0
1
0
1
Given sample Y1, . . . ,Yn from unknown
binary Markov source p
Transition probabilities?
Stationary probabilities?
 What we do:
fixed sample, best answer
 What we are not doing:
e.a.s. results
6 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Complications
Two major sources of difficulties:
 Long memory
 Slow mixing
May not estimate accurately/completely given n samples
Rather, want best possible answer with sample
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q2: If B ∼ i.i.d, what can be said?
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q2: If B ∼ i.i.d, what can be said?
Ans: P(1) P(0), (only 3 zeros in the sample)
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q2: If B ∼ i.i.d, what can be said?
Ans: P(1) P(0), (only 3 zeros in the sample)
Q3: If A ∼ Markov memory-1, what about transition probs?
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q2: If B ∼ i.i.d, what can be said?
Ans: P(1) P(0), (only 3 zeros in the sample)
Q3: If A ∼ Markov memory-1, what about transition probs?
Ans: P(1|1) is high
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q4: If B ∼ i.i.d, then what about P(0)?
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q4: If B ∼ i.i.d, then what about P(0)?
Ans: More 1’s than 0’s. With high confidence, P(0) small.
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q4: If B ∼ i.i.d, then what about P(0)?
Ans: More 1’s than 0’s. With high confidence, P(0) small.
Q5: If A ∼ Markov memory-1, what about P(0)?
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q4: If B ∼ i.i.d, then what about P(0)?
Ans: More 1’s than 0’s. With high confidence, P(0) small.
Q5: If A ∼ Markov memory-1, what about P(0)?
Ans: Can not judge with finite sample.
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q6: If B ∼ i.i.d and see more bits, then what?
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q6: If B ∼ i.i.d and see more bits, then what?
Ans: Likely lots of 1’s, few 0’s. P(0) ≈ #0’s
len is small.
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q6: If B ∼ i.i.d and see more bits, then what?
Ans: Likely lots of 1’s, few 0’s. P(0) ≈ #0’s
len is small.
Q7: If A ∼ Markov memory-1 and see more bits, can we say #0’s
len is small?
7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Two Samples with same # of 0’s and 1s:
A 11111111111111000 v.s. B 11011110111111011
Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
Q6: If B ∼ i.i.d and see more bits, then what?
Ans: Likely lots of 1’s, few 0’s. P(0) ≈ #0’s
len is small.
Q7: If A ∼ Markov memory-1 and see more bits, can we say #0’s
len is small?
Ans: NO. P(0) could be arbitrarily large!!
8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
Subsequence following w is i.i.d
(# wa)
(# w) ≈ P(a|w)
8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
Subsequence following w is i.i.d
(# wa)
(# w) ≈ P(a|w)
Harder: If memory unknown, but the source has mixed:
8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
Subsequence following w is i.i.d
(# wa)
(# w) ≈ P(a|w)
Harder: If memory unknown, but the source has mixed:
Both #w and #wa reflect P(w) and P(wa)
Still (# wa)
(# w) ≈ P(a|w)
8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
Subsequence following w is i.i.d
(# wa)
(# w) ≈ P(a|w)
Harder: If memory unknown, but the source has mixed:
Both #w and #wa reflect P(w) and P(wa)
Still (# wa)
(# w) ≈ P(a|w)
Difficult: If memory unknown and the source has not mixed:
8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Overview of Theoretical Results
Transition Probabilities:
Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
Easy: If |w| memory :
Subsequence following w is i.i.d
(# wa)
(# w) ≈ P(a|w)
Harder: If memory unknown, but the source has mixed:
Both #w and #wa reflect P(w) and P(wa)
Still (# wa)
(# w) ≈ P(a|w)
Difficult: If memory unknown and the source has not mixed:
 Non trivial
9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Given a length-n sample from binary model with dying dependencies
 Amount of information two bits provide about each other,
conditioned on middle, diminishes farther they are.
9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Given a length-n sample from binary model with dying dependencies
 Amount of information two bits provide about each other,
conditioned on middle, diminishes farther they are.
Identify from data which parameters can be estimated
Set ˜G of “good” strings w of length Θ(log n)
 Only those frequently occurred in the sample
9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Given a length-n sample from binary model with dying dependencies
 Amount of information two bits provide about each other,
conditioned on middle, diminishes farther they are.
Identify from data which parameters can be estimated
Set ˜G of “good” strings w of length Θ(log n)
 Only those frequently occurred in the sample
Provide (confidence and) accuracy bounds
9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Given a length-n sample from binary model with dying dependencies
 Amount of information two bits provide about each other,
conditioned on middle, diminishes farther they are.
Identify from data which parameters can be estimated
Set ˜G of “good” strings w of length Θ(log n)
 Only those frequently occurred in the sample
Provide (confidence and) accuracy bounds
Transition probabilities of w ∈ ˜G
Universal Compression + combinatorial arguments
(# wa)
(# w) ≈ P(a|w)
10 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Surprises:
!! even if (#wa) or (#w) ≈ stationary
! bound may not hold for shorter substrings of w
Stationary Transition
11 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Provide (confidence and) accuracy bounds
11 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
Provide (confidence and) accuracy bounds
Stationary probabilities of w ∈ ˜G
Doob Martingale + Coupling Markov chains + Azuma
#w
˜n ≈ P(w)
P( ˜G)
for w ∈ ˜G,
˜n = number of times strings in ˜G appear in the sample
12 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Results on Estimation
All bounds are entirely data dependent
 Confidence/accuracy obtained w.h.p. from sample
 Knowledge of the source is not required
13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Modeling using slow mixing
Provide randomized algorithm for community detection
13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Modeling using slow mixing
Provide randomized algorithm for community detection
13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Modeling using slow mixing
Provide randomized algorithm for community detection
Build slow mixing random walks on graphs
13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Modeling using slow mixing
Provide randomized algorithm for community detection
Build slow mixing random walks on graphs
Mixing properties of the walk → reveal community structure
Framework: Coupling From the Past
13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Modeling using slow mixing
Provide randomized algorithm for community detection
Build slow mixing random walks on graphs
Mixing properties of the walk → reveal community structure
Framework: Coupling From the Past
Simulation results on benchmark networks
14 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Part I:
Estimation in Slow Mixing
Markov Processes
15 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Arbitrary Mass
All strings in a finite sample, can have arbitrarily small mass.
1 −
m
1
0
p(0) = m
m+1
p(1) = 1
m+1
p(1) can be arbitrarily small for m large
enough.
If o(1/n), starting from 1, see a sequence
of O(1/ ) 1’s whp.
p(any seq. of 1’s) ≤ 1
m+1, can be arbitrarily
small.
16 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Long Memory
1/2
1/2
1/2
1/2
1/2
1/2
1/2
1/2
-------------
- - - -
- - - -
- - - -
- - - -
depth k
↑1
↓0
Bernoulli(1/2) i.i.d. source.
Over parameterized.
17 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Long Memory
1/2
1/2
1/2
1/2
1 −
2
1/2
1/2
-------------
- - - -
- - - -
- - - -
- - - -
depth k
↑1
↓0
If k ω(log n), starting from 1 we
will not see a sequence of k − 1
consecutive 0s whp in a sample of
size n.
17 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Long Memory
1/2
1/2
1/2
1/2
1 −
2
1/2
1/2
-------------
- - - -
- - - -
- - - -
- - - -
depth k
↑1
↓0
If k ω(log n), starting from 1 we
will not see a sequence of k − 1
consecutive 0s whp in a sample of
size n.
Therefore, all bits generated wp 1
2.
Cannot distinguish from i.i.d.
Bernoulli (1/2) whp
18 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Slow Mixing
p(0) = p(1) = 1
2
1 − 1
0
p (1) = 2
3, p (0) = 1
3
1 −
2
1
0
Starting from 1, whp both generate sequence of O(1/ ) 1s. If o(1/n), whp
cannot distinguish using length-n sample
18 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Challenges: Slow Mixing
p(0) = p(1) = 1
2
1 − 1
0
p (1) = 2
3, p (0) = 1
3
1 −
2
1
0
Starting from 1, whp both generate sequence of O(1/ ) 1s. If o(1/n), whp
cannot distinguish using length-n sample
Caution!
Slow mixing only hurts estimation, not compression!
good compression for memory-1 sources, slow mixing or not
19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Prior Work
Any consistent estimator converges pointwise (NOT uniformly) over the class
of stationary and ergodic Markov models.
Extensive work on
19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Prior Work
Any consistent estimator converges pointwise (NOT uniformly) over the class
of stationary and ergodic Markov models.
Extensive work on consistency,
19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Prior Work
Any consistent estimator converges pointwise (NOT uniformly) over the class
of stationary and ergodic Markov models.
Extensive work on consistency, e.a.s. results,
19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Prior Work
Any consistent estimator converges pointwise (NOT uniformly) over the class
of stationary and ergodic Markov models.
Extensive work on consistency, e.a.s. results, finite-sample but model
dependent results
[Buhlmann, Csiszár, alves, Gariviér, Leonardi, Marton, Maum-Deschamps,
Morvai, Rissanen, Schmitt,Shields, Talata, Weiss, Wyner]
Our Philosophy
Can we look at a length-n sample and identify what, if anything, can be
estimated accurately?
20 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Dependencies
Problem futile if the dependencies are arbitrary.
 We assume dependencies die down.
d(4)
Formalize with d : N → R+.
Siblings s, s (nodes of same color) satisfy
p(1|s)
p(1|s )
− 1 ≤ d(4).
Md = {srcs satisfying above for all siblings}.
21 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Dependencies Die Down
Information-theoretic interpretation
I(Y0; Yi+1|Y i
1 ) ≤ log(1 + d(i)).
b1 b2
b2b1
Not related to mixing properties of the source.
No bound on memory of the source.
Need d(i) summable over i, equivalently δj = i≥j d(i) → 0
22 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Aggregation
Unknown set of states T
kn = Θ(log n)
22 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Aggregation
Unknown set of states T
Consider a coarser model (aggregation)
Ask p(1|w), where |w| = kn
kn = Θ(log n)
22 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Aggregation
Unknown set of states T
Consider a coarser model (aggregation)
Ask p(1|w), where |w| = kn
 Set kn = Θ(log n)
 Makes sense to ask length-3 aggregations for
memory-2 source (source itself)
kn = Θ(log n)
23 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Goal
Unknown source p in Md
For any length-kn string w (known kn = Θ(log n))
• Estimate transition p(·|w)
• Estimate stationary p(w)
kn = Θ(log n)
23 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Goal
Unknown source p in Md
For any length-kn string w (known kn = Θ(log n))
• Estimate transition p(·|w)
• Estimate stationary p(w)
p
p
Unkown Src in Md
Space of memory-kn Srcs
Caution
The problem is not the same as estimating a memory-kn source.
We never see samples from the aggregated model.
24 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Naive Estimators
Example
Suppose Y 0
−∞ = · · · 00 and Y n
1 = 11010010011.
Depth-2 aggregated parameters, e,g., transition prb from w = 10.
· · · 00, 110 10 10 10 0
Subsequence following 10 is 1110
24 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Naive Estimators
Example
Suppose Y 0
−∞ = · · · 00 and Y n
1 = 11010010011.
Depth-2 aggregated parameters, e,g., transition prb from w = 10.
· · · 00, 110 10 10 10 0
Subsequence following 10 is 1110
This is not iid in general because sample is from true model, not aggregated model!
24 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Naive Estimators
Example
Suppose Y 0
−∞ = · · · 00 and Y n
1 = 11010010011.
Depth-2 aggregated parameters, e,g., transition prb from w = 10.
· · · 00, 110 10 10 10 0
Subsequence following 10 is 1110
This is not iid in general because sample is from true model, not aggregated model!
Naive (“1110 i.i.d.”): ˆp(1|w) = 3
4 and ˆp(0|w) = 1
4.
25 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Deviation Bounds for Conditional Probabilities
Theorem
Confidence ≥ 1 − 1
22kn+1 log n
(conditioned on any past Y 0
−∞), for all w ∈ {0, 1}kn
simultaneously
(#w·)
(#w)
− p(·|w)
1
≤ 2
(ln 2)(2kn+1 log n + nδkn )
Nw
.
Again, δkn = i≥kn
d(i).
The more occurrence, the stronger bound.
If d(i) decreases exponentially as γi and Nw = nβ, rhs diminishes as
O( 1√
nβ−1−γ log γ
).
26 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Proof Idea
 The result is built on following facts:
• Source belongs to Md, i.e., dependencies die down.
• Compression result on Md reminiscent of method of types.
• Arguments relating strong compression results to variational distance between
estimators.
27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Surprises
Proof order
Stationary Transition
27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Surprises
Proof order
Stationary Transition

Bound may hold for w
! even without empirical frequencies ≈ stationary probabilities
27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Surprises
Proof order
Stationary Transition

Bound may hold for w
! even without empirical frequencies ≈ stationary probabilities
!! but not for its suffixes
27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Surprises
Proof order
Stationary Transition

Bound may hold for w
! even without empirical frequencies ≈ stationary probabilities
!! but not for its suffixes
Possible  p(1|100 zeros), but  p(1|0)
28 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Good States
Good “states” ˜G are length-kn strings appearing frequently enough
˜G = w : count(w) ≥ max{nδkn log
1
δkn
, 2kn+1
log2
n}
Concentration bound is at least as fast as 1√
log n
.
If d(i) decreases exponentially as γi, concentration bound is poly in n.
29 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Stationary Probabilities
Sensitive function of conditional probabilities
How interpret counts of w ∈ ˜G in the sample?
30 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Deviation Bounds for Stationary Probabilities
Theorem
For any t  0, Y 0
−∞ and w ∈ ˜G,
P
(#w)
˜n
−
p(w)
p( ˜G)
≥ t|Y 0
−∞ ≤ 2 exp −
(˜nt − B)2
2˜nB2
where ˜n is the sum count of states in ˜G, B depends on n and how quickly
dependencies d(i) die off.
30 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Deviation Bounds for Stationary Probabilities
Theorem
For any t  0, Y 0
−∞ and w ∈ ˜G,
P
(#w)
˜n
−
p(w)
p( ˜G)
≥ t|Y 0
−∞ ≤ 2 exp −
(˜nt − B)2
2˜nB2
where ˜n is the sum count of states in ˜G, B depends on n and how quickly
dependencies d(i) die off.
B = O(log n) if d(i) = γi
For bound to be non-vacuously true, need d to be “twice summable”, or
δi = j≥i d(j) to be summable
31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
 Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
↑1
↓0
31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
 Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
↑1
↓0
˜G = {01, 10}
1
0
31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
 Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
1 0
↑
τ0
↑1
↓0
w = 10 1
1
110
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Z0
1 1 0
31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
 Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
0 1
↑
τ1
↑1
↓0
w = 01
0
0
001
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Z1
0 0 1
31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
 Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
1 0
↑
τ2
↑1
↓0
w = 10 1
1
110
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Z2
1 1 0
31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
 Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
0 1
↑
τ3
↑1
↓0
w = 01
0
0
001
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Z3
0 0 1
31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Estimation Along a Sequence of Stopping Times
 Restriction of the process to Good states.
Y 0
−∞ Y1 Y2
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Yn
0 1
↑
τ˜n
↑1
↓0
w = 01
0
0
001
0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1
Z˜n
0 0 1
32 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Proof Outline
1 Construct a natural Doob martingale
Vm = E(# 01|Z0, Z1, · · · , Zm), m = 0, · · · , ˜n
32 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Proof Outline
1 Construct a natural Doob martingale
Vm = E(# 01|Z0, Z1, · · · , Zm), m = 0, · · · , ˜n
2 Bound |Vm − Vm−1| using Coupling argument
32 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Proof Outline
1 Construct a natural Doob martingale
Vm = E(# 01|Z0, Z1, · · · , Zm), m = 0, · · · , ˜n
2 Bound |Vm − Vm−1| using Coupling argument
3 Azuma’s inequality closes the bound
33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling Technique
X0
Y0
X1
Y1
X2
Y2
X3
Y3
X4
Y4
· · ·
· · ·
Xn−1
Yn−1
Xn
Yn
33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling Technique
X0
Y0
X1
Y1
X2
Y2
X3
Y3
X4
Y4
· · ·
· · ·
Xn−1
Yn−1
Xn
Yn
33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling Technique
X0
Y0
X1
Y1
X2
Y2
X3
Y3
X4
Y4
· · ·
· · ·
Xn−1
Yn−1
Xn
Yn
33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling Technique
X0
Y0
X1
Y1
X2
Y2
X3
Y3
X4
Y4
· · ·
· · ·
Xn−1
Yn−1
Xn
Yn
Jointly evolve according to ω
33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling Technique
X0
Y0
X1
Y1
X2
Y2
X3
Y3
X4
Y4
· · ·
· · ·
Xn−1
Yn−1
Xn
Yn
Jointly evolve according to ω
Encourage evolution st Xt = Yt for all t after (random) τ steps
P(τ  i) = ω(Xi = Yi)
E(τ) =
i≥1
P(τ  i) =
i≥1
ω(Xi=Yi)
Sample size Eτ ⇒ empirical ≈ stationary [Aldous]
34 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Proof Outline (contd)
• Run two coupled copies Zj and Zj of the restricted process
• Long and unknown memory ⇒ They do not coalesce the usual way
• Instead, approximate coalescence ⇒ Longer together, harder to separate out
• Reason: Dependencies die down
|Vm − Vm−1| ≤
n
j=m+1
ω(Zj≈Zj )
35 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Part II:
Sampling From Slow Mixing
Markov Processes
36 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling From the Past (CFTP)
Markov Chain Over S
Stationary Distribution is π
Goal: Sample from exactly π(.)
Idea (Propp and Wilson):
 Expose states to shared source of randomness
 Simulate chains backward in time
 Wait until all chains merge
A
B C
1
3
2
3
1
2
3
4
1
4
1
2
π(A) = 9
19, π(B) = 4
19, π(C) = 6
19
37 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Coupling
A
B C
1
3
2
3
1
2
3
4
1
4
1
2
A B C
A
C
B
A
A
B
[0, 1]
1
3
38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
A B C
C
B
A
C
B
A
38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
A B C
C
B
A
C
B
A
C
B
A
38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
A B C
C
B
A
C
B
A
C
B
A
C
B
A
38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
A B C
C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
A B C
C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Example
time
-6 -5 -4 -3 -2 -1 0
C
Coalescence
Output ∼ π(.)C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
C
B
A
39 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Graph Clustering
Similarity Graph
39 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Graph Clustering
Similarity Graph
Reorganizing
39 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Graph Clustering
Similarity Graph
Reorganizing
Complexity: NP-hard
40 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Community Detection on Graphs
Number of clusters?
 Usually not known in advance.
What is a good measure?
 Nodes within a cluster tightly connected.
 Nodes in disparate clusters loosely connected.
Correlation Clustering
 Cost: Graph clustering distance
 Min # of edge add/deletion s.t.
Graph G © disjoint cliques
41 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Approaches
Spectral Clustering
• Eigenvectors of Laplacian of similarity graph
 Simple implementation
 Laplacian could be ill-conditioned
 # of clusters need to be known in advance
[Ng, Jordan, White, Smyth, Weiss, Shi, Malik, Kannan, Vempala, Vetta, Meila, · · · ]
Semi-Definite Programming (SDP)
• LP relaxation
 Asymptotically optimal for Stochastic Block Models
 Implicit assumption on generative model
 # of clusters need to be known in advance
[Abbe, Sandon, Hajeck, Bandeira, Hall, Decelle, Mossel, Neeman, Sly, Rao, Chen, Wu, Xu, · · · ]
42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Similarity graph.
42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Similarity graph.
Define random walk.
42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Non-uniform
Probability of following a link ∝ # of common neighbors
Similarity graph.
Define random walk.
Couple random walks
starting from different nodes
42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Similarity graph.
Define random walk.
Couple random walks
starting from different nodes
Adapt CFTP algorithm
42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Similarity graph.
Define random walk.
Couple random walks
starting from different nodes
Adapt CFTP algorithm
 Do not care about exact
sample
42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Similarity graph.
Define random walk.
Couple random walks
starting from different nodes
Adapt CFTP algorithm
 Do not care about exact
sample
 Identify clusters “before
coalescence happens”
42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Core Idea
a
b
c
d
e
f
g
h
i
Restricted walk within a cluster mixes faster.
Similarity graph.
Define random walk.
Couple random walks
starting from different nodes
Adapt CFTP algorithm
 Do not care about exact
sample
 Identify clusters “before
coalescence happens”
43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Initially: All singletons
43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Partial coalescence
Small Clusters formation
43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Critical times
Clusters merge
43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Critical times
Clusters merge
43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Critical times
Clusters merge
43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Algorithm Overview
S
Full coalescence
44 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Partial Coalescence
time
S
G
Gc
-10 -9 -8 -7 -6 -5 -4 -3 -2 -1 0
Paths starting from G ⊂ S
45 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Remarks
 Algorithm can stop at critical times
 Still yield a cluster
 Choose cluster with optimal cost
 Purely based on random walk
 Use of mixing properties
 Circumvent issues with ill-conditioned matrices in spectral based approaches
 No prior assumption on generative model
 No prior assumption on number of clusters
46 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Benchmark Networks
Stochastic Block Models (SBM)
 Size of communities equal
 Average degree equal for all nodes
LFR Models
 More realistic
 # of communities and sizes admit power law.
Real world networks
47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Stochastic Block Model
47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Stochastic Block Model
Ber(p)
47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Stochastic Block Model
Ber(p)
Ber(q)
47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Stochastic Block Model
Ber(p)
Ber(q)
p q
48 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
SBM Realization
500 nodes
p = 0.5
q = 0.1
5 communities
Randomly permuted.
49 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Identifying communities
50 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
CC-PIVOT Output
50 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
LFR Model
200 nodes
6 communities
51 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Identifying communities
52 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
CC-PIVOT output
53 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
American College Football
115 teams.
Divided into conferences.
54 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Identified communities
55 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
CC-PIVOT Output
56 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Future Directions
Open: Theoretical guarantees for recovery in SBMs
Open: Stationary is more difficult (needs twice summability of d) than
transition (just summability of d)
Theoretical foundation of the proposed algorithm
Extension to broader set of community detection problems
Mahalo!
58 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally
Acknowledgement
My Advisor: Prasad
Committee Members
Dr. Rui Zhang
Office Mates:
• Maryam and Meysam
My Friends:
• Masoud, Saeed, Navid, Elyas, Harir Chee, Reza, Ehsaneh, Ali, Ashkan, Hamed,
Ehsan, Seyed, Alireza,...
 My Family:
• My parents, my sister Nasrin and my brother Naser

More Related Content

Viewers also liked

Legge di Bilancio 2017 Iperammortamento 250%
Legge di Bilancio 2017 Iperammortamento 250%Legge di Bilancio 2017 Iperammortamento 250%
Legge di Bilancio 2017 Iperammortamento 250%Mattia Pilosio
 
Toovey_et_al-2016-DMCN-casecontrolstudy
Toovey_et_al-2016-DMCN-casecontrolstudyToovey_et_al-2016-DMCN-casecontrolstudy
Toovey_et_al-2016-DMCN-casecontrolstudyDr Rachel Toovey
 
Compiler Design Introduction
Compiler Design IntroductionCompiler Design Introduction
Compiler Design IntroductionRicha Sharma
 
Deep learning - A Visual Introduction
Deep learning - A Visual IntroductionDeep learning - A Visual Introduction
Deep learning - A Visual IntroductionLukas Masuch
 
La organización dentro de la administración
La organización dentro de la administraciónLa organización dentro de la administración
La organización dentro de la administraciónCris Guaicara
 
Deep learning - Conceptual understanding and applications
Deep learning - Conceptual understanding and applicationsDeep learning - Conceptual understanding and applications
Deep learning - Conceptual understanding and applicationsBuhwan Jeong
 
A1. MEP 1 - Mod 1 - Tema1
A1. MEP 1 - Mod 1 - Tema1A1. MEP 1 - Mod 1 - Tema1
A1. MEP 1 - Mod 1 - Tema1Poliana Bellan
 
Deep Learning and the state of AI / 2016
Deep Learning and the state of AI / 2016Deep Learning and the state of AI / 2016
Deep Learning and the state of AI / 2016Grigory Sapunov
 
Deep Learning through Examples
Deep Learning through ExamplesDeep Learning through Examples
Deep Learning through ExamplesSri Ambati
 

Viewers also liked (9)

Legge di Bilancio 2017 Iperammortamento 250%
Legge di Bilancio 2017 Iperammortamento 250%Legge di Bilancio 2017 Iperammortamento 250%
Legge di Bilancio 2017 Iperammortamento 250%
 
Toovey_et_al-2016-DMCN-casecontrolstudy
Toovey_et_al-2016-DMCN-casecontrolstudyToovey_et_al-2016-DMCN-casecontrolstudy
Toovey_et_al-2016-DMCN-casecontrolstudy
 
Compiler Design Introduction
Compiler Design IntroductionCompiler Design Introduction
Compiler Design Introduction
 
Deep learning - A Visual Introduction
Deep learning - A Visual IntroductionDeep learning - A Visual Introduction
Deep learning - A Visual Introduction
 
La organización dentro de la administración
La organización dentro de la administraciónLa organización dentro de la administración
La organización dentro de la administración
 
Deep learning - Conceptual understanding and applications
Deep learning - Conceptual understanding and applicationsDeep learning - Conceptual understanding and applications
Deep learning - Conceptual understanding and applications
 
A1. MEP 1 - Mod 1 - Tema1
A1. MEP 1 - Mod 1 - Tema1A1. MEP 1 - Mod 1 - Tema1
A1. MEP 1 - Mod 1 - Tema1
 
Deep Learning and the state of AI / 2016
Deep Learning and the state of AI / 2016Deep Learning and the state of AI / 2016
Deep Learning and the state of AI / 2016
 
Deep Learning through Examples
Deep Learning through ExamplesDeep Learning through Examples
Deep Learning through Examples
 

Similar to RamezanPhDSlides

Answering complex open domain questions with multi-hop dense retrieval
Answering complex open domain questions with multi-hop dense retrievalAnswering complex open domain questions with multi-hop dense retrieval
Answering complex open domain questions with multi-hop dense retrievalSan Kim
 
CS221: HMM and Particle Filters
CS221: HMM and Particle FiltersCS221: HMM and Particle Filters
CS221: HMM and Particle Filterszukun
 
A measure to evaluate latent variable model fit by sensitivity analysis
A measure to evaluate latent variable model fit by sensitivity analysisA measure to evaluate latent variable model fit by sensitivity analysis
A measure to evaluate latent variable model fit by sensitivity analysisDaniel Oberski
 
Complex sampling in latent variable models
Complex sampling in latent variable modelsComplex sampling in latent variable models
Complex sampling in latent variable modelsDaniel Oberski
 
Csr2011 june14 11_00_aaronson
Csr2011 june14 11_00_aaronsonCsr2011 june14 11_00_aaronson
Csr2011 june14 11_00_aaronsonCSR2011
 
An Introduction to Quantum Programming Languages
An Introduction to Quantum Programming LanguagesAn Introduction to Quantum Programming Languages
An Introduction to Quantum Programming LanguagesDavid Yonge-Mallo
 
Paper Study - Demand-Driven Computation of Interprocedural Data Flow
Paper Study - Demand-Driven Computation of Interprocedural Data FlowPaper Study - Demand-Driven Computation of Interprocedural Data Flow
Paper Study - Demand-Driven Computation of Interprocedural Data FlowMin-Yih Hsu
 
Functional specialization in human cognition: a large-scale neuroimaging init...
Functional specialization in human cognition: a large-scale neuroimaging init...Functional specialization in human cognition: a large-scale neuroimaging init...
Functional specialization in human cognition: a large-scale neuroimaging init...Ana Luísa Pinho
 
The Automated-Reasoning Revolution: from Theory to Practice and Back
The Automated-Reasoning Revolution: from Theory to Practice and BackThe Automated-Reasoning Revolution: from Theory to Practice and Back
The Automated-Reasoning Revolution: from Theory to Practice and BackMoshe Vardi
 
Fast ALS-based matrix factorization for explicit and implicit feedback datasets
Fast ALS-based matrix factorization for explicit and implicit feedback datasetsFast ALS-based matrix factorization for explicit and implicit feedback datasets
Fast ALS-based matrix factorization for explicit and implicit feedback datasetsGravity - Rock Solid Recommendations
 
SPIE Conference V3.0
SPIE Conference V3.0SPIE Conference V3.0
SPIE Conference V3.0Robert Fry
 
Scalable Dynamic Graph Summarization
Scalable Dynamic Graph SummarizationScalable Dynamic Graph Summarization
Scalable Dynamic Graph SummarizationIoanna Tsalouchidou
 
Algorithm Portfolios for Noisy Optimization: Compare Solvers Early (LION8)
Algorithm Portfolios for Noisy Optimization: Compare Solvers Early (LION8)Algorithm Portfolios for Noisy Optimization: Compare Solvers Early (LION8)
Algorithm Portfolios for Noisy Optimization: Compare Solvers Early (LION8)Jialin LIU
 
Introduction to recommender systems
Introduction to recommender systemsIntroduction to recommender systems
Introduction to recommender systemsArnaud de Myttenaere
 
Pwl rewal-slideshare
Pwl rewal-slidesharePwl rewal-slideshare
Pwl rewal-slidesharepalvaro
 

Similar to RamezanPhDSlides (20)

Answering complex open domain questions with multi-hop dense retrieval
Answering complex open domain questions with multi-hop dense retrievalAnswering complex open domain questions with multi-hop dense retrieval
Answering complex open domain questions with multi-hop dense retrieval
 
CS221: HMM and Particle Filters
CS221: HMM and Particle FiltersCS221: HMM and Particle Filters
CS221: HMM and Particle Filters
 
A measure to evaluate latent variable model fit by sensitivity analysis
A measure to evaluate latent variable model fit by sensitivity analysisA measure to evaluate latent variable model fit by sensitivity analysis
A measure to evaluate latent variable model fit by sensitivity analysis
 
Complex sampling in latent variable models
Complex sampling in latent variable modelsComplex sampling in latent variable models
Complex sampling in latent variable models
 
Csr2011 june14 11_00_aaronson
Csr2011 june14 11_00_aaronsonCsr2011 june14 11_00_aaronson
Csr2011 june14 11_00_aaronson
 
An Introduction to Quantum Programming Languages
An Introduction to Quantum Programming LanguagesAn Introduction to Quantum Programming Languages
An Introduction to Quantum Programming Languages
 
Paper Study - Demand-Driven Computation of Interprocedural Data Flow
Paper Study - Demand-Driven Computation of Interprocedural Data FlowPaper Study - Demand-Driven Computation of Interprocedural Data Flow
Paper Study - Demand-Driven Computation of Interprocedural Data Flow
 
Functional specialization in human cognition: a large-scale neuroimaging init...
Functional specialization in human cognition: a large-scale neuroimaging init...Functional specialization in human cognition: a large-scale neuroimaging init...
Functional specialization in human cognition: a large-scale neuroimaging init...
 
The Automated-Reasoning Revolution: from Theory to Practice and Back
The Automated-Reasoning Revolution: from Theory to Practice and BackThe Automated-Reasoning Revolution: from Theory to Practice and Back
The Automated-Reasoning Revolution: from Theory to Practice and Back
 
Fast ALS-based matrix factorization for explicit and implicit feedback datasets
Fast ALS-based matrix factorization for explicit and implicit feedback datasetsFast ALS-based matrix factorization for explicit and implicit feedback datasets
Fast ALS-based matrix factorization for explicit and implicit feedback datasets
 
Pnp
PnpPnp
Pnp
 
SM2701 Class 11
SM2701 Class 11SM2701 Class 11
SM2701 Class 11
 
SPIE Conference V3.0
SPIE Conference V3.0SPIE Conference V3.0
SPIE Conference V3.0
 
Blinkdb
BlinkdbBlinkdb
Blinkdb
 
Scalable Dynamic Graph Summarization
Scalable Dynamic Graph SummarizationScalable Dynamic Graph Summarization
Scalable Dynamic Graph Summarization
 
Algorithm Portfolios for Noisy Optimization: Compare Solvers Early (LION8)
Algorithm Portfolios for Noisy Optimization: Compare Solvers Early (LION8)Algorithm Portfolios for Noisy Optimization: Compare Solvers Early (LION8)
Algorithm Portfolios for Noisy Optimization: Compare Solvers Early (LION8)
 
Introduction to recommender systems
Introduction to recommender systemsIntroduction to recommender systems
Introduction to recommender systems
 
Pwl rewal-slideshare
Pwl rewal-slidesharePwl rewal-slideshare
Pwl rewal-slideshare
 
ilp-nlp-slides.pdf
ilp-nlp-slides.pdfilp-nlp-slides.pdf
ilp-nlp-slides.pdf
 
Fine Grained Complexity
Fine Grained ComplexityFine Grained Complexity
Fine Grained Complexity
 

RamezanPhDSlides

  • 1. ESTIMATION AND sampling IN SLOW MIXING MARKOV PROCESSES Ramezan Paravi | Ph.D. Candidate EE Department, UH Manoa Advisor: Dr. Santhanam August 2015
  • 2. 2 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview Markov Sources in slow mixing regime: Empirical counts do not reflect stationary Analysis of samples before mixing happens Part I: Statistical properties of finite samples Part II: Modeling using slow mixing Markov process Andrey Markov A B C 1 3 2 31 2 3 4 1 4 1 2
  • 3. 3 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Motivation Any structure?
  • 4. 3 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Motivation Any structure? Reorganizing
  • 5. 3 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Motivation Any structure? Reorganizing • Social Networks • Biological Networks • Recommender Systems
  • 6. 4 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Motivation 1 2 3 4 5 6 7 8 9 10 11 1213 1415 Random walk on graph
  • 7. 4 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Motivation 1 2 3 4 5 6 7 8 9 10 11 1213 1415 Uniform random walk. Explore state space fast.
  • 8. 4 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Motivation 1 2 3 4 5 6 7 8 9 10 11 1213 1415 Non-uniform random walk. Polarized state space. Walks starting here will be trapped.
  • 9. 5 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results p(1|111) p(1|011) p(1|101) p(1|001) 1 0 1 0 1 0 1 Given sample Y1, . . . ,Yn from unknown binary Markov source p Transition probabilities? Stationary probabilities? What we do: fixed sample, best answer What we are not doing: e.a.s. results
  • 10. 6 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Complications Two major sources of difficulties: Long memory Slow mixing May not estimate accurately/completely given n samples Rather, want best possible answer with sample
  • 11. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011
  • 12. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source?
  • 13. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1
  • 14. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q2: If B ∼ i.i.d, what can be said?
  • 15. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q2: If B ∼ i.i.d, what can be said? Ans: P(1) P(0), (only 3 zeros in the sample)
  • 16. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q2: If B ∼ i.i.d, what can be said? Ans: P(1) P(0), (only 3 zeros in the sample) Q3: If A ∼ Markov memory-1, what about transition probs?
  • 17. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q2: If B ∼ i.i.d, what can be said? Ans: P(1) P(0), (only 3 zeros in the sample) Q3: If A ∼ Markov memory-1, what about transition probs? Ans: P(1|1) is high
  • 18. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q4: If B ∼ i.i.d, then what about P(0)?
  • 19. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q4: If B ∼ i.i.d, then what about P(0)? Ans: More 1’s than 0’s. With high confidence, P(0) small.
  • 20. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q4: If B ∼ i.i.d, then what about P(0)? Ans: More 1’s than 0’s. With high confidence, P(0) small. Q5: If A ∼ Markov memory-1, what about P(0)?
  • 21. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q4: If B ∼ i.i.d, then what about P(0)? Ans: More 1’s than 0’s. With high confidence, P(0) small. Q5: If A ∼ Markov memory-1, what about P(0)? Ans: Can not judge with finite sample.
  • 22. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q6: If B ∼ i.i.d and see more bits, then what?
  • 23. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q6: If B ∼ i.i.d and see more bits, then what? Ans: Likely lots of 1’s, few 0’s. P(0) ≈ #0’s len is small.
  • 24. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q6: If B ∼ i.i.d and see more bits, then what? Ans: Likely lots of 1’s, few 0’s. P(0) ≈ #0’s len is small. Q7: If A ∼ Markov memory-1 and see more bits, can we say #0’s len is small?
  • 25. 7 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Two Samples with same # of 0’s and 1s: A 11111111111111000 v.s. B 11011110111111011 Q1: Which one is generated by a memory-1 Markov source v.s an i.i.d source? Ans: Probably, B ∼ i.i.d, while A ∼ Markov memory-1 Q6: If B ∼ i.i.d and see more bits, then what? Ans: Likely lots of 1’s, few 0’s. P(0) ≈ #0’s len is small. Q7: If A ∼ Markov memory-1 and see more bits, can we say #0’s len is small? Ans: NO. P(0) could be arbitrarily large!!
  • 26. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Transition Probabilities: Given sample x, string w and a ∈ A, can P(a|w) be estimated from x?
  • 27. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Transition Probabilities: Given sample x, string w and a ∈ A, can P(a|w) be estimated from x? Easy: If |w| memory :
  • 28. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Transition Probabilities: Given sample x, string w and a ∈ A, can P(a|w) be estimated from x? Easy: If |w| memory : Subsequence following w is i.i.d (# wa) (# w) ≈ P(a|w)
  • 29. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Transition Probabilities: Given sample x, string w and a ∈ A, can P(a|w) be estimated from x? Easy: If |w| memory : Subsequence following w is i.i.d (# wa) (# w) ≈ P(a|w) Harder: If memory unknown, but the source has mixed:
  • 30. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Transition Probabilities: Given sample x, string w and a ∈ A, can P(a|w) be estimated from x? Easy: If |w| memory : Subsequence following w is i.i.d (# wa) (# w) ≈ P(a|w) Harder: If memory unknown, but the source has mixed: Both #w and #wa reflect P(w) and P(wa) Still (# wa) (# w) ≈ P(a|w)
  • 31. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Transition Probabilities: Given sample x, string w and a ∈ A, can P(a|w) be estimated from x? Easy: If |w| memory : Subsequence following w is i.i.d (# wa) (# w) ≈ P(a|w) Harder: If memory unknown, but the source has mixed: Both #w and #wa reflect P(w) and P(wa) Still (# wa) (# w) ≈ P(a|w) Difficult: If memory unknown and the source has not mixed:
  • 32. 8 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Overview of Theoretical Results Transition Probabilities: Given sample x, string w and a ∈ A, can P(a|w) be estimated from x? Easy: If |w| memory : Subsequence following w is i.i.d (# wa) (# w) ≈ P(a|w) Harder: If memory unknown, but the source has mixed: Both #w and #wa reflect P(w) and P(wa) Still (# wa) (# w) ≈ P(a|w) Difficult: If memory unknown and the source has not mixed: Non trivial
  • 33. 9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Results on Estimation Given a length-n sample from binary model with dying dependencies Amount of information two bits provide about each other, conditioned on middle, diminishes farther they are.
  • 34. 9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Results on Estimation Given a length-n sample from binary model with dying dependencies Amount of information two bits provide about each other, conditioned on middle, diminishes farther they are. Identify from data which parameters can be estimated Set ˜G of “good” strings w of length Θ(log n) Only those frequently occurred in the sample
  • 35. 9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Results on Estimation Given a length-n sample from binary model with dying dependencies Amount of information two bits provide about each other, conditioned on middle, diminishes farther they are. Identify from data which parameters can be estimated Set ˜G of “good” strings w of length Θ(log n) Only those frequently occurred in the sample Provide (confidence and) accuracy bounds
  • 36. 9 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Results on Estimation Given a length-n sample from binary model with dying dependencies Amount of information two bits provide about each other, conditioned on middle, diminishes farther they are. Identify from data which parameters can be estimated Set ˜G of “good” strings w of length Θ(log n) Only those frequently occurred in the sample Provide (confidence and) accuracy bounds Transition probabilities of w ∈ ˜G Universal Compression + combinatorial arguments (# wa) (# w) ≈ P(a|w)
  • 37. 10 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Results on Estimation Surprises: !! even if (#wa) or (#w) ≈ stationary ! bound may not hold for shorter substrings of w Stationary Transition
  • 38. 11 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Results on Estimation Provide (confidence and) accuracy bounds
  • 39. 11 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Results on Estimation Provide (confidence and) accuracy bounds Stationary probabilities of w ∈ ˜G Doob Martingale + Coupling Markov chains + Azuma #w ˜n ≈ P(w) P( ˜G) for w ∈ ˜G, ˜n = number of times strings in ˜G appear in the sample
  • 40. 12 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Results on Estimation All bounds are entirely data dependent Confidence/accuracy obtained w.h.p. from sample Knowledge of the source is not required
  • 41. 13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Modeling using slow mixing Provide randomized algorithm for community detection
  • 42. 13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Modeling using slow mixing Provide randomized algorithm for community detection
  • 43. 13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Modeling using slow mixing Provide randomized algorithm for community detection Build slow mixing random walks on graphs
  • 44. 13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Modeling using slow mixing Provide randomized algorithm for community detection Build slow mixing random walks on graphs Mixing properties of the walk → reveal community structure Framework: Coupling From the Past
  • 45. 13 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Modeling using slow mixing Provide randomized algorithm for community detection Build slow mixing random walks on graphs Mixing properties of the walk → reveal community structure Framework: Coupling From the Past Simulation results on benchmark networks
  • 46. 14 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Part I: Estimation in Slow Mixing Markov Processes
  • 47. 15 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Challenges: Arbitrary Mass All strings in a finite sample, can have arbitrarily small mass. 1 − m 1 0 p(0) = m m+1 p(1) = 1 m+1 p(1) can be arbitrarily small for m large enough. If o(1/n), starting from 1, see a sequence of O(1/ ) 1’s whp. p(any seq. of 1’s) ≤ 1 m+1, can be arbitrarily small.
  • 48. 16 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Challenges: Long Memory 1/2 1/2 1/2 1/2 1/2 1/2 1/2 1/2 ------------- - - - - - - - - - - - - - - - - depth k ↑1 ↓0 Bernoulli(1/2) i.i.d. source. Over parameterized.
  • 49. 17 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Challenges: Long Memory 1/2 1/2 1/2 1/2 1 − 2 1/2 1/2 ------------- - - - - - - - - - - - - - - - - depth k ↑1 ↓0 If k ω(log n), starting from 1 we will not see a sequence of k − 1 consecutive 0s whp in a sample of size n.
  • 50. 17 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Challenges: Long Memory 1/2 1/2 1/2 1/2 1 − 2 1/2 1/2 ------------- - - - - - - - - - - - - - - - - depth k ↑1 ↓0 If k ω(log n), starting from 1 we will not see a sequence of k − 1 consecutive 0s whp in a sample of size n. Therefore, all bits generated wp 1 2. Cannot distinguish from i.i.d. Bernoulli (1/2) whp
  • 51. 18 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Challenges: Slow Mixing p(0) = p(1) = 1 2 1 − 1 0 p (1) = 2 3, p (0) = 1 3 1 − 2 1 0 Starting from 1, whp both generate sequence of O(1/ ) 1s. If o(1/n), whp cannot distinguish using length-n sample
  • 52. 18 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Challenges: Slow Mixing p(0) = p(1) = 1 2 1 − 1 0 p (1) = 2 3, p (0) = 1 3 1 − 2 1 0 Starting from 1, whp both generate sequence of O(1/ ) 1s. If o(1/n), whp cannot distinguish using length-n sample Caution! Slow mixing only hurts estimation, not compression! good compression for memory-1 sources, slow mixing or not
  • 53. 19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Prior Work Any consistent estimator converges pointwise (NOT uniformly) over the class of stationary and ergodic Markov models. Extensive work on
  • 54. 19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Prior Work Any consistent estimator converges pointwise (NOT uniformly) over the class of stationary and ergodic Markov models. Extensive work on consistency,
  • 55. 19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Prior Work Any consistent estimator converges pointwise (NOT uniformly) over the class of stationary and ergodic Markov models. Extensive work on consistency, e.a.s. results,
  • 56. 19 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Prior Work Any consistent estimator converges pointwise (NOT uniformly) over the class of stationary and ergodic Markov models. Extensive work on consistency, e.a.s. results, finite-sample but model dependent results [Buhlmann, Csiszár, alves, Gariviér, Leonardi, Marton, Maum-Deschamps, Morvai, Rissanen, Schmitt,Shields, Talata, Weiss, Wyner] Our Philosophy Can we look at a length-n sample and identify what, if anything, can be estimated accurately?
  • 57. 20 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Dependencies Problem futile if the dependencies are arbitrary. We assume dependencies die down. d(4) Formalize with d : N → R+. Siblings s, s (nodes of same color) satisfy p(1|s) p(1|s ) − 1 ≤ d(4). Md = {srcs satisfying above for all siblings}.
  • 58. 21 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Dependencies Die Down Information-theoretic interpretation I(Y0; Yi+1|Y i 1 ) ≤ log(1 + d(i)). b1 b2 b2b1 Not related to mixing properties of the source. No bound on memory of the source. Need d(i) summable over i, equivalently δj = i≥j d(i) → 0
  • 59. 22 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Aggregation Unknown set of states T kn = Θ(log n)
  • 60. 22 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Aggregation Unknown set of states T Consider a coarser model (aggregation) Ask p(1|w), where |w| = kn kn = Θ(log n)
  • 61. 22 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Aggregation Unknown set of states T Consider a coarser model (aggregation) Ask p(1|w), where |w| = kn Set kn = Θ(log n) Makes sense to ask length-3 aggregations for memory-2 source (source itself) kn = Θ(log n)
  • 62. 23 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Goal Unknown source p in Md For any length-kn string w (known kn = Θ(log n)) • Estimate transition p(·|w) • Estimate stationary p(w) kn = Θ(log n)
  • 63. 23 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Goal Unknown source p in Md For any length-kn string w (known kn = Θ(log n)) • Estimate transition p(·|w) • Estimate stationary p(w) p p Unkown Src in Md Space of memory-kn Srcs Caution The problem is not the same as estimating a memory-kn source. We never see samples from the aggregated model.
  • 64. 24 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Naive Estimators Example Suppose Y 0 −∞ = · · · 00 and Y n 1 = 11010010011. Depth-2 aggregated parameters, e,g., transition prb from w = 10. · · · 00, 110 10 10 10 0 Subsequence following 10 is 1110
  • 65. 24 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Naive Estimators Example Suppose Y 0 −∞ = · · · 00 and Y n 1 = 11010010011. Depth-2 aggregated parameters, e,g., transition prb from w = 10. · · · 00, 110 10 10 10 0 Subsequence following 10 is 1110 This is not iid in general because sample is from true model, not aggregated model!
  • 66. 24 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Naive Estimators Example Suppose Y 0 −∞ = · · · 00 and Y n 1 = 11010010011. Depth-2 aggregated parameters, e,g., transition prb from w = 10. · · · 00, 110 10 10 10 0 Subsequence following 10 is 1110 This is not iid in general because sample is from true model, not aggregated model! Naive (“1110 i.i.d.”): ˆp(1|w) = 3 4 and ˆp(0|w) = 1 4.
  • 67. 25 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Deviation Bounds for Conditional Probabilities Theorem Confidence ≥ 1 − 1 22kn+1 log n (conditioned on any past Y 0 −∞), for all w ∈ {0, 1}kn simultaneously (#w·) (#w) − p(·|w) 1 ≤ 2 (ln 2)(2kn+1 log n + nδkn ) Nw . Again, δkn = i≥kn d(i). The more occurrence, the stronger bound. If d(i) decreases exponentially as γi and Nw = nβ, rhs diminishes as O( 1√ nβ−1−γ log γ ).
  • 68. 26 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Proof Idea The result is built on following facts: • Source belongs to Md, i.e., dependencies die down. • Compression result on Md reminiscent of method of types. • Arguments relating strong compression results to variational distance between estimators.
  • 69. 27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Surprises Proof order Stationary Transition
  • 70. 27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Surprises Proof order Stationary Transition Bound may hold for w ! even without empirical frequencies ≈ stationary probabilities
  • 71. 27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Surprises Proof order Stationary Transition Bound may hold for w ! even without empirical frequencies ≈ stationary probabilities !! but not for its suffixes
  • 72. 27 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Surprises Proof order Stationary Transition Bound may hold for w ! even without empirical frequencies ≈ stationary probabilities !! but not for its suffixes Possible p(1|100 zeros), but p(1|0)
  • 73. 28 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Good States Good “states” ˜G are length-kn strings appearing frequently enough ˜G = w : count(w) ≥ max{nδkn log 1 δkn , 2kn+1 log2 n} Concentration bound is at least as fast as 1√ log n . If d(i) decreases exponentially as γi, concentration bound is poly in n.
  • 74. 29 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Stationary Probabilities Sensitive function of conditional probabilities How interpret counts of w ∈ ˜G in the sample?
  • 75. 30 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Deviation Bounds for Stationary Probabilities Theorem For any t 0, Y 0 −∞ and w ∈ ˜G, P (#w) ˜n − p(w) p( ˜G) ≥ t|Y 0 −∞ ≤ 2 exp − (˜nt − B)2 2˜nB2 where ˜n is the sum count of states in ˜G, B depends on n and how quickly dependencies d(i) die off.
  • 76. 30 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Deviation Bounds for Stationary Probabilities Theorem For any t 0, Y 0 −∞ and w ∈ ˜G, P (#w) ˜n − p(w) p( ˜G) ≥ t|Y 0 −∞ ≤ 2 exp − (˜nt − B)2 2˜nB2 where ˜n is the sum count of states in ˜G, B depends on n and how quickly dependencies d(i) die off. B = O(log n) if d(i) = γi For bound to be non-vacuously true, need d to be “twice summable”, or δi = j≥i d(j) to be summable
  • 77. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Along a Sequence of Stopping Times Restriction of the process to Good states. Y 0 −∞ Y1 Y2 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Yn ↑1 ↓0
  • 78. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Along a Sequence of Stopping Times Restriction of the process to Good states. Y 0 −∞ Y1 Y2 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Yn ↑1 ↓0 ˜G = {01, 10} 1 0
  • 79. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Along a Sequence of Stopping Times Restriction of the process to Good states. Y 0 −∞ Y1 Y2 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Yn 1 0 ↑ τ0 ↑1 ↓0 w = 10 1 1 110 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Z0 1 1 0
  • 80. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Along a Sequence of Stopping Times Restriction of the process to Good states. Y 0 −∞ Y1 Y2 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Yn 0 1 ↑ τ1 ↑1 ↓0 w = 01 0 0 001 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Z1 0 0 1
  • 81. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Along a Sequence of Stopping Times Restriction of the process to Good states. Y 0 −∞ Y1 Y2 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Yn 1 0 ↑ τ2 ↑1 ↓0 w = 10 1 1 110 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Z2 1 1 0
  • 82. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Along a Sequence of Stopping Times Restriction of the process to Good states. Y 0 −∞ Y1 Y2 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Yn 0 1 ↑ τ3 ↑1 ↓0 w = 01 0 0 001 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Z3 0 0 1
  • 83. 31 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Estimation Along a Sequence of Stopping Times Restriction of the process to Good states. Y 0 −∞ Y1 Y2 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Yn 0 1 ↑ τ˜n ↑1 ↓0 w = 01 0 0 001 0 1 1 1 0 0 0 1 1 1 1 0 0 1 1 1 0 0 1 1 Z˜n 0 0 1
  • 84. 32 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Proof Outline 1 Construct a natural Doob martingale Vm = E(# 01|Z0, Z1, · · · , Zm), m = 0, · · · , ˜n
  • 85. 32 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Proof Outline 1 Construct a natural Doob martingale Vm = E(# 01|Z0, Z1, · · · , Zm), m = 0, · · · , ˜n 2 Bound |Vm − Vm−1| using Coupling argument
  • 86. 32 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Proof Outline 1 Construct a natural Doob martingale Vm = E(# 01|Z0, Z1, · · · , Zm), m = 0, · · · , ˜n 2 Bound |Vm − Vm−1| using Coupling argument 3 Azuma’s inequality closes the bound
  • 87. 33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Coupling Technique X0 Y0 X1 Y1 X2 Y2 X3 Y3 X4 Y4 · · · · · · Xn−1 Yn−1 Xn Yn
  • 88. 33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Coupling Technique X0 Y0 X1 Y1 X2 Y2 X3 Y3 X4 Y4 · · · · · · Xn−1 Yn−1 Xn Yn
  • 89. 33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Coupling Technique X0 Y0 X1 Y1 X2 Y2 X3 Y3 X4 Y4 · · · · · · Xn−1 Yn−1 Xn Yn
  • 90. 33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Coupling Technique X0 Y0 X1 Y1 X2 Y2 X3 Y3 X4 Y4 · · · · · · Xn−1 Yn−1 Xn Yn Jointly evolve according to ω
  • 91. 33 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Coupling Technique X0 Y0 X1 Y1 X2 Y2 X3 Y3 X4 Y4 · · · · · · Xn−1 Yn−1 Xn Yn Jointly evolve according to ω Encourage evolution st Xt = Yt for all t after (random) τ steps P(τ i) = ω(Xi = Yi) E(τ) = i≥1 P(τ i) = i≥1 ω(Xi=Yi) Sample size Eτ ⇒ empirical ≈ stationary [Aldous]
  • 92. 34 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Proof Outline (contd) • Run two coupled copies Zj and Zj of the restricted process • Long and unknown memory ⇒ They do not coalesce the usual way • Instead, approximate coalescence ⇒ Longer together, harder to separate out • Reason: Dependencies die down |Vm − Vm−1| ≤ n j=m+1 ω(Zj≈Zj )
  • 93. 35 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Part II: Sampling From Slow Mixing Markov Processes
  • 94. 36 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Coupling From the Past (CFTP) Markov Chain Over S Stationary Distribution is π Goal: Sample from exactly π(.) Idea (Propp and Wilson): Expose states to shared source of randomness Simulate chains backward in time Wait until all chains merge A B C 1 3 2 3 1 2 3 4 1 4 1 2 π(A) = 9 19, π(B) = 4 19, π(C) = 6 19
  • 95. 37 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Coupling A B C 1 3 2 3 1 2 3 4 1 4 1 2 A B C A C B A A B [0, 1] 1 3
  • 96. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Example time -6 -5 -4 -3 -2 -1 0 A B C C B A C B A
  • 97. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Example time -6 -5 -4 -3 -2 -1 0 A B C C B A C B A C B A
  • 98. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Example time -6 -5 -4 -3 -2 -1 0 A B C C B A C B A C B A C B A
  • 99. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Example time -6 -5 -4 -3 -2 -1 0 A B C C B A C B A C B A C B A C B A
  • 100. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Example time -6 -5 -4 -3 -2 -1 0 A B C C B A C B A C B A C B A C B A C B A
  • 101. 38 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Example time -6 -5 -4 -3 -2 -1 0 C Coalescence Output ∼ π(.)C B A C B A C B A C B A C B A C B A C B A
  • 102. 39 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Graph Clustering Similarity Graph
  • 103. 39 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Graph Clustering Similarity Graph Reorganizing
  • 104. 39 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Graph Clustering Similarity Graph Reorganizing Complexity: NP-hard
  • 105. 40 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Community Detection on Graphs Number of clusters? Usually not known in advance. What is a good measure? Nodes within a cluster tightly connected. Nodes in disparate clusters loosely connected. Correlation Clustering Cost: Graph clustering distance Min # of edge add/deletion s.t. Graph G © disjoint cliques
  • 106. 41 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Approaches Spectral Clustering • Eigenvectors of Laplacian of similarity graph Simple implementation Laplacian could be ill-conditioned # of clusters need to be known in advance [Ng, Jordan, White, Smyth, Weiss, Shi, Malik, Kannan, Vempala, Vetta, Meila, · · · ] Semi-Definite Programming (SDP) • LP relaxation Asymptotically optimal for Stochastic Block Models Implicit assumption on generative model # of clusters need to be known in advance [Abbe, Sandon, Hajeck, Bandeira, Hall, Decelle, Mossel, Neeman, Sly, Rao, Chen, Wu, Xu, · · · ]
  • 107. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Core Idea a b c d e f g h i Similarity graph.
  • 108. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Core Idea a b c d e f g h i Similarity graph. Define random walk.
  • 109. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Core Idea a b c d e f g h i Non-uniform Probability of following a link ∝ # of common neighbors Similarity graph. Define random walk. Couple random walks starting from different nodes
  • 110. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Core Idea a b c d e f g h i Similarity graph. Define random walk. Couple random walks starting from different nodes Adapt CFTP algorithm
  • 111. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Core Idea a b c d e f g h i Similarity graph. Define random walk. Couple random walks starting from different nodes Adapt CFTP algorithm Do not care about exact sample
  • 112. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Core Idea a b c d e f g h i Similarity graph. Define random walk. Couple random walks starting from different nodes Adapt CFTP algorithm Do not care about exact sample Identify clusters “before coalescence happens”
  • 113. 42 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Core Idea a b c d e f g h i Restricted walk within a cluster mixes faster. Similarity graph. Define random walk. Couple random walks starting from different nodes Adapt CFTP algorithm Do not care about exact sample Identify clusters “before coalescence happens”
  • 114. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Algorithm Overview S
  • 115. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Algorithm Overview S Initially: All singletons
  • 116. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Algorithm Overview S Partial coalescence Small Clusters formation
  • 117. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Algorithm Overview S Critical times Clusters merge
  • 118. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Algorithm Overview S Critical times Clusters merge
  • 119. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Algorithm Overview S Critical times Clusters merge
  • 120. 43 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Algorithm Overview S Full coalescence
  • 121. 44 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Partial Coalescence time S G Gc -10 -9 -8 -7 -6 -5 -4 -3 -2 -1 0 Paths starting from G ⊂ S
  • 122. 45 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Remarks Algorithm can stop at critical times Still yield a cluster Choose cluster with optimal cost Purely based on random walk Use of mixing properties Circumvent issues with ill-conditioned matrices in spectral based approaches No prior assumption on generative model No prior assumption on number of clusters
  • 123. 46 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Benchmark Networks Stochastic Block Models (SBM) Size of communities equal Average degree equal for all nodes LFR Models More realistic # of communities and sizes admit power law. Real world networks
  • 124. 47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Stochastic Block Model
  • 125. 47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Stochastic Block Model Ber(p)
  • 126. 47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Stochastic Block Model Ber(p) Ber(q)
  • 127. 47 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Stochastic Block Model Ber(p) Ber(q) p q
  • 128. 48 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally SBM Realization 500 nodes p = 0.5 q = 0.1 5 communities Randomly permuted.
  • 129. 49 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Identifying communities
  • 130. 50 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally CC-PIVOT Output
  • 131. 50 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally LFR Model 200 nodes 6 communities
  • 132. 51 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Identifying communities
  • 133. 52 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally CC-PIVOT output
  • 134. 53 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally American College Football 115 teams. Divided into conferences.
  • 135. 54 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Identified communities
  • 136. 55 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally CC-PIVOT Output
  • 137. 56 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Future Directions Open: Theoretical guarantees for recovery in SBMs Open: Stationary is more difficult (needs twice summability of d) than transition (just summability of d) Theoretical foundation of the proposed algorithm Extension to broader set of community detection problems
  • 139. 58 Intro Challenges Formal Transition Stationary Sampling Community Algorithm Finally Acknowledgement My Advisor: Prasad Committee Members Dr. Rui Zhang Office Mates: • Maryam and Meysam My Friends: • Masoud, Saeed, Navid, Elyas, Harir Chee, Reza, Ehsaneh, Ali, Ashkan, Hamed, Ehsan, Seyed, Alireza,... My Family: • My parents, my sister Nasrin and my brother Naser