SlideShare a Scribd company logo
1 of 41
Download to read offline
aalto-logo-en-3
Graphical Model Selection for Big Data over
Networks
Alexander Jung
Department of Computer Science
Aalto University, School of Science
20.12.2016
1 / 41
aalto-logo-en-3
Outline
1 Big Data over Networks
2 Probabilistic Graphical Models
3 GMS via Sparse Neighbourhood Regression
IID Training
Non-IID Training
4 Efficient GMS via Convex Optimization
2 / 41
aalto-logo-en-3
Data is New Oil
3 / 41
aalto-logo-en-3
Data, Data Everywhere
Atacama Large Millimeter Array yields hundreds of TB (1012
bytes)/year
CERN experiment generates data at rate of PB (1015
bytes)/second
annual internet traffic crushed ZB (1021 bytes) borderline
Big Data challenge: how to process data at internet scale?
4 / 41
aalto-logo-en-3
The Three V’s of Big Data
Volume: deal with Zettabyte (= 1021 bytes) scale data
Velocity: CERN experiment generates Petabytes (= 1015
bytes) per second
Variety: heterogeneous data (text, audio, video...)
5 / 41
aalto-logo-en-3
Big Data over Networks
often the datasets have an intrinsic network structure
chip design internet bioinformatics
social networks universe material science
L. Lov´asz, “Large Networks and Graph Limits”
6 / 41
aalto-logo-en-3
Big Data over Networks
observe dataset D={z1, . . . , zN} with datapoints zi
a particular zi might be audio, video or text data
datapoints are structured by some notion of “similarity”
e.g., zi , zj similar if from the same user account
represent datapoint zi by node i ∈ V of graph G = (V, E)
nodes i, j ∈ V connected by edge (i, j), if zi is similar to zj
7 / 41
aalto-logo-en-3
Graph Signals for Supervised Learning
consider supervised machine learning from dataset D
each datapoint zi has a label r[i] (e.g., persons preference for
buying certain product)
entire labelling is a graph signal r[·] : V → R
graph signal r[·] maps node i ∈ V to its label r[i]
graph signal processing (GSP) provides efficient tools for
handling large-scale graph signals
8 / 41
aalto-logo-en-3
GSP generalizes DSP
discrete-time signals are graph signals over chain graph
· · ·· · ·
r[−1] r[0] r[1]
label r[i] might correspond to presence of “clipping” at time i
(greyscale) images are signals over grid graph
· · ·
···
r[i]
label r[i] might be presence of certain object at location i
GSP yields efficient learning by fast graph Fourier transforms
9 / 41
aalto-logo-en-3
Fast Algorithms on Graphs
GSP theory yields fast algorithms for large-scale graphs
generalizes FFT from chain graph to general graphs
parallelization/vectorization possible for product graph
structure
6
graph product, denoted as G = G1 × G2,
is
= A1 ⊗ IN2 + IN1 ⊗A2. (25)
g product, denoted as G = G1 G2, the
⊗ A2 + A1 ⊗ IN2 + IN1 ⊗A2. (26)
can be seen as a combination of the Kro-
n products. Since the products (23), (25),
ve, Kronecker, Cartesian, and strong graph
ned for an arbitrary number of graphs.
rise in different applications, including
processing [32], computational sciences
], and computational biology [34]. Their
parts are used in network modeling and
], [37]. Multiple approaches have been
composition and approximation of graphs
[38], [30], [31], [39].
er a versatile graph model for the represen-
atasets in multi-level and multi-parameter
DSP, multi-dimensional signals, such as
video, reside on rectangular lattices that
ts of line graphs. Fig. 2(a) shows a two-
formed by the Cartesian product of two
ces.
of graph signals residing on product graphs
Digital image Row Column
(a)
Sensor network Time series
Measurements
atonetimestep
Measurements of one sensor
Sensor network measurements
(b)
Social network with communities
Community
structure
Intercommunity
communication
structure
(c)
10 / 41
aalto-logo-en-3
Graph Models: Perfect Match for 3 Vs of Big Data
graph based models lead to message passing algorithms
message passing algorithms are perfectly scalable
copes with volume (distributed computing) and velocity
(parallel computing) of big data
ship computation to data and not vice-versa!
graph-based models allow to process heterogeneous data
11 / 41
aalto-logo-en-3
Semi-Supervised Learning for Big Data over Networks
consider graph signal x[i] representing big dataset D
we only observe labels at sampling set S ⊆ V
aquiring labels can be costly
central hypothesis of supervised learning
close-by data points in high-density regions have similar labels
12 / 41
aalto-logo-en-3
Aquiring Labels (Sampling) in Marine Biology
13 / 41
aalto-logo-en-3
Aquiring Labels (Sampling) in Particle Physics
14 / 41
aalto-logo-en-3
Aquiring Labels (Sampling) in Pharmacology
15 / 41
aalto-logo-en-3
Key Problems for Learning from Big Data over Networks
given a graph signal representation of the learning problem:
how many labels (samples) do we need?
which nodes should we sample ?
what are efficient learning algorithms ?
all this presupposes that we know the graph structure!
16 / 41
aalto-logo-en-3
Outline
1 Big Data over Networks
2 Probabilistic Graphical Models
3 GMS via Sparse Neighbourhood Regression
IID Training
Non-IID Training
4 Efficient GMS via Convex Optimization
17 / 41
aalto-logo-en-3
Conditional Independence Graph
interpret ith data point as realization of random variable zi
associate zi with node i ∈ V of undirected graph G = (V, E)
connect i, j ∈ V by undirected edge if zi and zj conditionally
independent given {zl }l∈V{i,j}
G is the conditional independence graph (CIG) of {zi }p
i=1
18 / 41
aalto-logo-en-3
Modelling Assumptions
assume the zj to be jointly Gaussian N(0, C) with C 0
edges of CIG G characterized by non-zero entries of K = C−1:
{i, j} ∈ E ⇐⇒ Ki,j = 0
define neighbourhood N(i) := {j ∈ V : {i, j} ∈ E} of node i
we assume CIG to be sparse, i.e.,
|N(i)| ≤ smax
for some fixed sparsity smax
19 / 41
aalto-logo-en-3
The Global Markov Property
CIG G allows to read off conditional independence relations
A B C
node set B separates A from C.
global Markov property: {zi }i∈A conditionally independent of
{Zi }i∈C given {Zi }i∈B
Bayes optimal estimator for Zi depends only on {Zj }j∈N(i)
for sparse graphs, i.e., |N(i)| p, Bayes optimal estimator
can be implemented efficiently!
20 / 41
aalto-logo-en-3
The Training Data
for each datapoint Zi , observe training samples
xi [1], . . . , xi [N]
stack samples into vector x[t] = (x1[t], . . . , xp[t])T
each vector sample x[t] multivariate normal ∼ N(0, C[t])
graphical model selection (GMS): learn G from x[t]
x1[1], . . . , x1[N]
x2[1], . . . , x2[N]
x3[1], . . . , x3[N]
21 / 41
aalto-logo-en-3
Outline
1 Big Data over Networks
2 Probabilistic Graphical Models
3 GMS via Sparse Neighbourhood Regression
IID Training
Non-IID Training
4 Efficient GMS via Convex Optimization
22 / 41
aalto-logo-en-3
Outline
1 Big Data over Networks
2 Probabilistic Graphical Models
3 GMS via Sparse Neighbourhood Regression
IID Training
Non-IID Training
4 Efficient GMS via Convex Optimization
23 / 41
aalto-logo-en-3
The Idea of Neighbourhood Regression
consider the problem of finding the neighbourhood N(i)
once we found all neighbourhoods, we have the CIG (trivial!)
neighbourhood N(i) pops up in regression model for zi :
zi =
j∈N(i)
aj zj + ei
regression coeffs aj = −Ki,j /Ki,i
error term ei uncorrelated with {zj }j∈V{i}
neighbourhood N(i) solves sparse regression problem
N(i) = arg min
| supp(a)|≤smax
E{(zi −
j∈supp(a)
aj zj )2
}
24 / 41
aalto-logo-en-3
Learning based on Sparse Regression
neighbourhood N(i) characterized by
N(i) = arg min
| supp(a)|≤smax
E{(zi −
j∈supp(a)
aj zj )2
}
not implementable since cannot evaluate expectation E
consider i.i.d. training x[t] ∼ N(0, C)
replace E with sample average
N(i) ≈ arg min
| supp(a)|≤smax
(1/N)
N
t=1
(xi [t] −
j∈supp(a)
aj xj [t])2
(1)
involves search over p
smax
subsets (more on this later!)
25 / 41
aalto-logo-en-3
The Test Statistic of Sparse Regression
in what follows assume |N(i)| = smax
for index set I = {i1, . . . , ismax }, define statistic
Z(I) := arg min
| supp(a)|⊆I
(1/N)
N
t=1
(xi [t] −
j∈supp(a)
aj xj [t])2
= (1/N) P⊥
I xi
2
2
with the vector xi = (xi [1], . . . , xi [N])T
projection P⊥
I on orthogonal complement of span{xj }j∈I
P⊥
I := I − XI XT
I XI
−1
XT
I
26 / 41
aalto-logo-en-3
Pairwise Error
consider index set T with |T | ≤ smax and N(i)  T = ∅
i ∈ V
N(i) T
with high probability Z(N(i)) < Z(T )
Z(I)
prob(Z(I))
I =N (i) I =T
27 / 41
aalto-logo-en-3
Pairwise Error
assume samples x[t] i.i.d. with ∼ N(0, C)
edges of CIG correspond to non-zero locations in K = C−1
define condition number κ := λmin(C)/λmax(C)
strength of edge (i, j) quantified by partial correlation
ρi,j := |Ki,j |/ Ki,i Kj,j
minimum partial correlation ρmin ≤ ρi,j
the probability of confusing T with N(i) is bounded as
Prob{Z(N(i)) ≥ Z(T )} ≤ 4 exp −(N−smax)
ρ2
minκ
64(ρ2
minκ + 8)
28 / 41
aalto-logo-en-3
Sample Complexity of GMS
by union bound over all p
smax
subsets T , we obtain
Prob(N(i) = N(i))≤c1 exp smax log p−(N−smax)
ρ2
minκ
64(ρ2
minκ+8)
another union bound over all nodes i yields
Prob(N(i) = N(i))≤c1 exp 2smax log p−(N−smax)
ρ2
minκ
64(ρ2
minκ+8)
accurate GMS for sample size
N ≥ c1 max log
p−smax
smax
,
log(p−smax)
κρ2
min
allows for high-dimensional regime: large p, small N
29 / 41
aalto-logo-en-3
How Low Can We Go?
what is the fundamental lower bound on sample size N?
it can be shown that for sample size
N ≤ c2
log(p−smax)
κρ2
min
ANY GMS fails with non-negligible probability
thus, sample complexity of GMS is
N ∝
log(p−smax)
κρ2
min
sample complexity driven by minimum partial correlation ρmin
30 / 41
aalto-logo-en-3
Outline
1 Big Data over Networks
2 Probabilistic Graphical Models
3 GMS via Sparse Neighbourhood Regression
IID Training
Non-IID Training
4 Efficient GMS via Convex Optimization
31 / 41
aalto-logo-en-3
Non-IID Samples
samples x[f ]∼N(0, C[f ]) still independent but varying C[f ]
model useful for two particular process classes:
x[f ] are Fourier coeffs of stationary process y[t]:
x[f ] =
N
t=1
y[t] exp(−(2π/N)(t−1)(f −1))
x[f ] are LCB coeffs of locally stationary process y[t]:
x[f ] =
N
t=1
y[t]g(f )
[t]
32 / 41
aalto-logo-en-3
Sample Complexity for Non-IID Samples
RECALL sample complexity for iid case:
N ∝
log(p−smax)
κρ2
min
replace ρmin with minimum average partial correlation
¯ρmin := min
(i,j)∈E
(1/N)
N
f =1
K2
i,j [f ]/(Ki,i [f ]Kj,j [f ])
with K[f ] := C−1[f ]
sample complex for non-iid case:
N ∝
log(p−smax)
κ¯ρ2
min
33 / 41
aalto-logo-en-3
Average Partial Correlation
N
¯ρ2
min
f
K2
i,j [f]
Ki,i[f]Kj,j [f]
34 / 41
aalto-logo-en-3
Outline
1 Big Data over Networks
2 Probabilistic Graphical Models
3 GMS via Sparse Neighbourhood Regression
IID Training
Non-IID Training
4 Efficient GMS via Convex Optimization
35 / 41
aalto-logo-en-3
Just Relax
consider (vectorized) sparse neighbourhood regression
N(i) ≈ arg min
| supp(a)|≤smax
xi −
j∈supp(a)
aj xj
2
2
= arg min
a
xi − Xa 2
2 + λ| supp(a)|
with X := {xj }j=i and some multiplier λ
involves intractable search over p
smax
supports supp(a)
RELAX penalty λ| supp(a)| with convex function λ a 1
36 / 41
aalto-logo-en-3
The LASSO
convex relaxation of sparse neighbourhood regression
aLasso ∈ arg min
a
xi −Xaj
2
2 + λ a 1
for some suitable multiplier λ
convex optimization problem
(Lagrangian form) least absolute shrinkage and selection
operator (LASSO)
LASSO found widespread use in statistics/signal processing
37 / 41
aalto-logo-en-3
Sample Complexity of LASSO
consider LASSO estimator
aLasso ∈ arg min
a
xi −Xaj
2
2 + λ a 1
assume that zi are only weakly correlated
|E{zi zj }|/ E{z2
i }E{z2
j } ≤ /(2smax)
we have supp(aLasso) = N(i) with high prob. if
N ≥ (c1smax + c2/(κρ2
min)) log(p−smax)
38 / 41
aalto-logo-en-3
Structure of LASSO
consider LASSO
arg min
a
xi − Xa 2
2 + λ a 1
sum of smooth term xi − Xa 2
2 and non-smooth a 1
highly developed methods for such optimization problems
e.g., proximal gradient method, alternating direction method
of multipliers (ADMM), Pock-Chambolle primal-dual, etc....
basic idea: splitting of minimization of the two terms
39 / 41
aalto-logo-en-3
ADMM for LASSO
LASSO: aLasso ∈ arg mina=b xi − Xa 2
2 + λ b 1
ADMM amounts to iterating
a(k+1)
= XT
X + ρI
−1
XT
xi + ρ(b(k)
− c(k)
)
b(k+1)
= Sλ/ρ(a(k+1)
+ c(k)
)
c(k+1)
= c(k)
+ a(k+1)
− b(k+1)
a
Sκ(a)
κ
κ
2κ
convergence under very general conditions
40 / 41
aalto-logo-en-3
Future Directions
consider directed graphical models (Bayesian networks)
analyze intrinsic computational complexity of GMS (is group
LASSO computationally efficient)
extend approach to vector-valued data points Zi ∈ Rp
incorporate additional structure on top of sparsity, e.g.,
product graphs
41 / 41

More Related Content

What's hot

D I G I T A L I C A P P L I C A T I O N S J N T U M O D E L P A P E R{Www
D I G I T A L  I C  A P P L I C A T I O N S  J N T U  M O D E L  P A P E R{WwwD I G I T A L  I C  A P P L I C A T I O N S  J N T U  M O D E L  P A P E R{Www
D I G I T A L I C A P P L I C A T I O N S J N T U M O D E L P A P E R{Wwwguest3f9c6b
 
Injecting image priors into Learnable Compressive Subsampling
Injecting image priors into Learnable Compressive SubsamplingInjecting image priors into Learnable Compressive Subsampling
Injecting image priors into Learnable Compressive SubsamplingMartino Ferrari
 
Implementation of Efficiency CORDIC Algorithmfor Sine & Cosine Generation
Implementation of Efficiency CORDIC Algorithmfor Sine & Cosine GenerationImplementation of Efficiency CORDIC Algorithmfor Sine & Cosine Generation
Implementation of Efficiency CORDIC Algorithmfor Sine & Cosine GenerationIOSR Journals
 
Performance analysis of transmission of 5 users based on model b using gf (5)...
Performance analysis of transmission of 5 users based on model b using gf (5)...Performance analysis of transmission of 5 users based on model b using gf (5)...
Performance analysis of transmission of 5 users based on model b using gf (5)...csandit
 
Lecture 3 image sampling and quantization
Lecture 3 image sampling and quantizationLecture 3 image sampling and quantization
Lecture 3 image sampling and quantizationVARUN KUMAR
 
GATE Computer Science Solved Paper 2004
GATE Computer Science Solved Paper 2004GATE Computer Science Solved Paper 2004
GATE Computer Science Solved Paper 2004Rohit Garg
 
Implementation of Energy Efficient Scalar Point Multiplication Techniques for...
Implementation of Energy Efficient Scalar Point Multiplication Techniques for...Implementation of Energy Efficient Scalar Point Multiplication Techniques for...
Implementation of Energy Efficient Scalar Point Multiplication Techniques for...idescitation
 
Point Cloud Compression in MPEG
Point Cloud Compression in MPEGPoint Cloud Compression in MPEG
Point Cloud Compression in MPEGMarius Preda PhD
 
Parallel algorithm for computing edt with new architecture
Parallel algorithm for computing edt with new architectureParallel algorithm for computing edt with new architecture
Parallel algorithm for computing edt with new architectureIAEME Publication
 
Predictive coding
Predictive codingPredictive coding
Predictive codingp_ayal
 
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...IJERA Editor
 
VHDL and Cordic Algorithim
VHDL and Cordic AlgorithimVHDL and Cordic Algorithim
VHDL and Cordic AlgorithimSubeer Rangra
 
Power and Delay Analysis of Logic Circuits Using Reversible Gates
Power and Delay Analysis of Logic Circuits Using Reversible GatesPower and Delay Analysis of Logic Circuits Using Reversible Gates
Power and Delay Analysis of Logic Circuits Using Reversible GatesRSIS International
 

What's hot (18)

D I G I T A L I C A P P L I C A T I O N S J N T U M O D E L P A P E R{Www
D I G I T A L  I C  A P P L I C A T I O N S  J N T U  M O D E L  P A P E R{WwwD I G I T A L  I C  A P P L I C A T I O N S  J N T U  M O D E L  P A P E R{Www
D I G I T A L I C A P P L I C A T I O N S J N T U M O D E L P A P E R{Www
 
Injecting image priors into Learnable Compressive Subsampling
Injecting image priors into Learnable Compressive SubsamplingInjecting image priors into Learnable Compressive Subsampling
Injecting image priors into Learnable Compressive Subsampling
 
Implementation of Efficiency CORDIC Algorithmfor Sine & Cosine Generation
Implementation of Efficiency CORDIC Algorithmfor Sine & Cosine GenerationImplementation of Efficiency CORDIC Algorithmfor Sine & Cosine Generation
Implementation of Efficiency CORDIC Algorithmfor Sine & Cosine Generation
 
Performance analysis of transmission of 5 users based on model b using gf (5)...
Performance analysis of transmission of 5 users based on model b using gf (5)...Performance analysis of transmission of 5 users based on model b using gf (5)...
Performance analysis of transmission of 5 users based on model b using gf (5)...
 
An35225228
An35225228An35225228
An35225228
 
Lecture 3 image sampling and quantization
Lecture 3 image sampling and quantizationLecture 3 image sampling and quantization
Lecture 3 image sampling and quantization
 
GATE Computer Science Solved Paper 2004
GATE Computer Science Solved Paper 2004GATE Computer Science Solved Paper 2004
GATE Computer Science Solved Paper 2004
 
Compression Ii
Compression IiCompression Ii
Compression Ii
 
Implementation of Energy Efficient Scalar Point Multiplication Techniques for...
Implementation of Energy Efficient Scalar Point Multiplication Techniques for...Implementation of Energy Efficient Scalar Point Multiplication Techniques for...
Implementation of Energy Efficient Scalar Point Multiplication Techniques for...
 
ICRA Nathan Piasco
ICRA Nathan PiascoICRA Nathan Piasco
ICRA Nathan Piasco
 
Cordic Code
Cordic CodeCordic Code
Cordic Code
 
Point Cloud Compression in MPEG
Point Cloud Compression in MPEGPoint Cloud Compression in MPEG
Point Cloud Compression in MPEG
 
Review_Cibe Sridharan
Review_Cibe SridharanReview_Cibe Sridharan
Review_Cibe Sridharan
 
Parallel algorithm for computing edt with new architecture
Parallel algorithm for computing edt with new architectureParallel algorithm for computing edt with new architecture
Parallel algorithm for computing edt with new architecture
 
Predictive coding
Predictive codingPredictive coding
Predictive coding
 
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...
Image Restoration and Denoising By Using Nonlocally Centralized Sparse Repres...
 
VHDL and Cordic Algorithim
VHDL and Cordic AlgorithimVHDL and Cordic Algorithim
VHDL and Cordic Algorithim
 
Power and Delay Analysis of Logic Circuits Using Reversible Gates
Power and Delay Analysis of Logic Circuits Using Reversible GatesPower and Delay Analysis of Logic Circuits Using Reversible Gates
Power and Delay Analysis of Logic Circuits Using Reversible Gates
 

Similar to Graphical Model Selection for Big Data

VJAI Paper Reading#3-KDD2019-ClusterGCN
VJAI Paper Reading#3-KDD2019-ClusterGCNVJAI Paper Reading#3-KDD2019-ClusterGCN
VJAI Paper Reading#3-KDD2019-ClusterGCNDat Nguyen
 
Paper Summary of Infogan-CR : Disentangling Generative Adversarial Networks w...
Paper Summary of Infogan-CR : Disentangling Generative Adversarial Networks w...Paper Summary of Infogan-CR : Disentangling Generative Adversarial Networks w...
Paper Summary of Infogan-CR : Disentangling Generative Adversarial Networks w...준식 최
 
DAOR - Bridging the Gap between Community and Node Representations: Graph Emb...
DAOR - Bridging the Gap between Community and Node Representations: Graph Emb...DAOR - Bridging the Gap between Community and Node Representations: Graph Emb...
DAOR - Bridging the Gap between Community and Node Representations: Graph Emb...Artem Lutov
 
Information Content of Complex Networks
Information Content of Complex NetworksInformation Content of Complex Networks
Information Content of Complex NetworksHector Zenil
 
parameterized complexity for graph Motif
parameterized complexity for graph Motifparameterized complexity for graph Motif
parameterized complexity for graph MotifAMR koura
 
FPGA based BCH Decoder
FPGA based BCH DecoderFPGA based BCH Decoder
FPGA based BCH Decoderijsrd.com
 
Design and Implementation of Encoder for (15, k) Binary BCH Code Using VHDL a...
Design and Implementation of Encoder for (15, k) Binary BCH Code Using VHDL a...Design and Implementation of Encoder for (15, k) Binary BCH Code Using VHDL a...
Design and Implementation of Encoder for (15, k) Binary BCH Code Using VHDL a...IOSR Journals
 
An Analysis of Graph Cut Size for Transductive Learning
An Analysis of Graph Cut Size for Transductive LearningAn Analysis of Graph Cut Size for Transductive Learning
An Analysis of Graph Cut Size for Transductive Learningbutest
 
Scalable Global Alignment Graph Kernel Using Random Features: From Node Embed...
Scalable Global Alignment Graph Kernel Using Random Features: From Node Embed...Scalable Global Alignment Graph Kernel Using Random Features: From Node Embed...
Scalable Global Alignment Graph Kernel Using Random Features: From Node Embed...seijihagawa
 
GASGD: Stochastic Gradient Descent for Distributed Asynchronous Matrix Comple...
GASGD: Stochastic Gradient Descent for Distributed Asynchronous Matrix Comple...GASGD: Stochastic Gradient Descent for Distributed Asynchronous Matrix Comple...
GASGD: Stochastic Gradient Descent for Distributed Asynchronous Matrix Comple...Fabio Petroni, PhD
 
A simple framework for contrastive learning of visual representations
A simple framework for contrastive learning of visual representationsA simple framework for contrastive learning of visual representations
A simple framework for contrastive learning of visual representationsDevansh16
 
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...Matt Moores
 
Paired-end alignments in sequence graphs
Paired-end alignments in sequence graphsPaired-end alignments in sequence graphs
Paired-end alignments in sequence graphsChirag Jain
 
Overlap Layout Consensus assembly
Overlap Layout Consensus assemblyOverlap Layout Consensus assembly
Overlap Layout Consensus assemblyZhuyi Xue
 
論文紹介:Towards Robust Adaptive Object Detection Under Noisy Annotations
論文紹介:Towards Robust Adaptive Object Detection Under Noisy Annotations論文紹介:Towards Robust Adaptive Object Detection Under Noisy Annotations
論文紹介:Towards Robust Adaptive Object Detection Under Noisy AnnotationsToru Tamaki
 
DISTANCE TWO LABELING FOR MULTI-STOREY GRAPHS
DISTANCE TWO LABELING FOR MULTI-STOREY GRAPHSDISTANCE TWO LABELING FOR MULTI-STOREY GRAPHS
DISTANCE TWO LABELING FOR MULTI-STOREY GRAPHSgraphhoc
 
DLD BOOLEAN EXPRESSIONS
DLD BOOLEAN EXPRESSIONSDLD BOOLEAN EXPRESSIONS
DLD BOOLEAN EXPRESSIONSnaresh414857
 

Similar to Graphical Model Selection for Big Data (20)

VJAI Paper Reading#3-KDD2019-ClusterGCN
VJAI Paper Reading#3-KDD2019-ClusterGCNVJAI Paper Reading#3-KDD2019-ClusterGCN
VJAI Paper Reading#3-KDD2019-ClusterGCN
 
Paper Summary of Infogan-CR : Disentangling Generative Adversarial Networks w...
Paper Summary of Infogan-CR : Disentangling Generative Adversarial Networks w...Paper Summary of Infogan-CR : Disentangling Generative Adversarial Networks w...
Paper Summary of Infogan-CR : Disentangling Generative Adversarial Networks w...
 
DAOR - Bridging the Gap between Community and Node Representations: Graph Emb...
DAOR - Bridging the Gap between Community and Node Representations: Graph Emb...DAOR - Bridging the Gap between Community and Node Representations: Graph Emb...
DAOR - Bridging the Gap between Community and Node Representations: Graph Emb...
 
Information Content of Complex Networks
Information Content of Complex NetworksInformation Content of Complex Networks
Information Content of Complex Networks
 
parameterized complexity for graph Motif
parameterized complexity for graph Motifparameterized complexity for graph Motif
parameterized complexity for graph Motif
 
FPGA based BCH Decoder
FPGA based BCH DecoderFPGA based BCH Decoder
FPGA based BCH Decoder
 
Design and Implementation of Encoder for (15, k) Binary BCH Code Using VHDL a...
Design and Implementation of Encoder for (15, k) Binary BCH Code Using VHDL a...Design and Implementation of Encoder for (15, k) Binary BCH Code Using VHDL a...
Design and Implementation of Encoder for (15, k) Binary BCH Code Using VHDL a...
 
An Analysis of Graph Cut Size for Transductive Learning
An Analysis of Graph Cut Size for Transductive LearningAn Analysis of Graph Cut Size for Transductive Learning
An Analysis of Graph Cut Size for Transductive Learning
 
Scalable Global Alignment Graph Kernel Using Random Features: From Node Embed...
Scalable Global Alignment Graph Kernel Using Random Features: From Node Embed...Scalable Global Alignment Graph Kernel Using Random Features: From Node Embed...
Scalable Global Alignment Graph Kernel Using Random Features: From Node Embed...
 
GASGD: Stochastic Gradient Descent for Distributed Asynchronous Matrix Comple...
GASGD: Stochastic Gradient Descent for Distributed Asynchronous Matrix Comple...GASGD: Stochastic Gradient Descent for Distributed Asynchronous Matrix Comple...
GASGD: Stochastic Gradient Descent for Distributed Asynchronous Matrix Comple...
 
A simple framework for contrastive learning of visual representations
A simple framework for contrastive learning of visual representationsA simple framework for contrastive learning of visual representations
A simple framework for contrastive learning of visual representations
 
Ivd soda-2019
Ivd soda-2019Ivd soda-2019
Ivd soda-2019
 
Subquad multi ff
Subquad multi ffSubquad multi ff
Subquad multi ff
 
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...
R package 'bayesImageS': a case study in Bayesian computation using Rcpp and ...
 
Lec 2-2
Lec 2-2Lec 2-2
Lec 2-2
 
Paired-end alignments in sequence graphs
Paired-end alignments in sequence graphsPaired-end alignments in sequence graphs
Paired-end alignments in sequence graphs
 
Overlap Layout Consensus assembly
Overlap Layout Consensus assemblyOverlap Layout Consensus assembly
Overlap Layout Consensus assembly
 
論文紹介:Towards Robust Adaptive Object Detection Under Noisy Annotations
論文紹介:Towards Robust Adaptive Object Detection Under Noisy Annotations論文紹介:Towards Robust Adaptive Object Detection Under Noisy Annotations
論文紹介:Towards Robust Adaptive Object Detection Under Noisy Annotations
 
DISTANCE TWO LABELING FOR MULTI-STOREY GRAPHS
DISTANCE TWO LABELING FOR MULTI-STOREY GRAPHSDISTANCE TWO LABELING FOR MULTI-STOREY GRAPHS
DISTANCE TWO LABELING FOR MULTI-STOREY GRAPHS
 
DLD BOOLEAN EXPRESSIONS
DLD BOOLEAN EXPRESSIONSDLD BOOLEAN EXPRESSIONS
DLD BOOLEAN EXPRESSIONS
 

Recently uploaded

High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...soniya singh
 
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...Sapana Sha
 
Beautiful Sapna Vip Call Girls Hauz Khas 9711199012 Call /Whatsapps
Beautiful Sapna Vip  Call Girls Hauz Khas 9711199012 Call /WhatsappsBeautiful Sapna Vip  Call Girls Hauz Khas 9711199012 Call /Whatsapps
Beautiful Sapna Vip Call Girls Hauz Khas 9711199012 Call /Whatsappssapnasaifi408
 
ASML's Taxonomy Adventure by Daniel Canter
ASML's Taxonomy Adventure by Daniel CanterASML's Taxonomy Adventure by Daniel Canter
ASML's Taxonomy Adventure by Daniel Cantervoginip
 
RS 9000 Call In girls Dwarka Mor (DELHI)⇛9711147426🔝Delhi
RS 9000 Call In girls Dwarka Mor (DELHI)⇛9711147426🔝DelhiRS 9000 Call In girls Dwarka Mor (DELHI)⇛9711147426🔝Delhi
RS 9000 Call In girls Dwarka Mor (DELHI)⇛9711147426🔝Delhijennyeacort
 
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一F La
 
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样vhwb25kk
 
How we prevented account sharing with MFA
How we prevented account sharing with MFAHow we prevented account sharing with MFA
How we prevented account sharing with MFAAndrei Kaleshka
 
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptx
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptxAmazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptx
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptxAbdelrhman abooda
 
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptdokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptSonatrach
 
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptxEMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptxthyngster
 
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130Suhani Kapoor
 
科罗拉多大学波尔得分校毕业证学位证成绩单-可办理
科罗拉多大学波尔得分校毕业证学位证成绩单-可办理科罗拉多大学波尔得分校毕业证学位证成绩单-可办理
科罗拉多大学波尔得分校毕业证学位证成绩单-可办理e4aez8ss
 
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...Jack DiGiovanna
 
Consent & Privacy Signals on Google *Pixels* - MeasureCamp Amsterdam 2024
Consent & Privacy Signals on Google *Pixels* - MeasureCamp Amsterdam 2024Consent & Privacy Signals on Google *Pixels* - MeasureCamp Amsterdam 2024
Consent & Privacy Signals on Google *Pixels* - MeasureCamp Amsterdam 2024thyngster
 
Dubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls DubaiDubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls Dubaihf8803863
 
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Callshivangimorya083
 
Call Girls In Mahipalpur O9654467111 Escorts Service
Call Girls In Mahipalpur O9654467111  Escorts ServiceCall Girls In Mahipalpur O9654467111  Escorts Service
Call Girls In Mahipalpur O9654467111 Escorts ServiceSapana Sha
 
From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...Florian Roscheck
 

Recently uploaded (20)

High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
High Class Call Girls Noida Sector 39 Aarushi 🔝8264348440🔝 Independent Escort...
 
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
Saket, (-DELHI )+91-9654467111-(=)CHEAP Call Girls in Escorts Service Saket C...
 
Beautiful Sapna Vip Call Girls Hauz Khas 9711199012 Call /Whatsapps
Beautiful Sapna Vip  Call Girls Hauz Khas 9711199012 Call /WhatsappsBeautiful Sapna Vip  Call Girls Hauz Khas 9711199012 Call /Whatsapps
Beautiful Sapna Vip Call Girls Hauz Khas 9711199012 Call /Whatsapps
 
ASML's Taxonomy Adventure by Daniel Canter
ASML's Taxonomy Adventure by Daniel CanterASML's Taxonomy Adventure by Daniel Canter
ASML's Taxonomy Adventure by Daniel Canter
 
RS 9000 Call In girls Dwarka Mor (DELHI)⇛9711147426🔝Delhi
RS 9000 Call In girls Dwarka Mor (DELHI)⇛9711147426🔝DelhiRS 9000 Call In girls Dwarka Mor (DELHI)⇛9711147426🔝Delhi
RS 9000 Call In girls Dwarka Mor (DELHI)⇛9711147426🔝Delhi
 
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一
办理(Vancouver毕业证书)加拿大温哥华岛大学毕业证成绩单原版一比一
 
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
1:1定制(UQ毕业证)昆士兰大学毕业证成绩单修改留信学历认证原版一模一样
 
How we prevented account sharing with MFA
How we prevented account sharing with MFAHow we prevented account sharing with MFA
How we prevented account sharing with MFA
 
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptx
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptxAmazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptx
Amazon TQM (2) Amazon TQM (2)Amazon TQM (2).pptx
 
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.pptdokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
dokumen.tips_chapter-4-transient-heat-conduction-mehmet-kanoglu.ppt
 
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptxEMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM  TRACKING WITH GOOGLE ANALYTICS.pptx
EMERCE - 2024 - AMSTERDAM - CROSS-PLATFORM TRACKING WITH GOOGLE ANALYTICS.pptx
 
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
VIP Call Girls Service Miyapur Hyderabad Call +91-8250192130
 
科罗拉多大学波尔得分校毕业证学位证成绩单-可办理
科罗拉多大学波尔得分校毕业证学位证成绩单-可办理科罗拉多大学波尔得分校毕业证学位证成绩单-可办理
科罗拉多大学波尔得分校毕业证学位证成绩单-可办理
 
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...
Building on a FAIRly Strong Foundation to Connect Academic Research to Transl...
 
Consent & Privacy Signals on Google *Pixels* - MeasureCamp Amsterdam 2024
Consent & Privacy Signals on Google *Pixels* - MeasureCamp Amsterdam 2024Consent & Privacy Signals on Google *Pixels* - MeasureCamp Amsterdam 2024
Consent & Privacy Signals on Google *Pixels* - MeasureCamp Amsterdam 2024
 
Dubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls DubaiDubai Call Girls Wifey O52&786472 Call Girls Dubai
Dubai Call Girls Wifey O52&786472 Call Girls Dubai
 
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
꧁❤ Greater Noida Call Girls Delhi ❤꧂ 9711199171 ☎️ Hard And Sexy Vip Call
 
Call Girls In Mahipalpur O9654467111 Escorts Service
Call Girls In Mahipalpur O9654467111  Escorts ServiceCall Girls In Mahipalpur O9654467111  Escorts Service
Call Girls In Mahipalpur O9654467111 Escorts Service
 
Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
Deep Generative Learning for All - The Gen AI Hype (Spring 2024)Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
Deep Generative Learning for All - The Gen AI Hype (Spring 2024)
 
From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...From idea to production in a day – Leveraging Azure ML and Streamlit to build...
From idea to production in a day – Leveraging Azure ML and Streamlit to build...
 

Graphical Model Selection for Big Data

  • 1. aalto-logo-en-3 Graphical Model Selection for Big Data over Networks Alexander Jung Department of Computer Science Aalto University, School of Science 20.12.2016 1 / 41
  • 2. aalto-logo-en-3 Outline 1 Big Data over Networks 2 Probabilistic Graphical Models 3 GMS via Sparse Neighbourhood Regression IID Training Non-IID Training 4 Efficient GMS via Convex Optimization 2 / 41
  • 4. aalto-logo-en-3 Data, Data Everywhere Atacama Large Millimeter Array yields hundreds of TB (1012 bytes)/year CERN experiment generates data at rate of PB (1015 bytes)/second annual internet traffic crushed ZB (1021 bytes) borderline Big Data challenge: how to process data at internet scale? 4 / 41
  • 5. aalto-logo-en-3 The Three V’s of Big Data Volume: deal with Zettabyte (= 1021 bytes) scale data Velocity: CERN experiment generates Petabytes (= 1015 bytes) per second Variety: heterogeneous data (text, audio, video...) 5 / 41
  • 6. aalto-logo-en-3 Big Data over Networks often the datasets have an intrinsic network structure chip design internet bioinformatics social networks universe material science L. Lov´asz, “Large Networks and Graph Limits” 6 / 41
  • 7. aalto-logo-en-3 Big Data over Networks observe dataset D={z1, . . . , zN} with datapoints zi a particular zi might be audio, video or text data datapoints are structured by some notion of “similarity” e.g., zi , zj similar if from the same user account represent datapoint zi by node i ∈ V of graph G = (V, E) nodes i, j ∈ V connected by edge (i, j), if zi is similar to zj 7 / 41
  • 8. aalto-logo-en-3 Graph Signals for Supervised Learning consider supervised machine learning from dataset D each datapoint zi has a label r[i] (e.g., persons preference for buying certain product) entire labelling is a graph signal r[·] : V → R graph signal r[·] maps node i ∈ V to its label r[i] graph signal processing (GSP) provides efficient tools for handling large-scale graph signals 8 / 41
  • 9. aalto-logo-en-3 GSP generalizes DSP discrete-time signals are graph signals over chain graph · · ·· · · r[−1] r[0] r[1] label r[i] might correspond to presence of “clipping” at time i (greyscale) images are signals over grid graph · · · ··· r[i] label r[i] might be presence of certain object at location i GSP yields efficient learning by fast graph Fourier transforms 9 / 41
  • 10. aalto-logo-en-3 Fast Algorithms on Graphs GSP theory yields fast algorithms for large-scale graphs generalizes FFT from chain graph to general graphs parallelization/vectorization possible for product graph structure 6 graph product, denoted as G = G1 × G2, is = A1 ⊗ IN2 + IN1 ⊗A2. (25) g product, denoted as G = G1 G2, the ⊗ A2 + A1 ⊗ IN2 + IN1 ⊗A2. (26) can be seen as a combination of the Kro- n products. Since the products (23), (25), ve, Kronecker, Cartesian, and strong graph ned for an arbitrary number of graphs. rise in different applications, including processing [32], computational sciences ], and computational biology [34]. Their parts are used in network modeling and ], [37]. Multiple approaches have been composition and approximation of graphs [38], [30], [31], [39]. er a versatile graph model for the represen- atasets in multi-level and multi-parameter DSP, multi-dimensional signals, such as video, reside on rectangular lattices that ts of line graphs. Fig. 2(a) shows a two- formed by the Cartesian product of two ces. of graph signals residing on product graphs Digital image Row Column (a) Sensor network Time series Measurements atonetimestep Measurements of one sensor Sensor network measurements (b) Social network with communities Community structure Intercommunity communication structure (c) 10 / 41
  • 11. aalto-logo-en-3 Graph Models: Perfect Match for 3 Vs of Big Data graph based models lead to message passing algorithms message passing algorithms are perfectly scalable copes with volume (distributed computing) and velocity (parallel computing) of big data ship computation to data and not vice-versa! graph-based models allow to process heterogeneous data 11 / 41
  • 12. aalto-logo-en-3 Semi-Supervised Learning for Big Data over Networks consider graph signal x[i] representing big dataset D we only observe labels at sampling set S ⊆ V aquiring labels can be costly central hypothesis of supervised learning close-by data points in high-density regions have similar labels 12 / 41
  • 13. aalto-logo-en-3 Aquiring Labels (Sampling) in Marine Biology 13 / 41
  • 14. aalto-logo-en-3 Aquiring Labels (Sampling) in Particle Physics 14 / 41
  • 16. aalto-logo-en-3 Key Problems for Learning from Big Data over Networks given a graph signal representation of the learning problem: how many labels (samples) do we need? which nodes should we sample ? what are efficient learning algorithms ? all this presupposes that we know the graph structure! 16 / 41
  • 17. aalto-logo-en-3 Outline 1 Big Data over Networks 2 Probabilistic Graphical Models 3 GMS via Sparse Neighbourhood Regression IID Training Non-IID Training 4 Efficient GMS via Convex Optimization 17 / 41
  • 18. aalto-logo-en-3 Conditional Independence Graph interpret ith data point as realization of random variable zi associate zi with node i ∈ V of undirected graph G = (V, E) connect i, j ∈ V by undirected edge if zi and zj conditionally independent given {zl }l∈V{i,j} G is the conditional independence graph (CIG) of {zi }p i=1 18 / 41
  • 19. aalto-logo-en-3 Modelling Assumptions assume the zj to be jointly Gaussian N(0, C) with C 0 edges of CIG G characterized by non-zero entries of K = C−1: {i, j} ∈ E ⇐⇒ Ki,j = 0 define neighbourhood N(i) := {j ∈ V : {i, j} ∈ E} of node i we assume CIG to be sparse, i.e., |N(i)| ≤ smax for some fixed sparsity smax 19 / 41
  • 20. aalto-logo-en-3 The Global Markov Property CIG G allows to read off conditional independence relations A B C node set B separates A from C. global Markov property: {zi }i∈A conditionally independent of {Zi }i∈C given {Zi }i∈B Bayes optimal estimator for Zi depends only on {Zj }j∈N(i) for sparse graphs, i.e., |N(i)| p, Bayes optimal estimator can be implemented efficiently! 20 / 41
  • 21. aalto-logo-en-3 The Training Data for each datapoint Zi , observe training samples xi [1], . . . , xi [N] stack samples into vector x[t] = (x1[t], . . . , xp[t])T each vector sample x[t] multivariate normal ∼ N(0, C[t]) graphical model selection (GMS): learn G from x[t] x1[1], . . . , x1[N] x2[1], . . . , x2[N] x3[1], . . . , x3[N] 21 / 41
  • 22. aalto-logo-en-3 Outline 1 Big Data over Networks 2 Probabilistic Graphical Models 3 GMS via Sparse Neighbourhood Regression IID Training Non-IID Training 4 Efficient GMS via Convex Optimization 22 / 41
  • 23. aalto-logo-en-3 Outline 1 Big Data over Networks 2 Probabilistic Graphical Models 3 GMS via Sparse Neighbourhood Regression IID Training Non-IID Training 4 Efficient GMS via Convex Optimization 23 / 41
  • 24. aalto-logo-en-3 The Idea of Neighbourhood Regression consider the problem of finding the neighbourhood N(i) once we found all neighbourhoods, we have the CIG (trivial!) neighbourhood N(i) pops up in regression model for zi : zi = j∈N(i) aj zj + ei regression coeffs aj = −Ki,j /Ki,i error term ei uncorrelated with {zj }j∈V{i} neighbourhood N(i) solves sparse regression problem N(i) = arg min | supp(a)|≤smax E{(zi − j∈supp(a) aj zj )2 } 24 / 41
  • 25. aalto-logo-en-3 Learning based on Sparse Regression neighbourhood N(i) characterized by N(i) = arg min | supp(a)|≤smax E{(zi − j∈supp(a) aj zj )2 } not implementable since cannot evaluate expectation E consider i.i.d. training x[t] ∼ N(0, C) replace E with sample average N(i) ≈ arg min | supp(a)|≤smax (1/N) N t=1 (xi [t] − j∈supp(a) aj xj [t])2 (1) involves search over p smax subsets (more on this later!) 25 / 41
  • 26. aalto-logo-en-3 The Test Statistic of Sparse Regression in what follows assume |N(i)| = smax for index set I = {i1, . . . , ismax }, define statistic Z(I) := arg min | supp(a)|⊆I (1/N) N t=1 (xi [t] − j∈supp(a) aj xj [t])2 = (1/N) P⊥ I xi 2 2 with the vector xi = (xi [1], . . . , xi [N])T projection P⊥ I on orthogonal complement of span{xj }j∈I P⊥ I := I − XI XT I XI −1 XT I 26 / 41
  • 27. aalto-logo-en-3 Pairwise Error consider index set T with |T | ≤ smax and N(i) T = ∅ i ∈ V N(i) T with high probability Z(N(i)) < Z(T ) Z(I) prob(Z(I)) I =N (i) I =T 27 / 41
  • 28. aalto-logo-en-3 Pairwise Error assume samples x[t] i.i.d. with ∼ N(0, C) edges of CIG correspond to non-zero locations in K = C−1 define condition number κ := λmin(C)/λmax(C) strength of edge (i, j) quantified by partial correlation ρi,j := |Ki,j |/ Ki,i Kj,j minimum partial correlation ρmin ≤ ρi,j the probability of confusing T with N(i) is bounded as Prob{Z(N(i)) ≥ Z(T )} ≤ 4 exp −(N−smax) ρ2 minκ 64(ρ2 minκ + 8) 28 / 41
  • 29. aalto-logo-en-3 Sample Complexity of GMS by union bound over all p smax subsets T , we obtain Prob(N(i) = N(i))≤c1 exp smax log p−(N−smax) ρ2 minκ 64(ρ2 minκ+8) another union bound over all nodes i yields Prob(N(i) = N(i))≤c1 exp 2smax log p−(N−smax) ρ2 minκ 64(ρ2 minκ+8) accurate GMS for sample size N ≥ c1 max log p−smax smax , log(p−smax) κρ2 min allows for high-dimensional regime: large p, small N 29 / 41
  • 30. aalto-logo-en-3 How Low Can We Go? what is the fundamental lower bound on sample size N? it can be shown that for sample size N ≤ c2 log(p−smax) κρ2 min ANY GMS fails with non-negligible probability thus, sample complexity of GMS is N ∝ log(p−smax) κρ2 min sample complexity driven by minimum partial correlation ρmin 30 / 41
  • 31. aalto-logo-en-3 Outline 1 Big Data over Networks 2 Probabilistic Graphical Models 3 GMS via Sparse Neighbourhood Regression IID Training Non-IID Training 4 Efficient GMS via Convex Optimization 31 / 41
  • 32. aalto-logo-en-3 Non-IID Samples samples x[f ]∼N(0, C[f ]) still independent but varying C[f ] model useful for two particular process classes: x[f ] are Fourier coeffs of stationary process y[t]: x[f ] = N t=1 y[t] exp(−(2π/N)(t−1)(f −1)) x[f ] are LCB coeffs of locally stationary process y[t]: x[f ] = N t=1 y[t]g(f ) [t] 32 / 41
  • 33. aalto-logo-en-3 Sample Complexity for Non-IID Samples RECALL sample complexity for iid case: N ∝ log(p−smax) κρ2 min replace ρmin with minimum average partial correlation ¯ρmin := min (i,j)∈E (1/N) N f =1 K2 i,j [f ]/(Ki,i [f ]Kj,j [f ]) with K[f ] := C−1[f ] sample complex for non-iid case: N ∝ log(p−smax) κ¯ρ2 min 33 / 41
  • 35. aalto-logo-en-3 Outline 1 Big Data over Networks 2 Probabilistic Graphical Models 3 GMS via Sparse Neighbourhood Regression IID Training Non-IID Training 4 Efficient GMS via Convex Optimization 35 / 41
  • 36. aalto-logo-en-3 Just Relax consider (vectorized) sparse neighbourhood regression N(i) ≈ arg min | supp(a)|≤smax xi − j∈supp(a) aj xj 2 2 = arg min a xi − Xa 2 2 + λ| supp(a)| with X := {xj }j=i and some multiplier λ involves intractable search over p smax supports supp(a) RELAX penalty λ| supp(a)| with convex function λ a 1 36 / 41
  • 37. aalto-logo-en-3 The LASSO convex relaxation of sparse neighbourhood regression aLasso ∈ arg min a xi −Xaj 2 2 + λ a 1 for some suitable multiplier λ convex optimization problem (Lagrangian form) least absolute shrinkage and selection operator (LASSO) LASSO found widespread use in statistics/signal processing 37 / 41
  • 38. aalto-logo-en-3 Sample Complexity of LASSO consider LASSO estimator aLasso ∈ arg min a xi −Xaj 2 2 + λ a 1 assume that zi are only weakly correlated |E{zi zj }|/ E{z2 i }E{z2 j } ≤ /(2smax) we have supp(aLasso) = N(i) with high prob. if N ≥ (c1smax + c2/(κρ2 min)) log(p−smax) 38 / 41
  • 39. aalto-logo-en-3 Structure of LASSO consider LASSO arg min a xi − Xa 2 2 + λ a 1 sum of smooth term xi − Xa 2 2 and non-smooth a 1 highly developed methods for such optimization problems e.g., proximal gradient method, alternating direction method of multipliers (ADMM), Pock-Chambolle primal-dual, etc.... basic idea: splitting of minimization of the two terms 39 / 41
  • 40. aalto-logo-en-3 ADMM for LASSO LASSO: aLasso ∈ arg mina=b xi − Xa 2 2 + λ b 1 ADMM amounts to iterating a(k+1) = XT X + ρI −1 XT xi + ρ(b(k) − c(k) ) b(k+1) = Sλ/ρ(a(k+1) + c(k) ) c(k+1) = c(k) + a(k+1) − b(k+1) a Sκ(a) κ κ 2κ convergence under very general conditions 40 / 41
  • 41. aalto-logo-en-3 Future Directions consider directed graphical models (Bayesian networks) analyze intrinsic computational complexity of GMS (is group LASSO computationally efficient) extend approach to vector-valued data points Zi ∈ Rp incorporate additional structure on top of sparsity, e.g., product graphs 41 / 41