SlideShare a Scribd company logo
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Spatially Informed Variable Selection Priors and
Applications to Large-Scale Data
Marina Vannucci
Department of Statistics
Rice University
Houston, TX
USA
April 2019 - SAMSI
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Outline of the Talk
Variable selection via spike-and-slab priors
Accounting for spatial information
Nonparametric constructions
Edge selection in graphical models
Single graphs
Multiple graphs
Beyond Gaussian data - count data
Computational issues - Variational Inference
Motivating applications: High-throughput genomics;
neuroimaging.
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Spike-and-slab Variable Selection Prior for Regression Models
Yn×1 = Xn×pβp×1 + ε, ε ∼ N(0, σ2
I)
Introduce latent variable γ = (γ1, . . . , γp) to select variables
γj = 1 if variable j included in model
γj = 0 otherwise
Specify priors for model parameters:
βj ∼ (1 − γj)δ0 + γjN(0, σ2
τ2
j ), σ2
∼ IG(ν/2, λ/2)
p(γ) =
p
j=1
ωγj (1 − ω)1−γj
Combine data and prior information into a posterior distribution
⇒ interest in posterior distribution p(γ|X, Y).
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Posterior Inference via MCMC
Use Metropolis as stochastic search. At each iteration
(i) Add or Delete one variable in γold
(ii) Swap a 0 and a 1 in γold
The proposed γnew is accepted with probability
min
p(γnew |X, Y)
p(γold |X, Y)
, 1 .
George & McCulloch (1993,1997); Brown et al. (1998 & 2002)
Sample (β, γ) jointly if marginalization not possible
Gottardo & Raftery (2008 JCGS); Savitsky et al. (2011 Stat Science)
Select variables in the “best” models or from largest
marginal posterior probabilities of inclusion (PPIs) via
median model or Bayesian FDR thresholds.
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Functional Magnetic Resonance Imaging (fMRI)
Source:(www.anc.ed.ac.uk((
Source:(photos.uc.wisc.edu((
Indirect measure of brain activity as changes in blood flow.
Collected at rest or during a sensorimotor task.
Observed data - time series of the blood oxygenation level
dependent (BOLD) response, at each voxel in the brain.
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Model for Single Subject
Yν = Xνβν + εν, ν = 1, . . . , V
fMRI = BOLD Signal + Noise
Yν = (Yν1, . . . , YνT )T time-series data for voxel ν.
Xν the convolved design matrix T × p (p stimuli).
Convolution of stimulus function with HRF (e.g., Poisson
with parameter λν) to account for time lapse between
stimulus and vascular response.
βν = (βν1, . . . , βνp)T regression coefficients vector
εν ∼ NT (0, Σν) caused by hardware and subjects
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Spatial Modeling
Spike-and-slab prior on to select activated voxels (p = 1)
βν ∼ γνN(0, τ) + (1 − γν)δ0, ν = 1, . . . , V,
MRF prior on voxel-dependent parameter γν
(Zhang et al., 2014 NeuroImage)
P(γν|d, e, γk , k ∈ Nν) ∝ exp(γν(d + e
k∈Nν
γν))
with Nν the set of neighbors of voxel ν, d ∈ R, e > 0.
Probit priors
(Chiang et al., 2017 Human Brain Mapping)
p(γν = 1) = Φ (α0 + α1Nν)
with Nν a set of covariates (e.g., multi-modal image data).
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Spatial Modeling - Nonparametric Prior
Objective: capture correlation among time series data of
possibly distant voxels
Proposed prior: spike-and-slab nonparametric prior,
βν|γν, G0 ∼ γνG0 + (1 − γν)δ0
with G0 a Dirichlet process prior with N(0, τ) as centering
distribution.
Achieves clustering of voxels within subject according to
similar activation strengths.
Particularly useful for multi-subject modeling
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Regression Model for Multiple Subjects
Extend the single-subject regression model to multiple subjects
Yiν = Xiνβiν + εiν, εiν ∼ NT (0, Σiν)
Yiν = (Yiν1, . . . , YiνT )T , T × 1 BOLD response data for the
νth voxel in the ith subject
Xiν, T × p design matrix, with Poisson HRF
βiν, p × 1 vector of regression coefficients (spatial prior)
Hierarchical Dirichlet process as slab distribution on βiν
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Spatial Nonparametric Prior
Objective: capture correlation within and across subjects.
Proposed prior: spike-and-slab nonparametric prior,
βiν|γiν, Gi ∼ γiνGi + (1 − γiν)δ0
Clustering of subjects based on similar activation patterns;
clustering of voxels within each subject according to similar
activation strengths.
Subject(1(
Subject(3(
Subject(4( Subject(2(
Subject(5(
(Zhang et al., 2016 Annals of Applied Statistics)
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Posterior Inference
Standard approach: MCMC
MH updates to jointly update (β, γ) with a combination of
add-delete-swap steps and sampling-based methods for
DP models (based on the Chinese restaurant franchise)
Update λ with Metropolis-Hastings (MH) schemes
Update other model parameters
Faster strategy: Variational Inference (Matlab GUI)
Combine VI schemes with importance sampling steps on
other model parameters (Carbonetto and Stephens, 2012)
Posterior maps on γiν, βiν, λiν’s
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Software: NPBayes-fMRI GUI (https://github.com/marinavannucci)
500
520
540
560
580
600
620
640
660
680
700
Cluster 1 (n = 12) Cluster 2 (n = 16)
RTtoCorrectTrials(ms)
Nonwords Pseudohomophones
*
**
* p < .05, ** p < .001 -30
-20
-10
0
10
20
30
40
50
60
PseudohomophoneEffect(ms)
Cluster 1 Cluster 2
Fischer-Baum et al. (2018). Individual differences in the neural and cognitive
mechanisms of single word reading. Frontiers in Human Neuroscience, 12:271.
(Kook	et	al.,	2019	Statistics	in	Biosciences)
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Edge Selection in Graphical Models
Undirected Graphical Models
Use a graph structure to
summarize conditional
independence between variables
Each node represents a variable
No edge exists between two
nodes if and only if variables are
independent given all others
Fundamental assumption
True conditional dependence structure is reasonably
sparse
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Gaussian Graphical Model (GGM)
Data follow a joint multivariate normal distribution
xl ∼ N(µ, Ω−1
) l = 1 . . . n
ωi,j = 0 if no edge exists
between nodes i and j in
graph G
For simplicity, we
column-center the data to
assume µ = 0
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Choices for Edge Selection Priors
G-Wishart priors enforce exact zeros but are limited in
scalability (Roverato 2002, Dobra et al. 2011)
Continuous shrinkage prior of Wang (BA 2015)
p(Ωk |G, v0, v1, λ) ∝
i<j
N(ωi,j|0, v2
gij
)
i
Exp(ωi,i|
λ
2
)
defined by introducing binary latent variables of
edge-inclusion G ≡ (gi,j)i<j
∈ G ≡ {0, 1}p(p−1)/2
Avoid use of priors over the space of positive definite
matrices with fixed zeroes
Improve computational efficiency
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Bayesian Framework for Multiple Graphs Estimation
xk,l ∼ N(µk , Ω−1
k ) l = 1, . . . , nk k = 1, . . . , K
Graphical model for each group over common set of
variables
xk,l, l = 1, . . . , nk data for subject l in group k
Ωk = Σ−1
k = (ωi,j,k ) is the precision matrix for group k
May share common features, or could differ substantially
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Linking Networks with a MRF Prior
Encourages selection of common edges for similar groups
p(gij|νij, Θ) =
exp(νij 1T gij+gij
T Θgij)
C(νij ,Θ)
with gij = {g1ij, . . . , gkij}
Spike-and-slab prior on
off-diagonal elements of
K × K matrix Θ
Nonzero entries indicate
network relatedness
Magnitude measures
pairwise graph similarity
Flexibility of model
Shares information when appropriate
Does not enforce similarity unless supported by data
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Selection Prior for Network Similarity
Spike and slab prior on off-diagonal entries of Θ
p(θkm|γkm) = (1 − γkm)δ0 + γkm
βα
Γ(α)
θα−1
km exp(−βθkm)
Spike-and-slab prior on parameters measuring network
relatedness, to allow shared structure to be favored only when
supported by the data
Fixed hyperparameters α and β
Latent indicator variable γkm of graph relatedness
Independent Bernoulli priors defined on latent indicators
with probability w ∈ [0, 1]
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Completing the Model
Edge specific prior for edge inclusion probability
P(νij) =
1
β(a, b)
exp(aνij)
(1 + exp(νij))a+b
Can encourage sparsity of the graphs G1, . . . , GK
Can be used to incorporate prior knowledge of connections
between nodes
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Diagram of the Joint Graph Model
Peterson et al. (2015, JASA); Shaddox et al. (2019, Biostatistics)
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Posterior Inference
MCMC Sampling Scheme
At each iteration t of our algorithm, sample from posterior full
conditionals
(i) Update graph G
(t)
k and precision matrix Ω
(t)
k for
k = 1, . . . , K
Block Gibbs sampler
p-coupled stochastic search variable selection algorithm
(ii) Update network relatedness parameters θkm and γkm for
1 ≤ k < m ≤ K using Metropolis-Hastings steps
Incorporates both between-model and within-model moves
(iii) Update νij for 1 ≤ i < j ≤ p using a standard
Metropolis-Hastings step
Model Selection
Posterior marginal probabilities of edge inclusion (PPIs)
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
[Control]
1
2
3
4
5
6
7
8
9
10
1112
13
14
15
16
17
18
19
20
21
22
23 24
25
[Mild]
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23 24
25
[Moderate]
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
[Severe]
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16 17
18
19
20
21
22
23 24
25
p = 100 TPR(SE) FPR (SE) MCC(SE) AUC(SE)
Fused GLasso 0.8658 (.0479) 0.2620 (.2632) 0.2775(.0940) 0.9688(.0016)
Group GLasso 0.8496 (.0095) 0.1739 (.0028) 0.3089 (.0042) 0.8982(.0048)
Our Method 0.7135 (.2929) 0.0053 (.0037) 0.7453 (.2334) 0.9578 (.0434)
ΘM =


· .95 .97
· .95
·


Pathway Platform Total Pairs 100 110 011 001 Total Disrupted
FcγR Metabolites 73 17 5 3 18 43
(Shaddox et al., 2019 Biostatistics)
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Beyond Gaussian Data
Linear models
Earlier work in GLM/GLMM models (Raftery 1996; Chen &
Dunson, 2003; Kinney & Dunson, 2007)
Data augmentation techniques for logit and log links
(Polson et al. (2013) and Zhu et al. (2012))
Recent interest in (zero-inflated) negative binomial and
Dirichlet-multinomial regression models for large-scale
genomic data (e.g. microbiome)
(Wadsworth et al. 2017, Koslovsky and Vannucci 2019)
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Dirichlet-Multinomial Regression
Multivariate counts modeled as Multinomial(y+, φ), with
y+ = q
j=1 yj, e.g. q microbial taxa, and φ ∼ Dirichlet(γ)
Integrate p covariates as log(γj(xi)) = xiθj
Obtain hierarchical Dirichlet-multinomial log-linear
regression
Yi | φi
ind.
∼ Multinomial(y+, φi)
φi | α, β1, . . . , βp
ind.
∼ Dirichlet

 exp(αj +
p
k=1
βkjxik ))
j=1,...,q


with φi taxa-specific proportions.
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Variable Selection Prior
Spike-and-slab selects relevant taxa/predictor associations
βkj | ξkj ∼ ξkj Normal(0 | t2
kj) + (1 − ξkj)δ0(βkj)
Stochastic search MCMC for posterior inference
1 Draw auxiliary variables for DM(γj (x))
2 Update α using Metropolis–Hastings
3 Update (ξ, β) jointly via SSVB MCMC (Savitsky et al., 2011)
Positive Associations (MPPI > 0.981)
ko00920
ko00513
ko00260
ko04660
ko00626
ko00620
ko01057
ko04512
ko04340
ko00330
f.Prevotellaceae
g.Parabacteroides
g.Prevotella
o.Bacteroidales
g.Roseburia
g.Subdoligranulum
g.Bacteroides
g.Ruminococcus
Negative Associations (MPPI > 0.981)
ko00680
ko00512
ko00904
ko00500
ko00620
ko00010
ko00534
ko05200
ko00053
ko04540
ko00650
ko00983
g.Sutterella
g.Prevotella
g.Dialister
g.Parabacteroides
g.Bacteroides
g.Ruminococcus
o.Bacteroidales
Wadsworth et al. (2017, BMC Bioinformatics); Koslovsky and Vannucci
(2019, submitted) for Dirichlet-tree multinomial distributions
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Variational Inference
Variational inference turns inference into an optimization
problem. Faster and more scalable than MCMC.
Underlying idea: pick family of distributions q (β, γ) ∈ Q,
with free variational parameters; use gradient descent to
minimize KL divergence between q and posterior f (β, γ)
q∗
= arg min
q∈Q
KL (q f) = q (β, γ) log
q (β, γ)
f (β, γ)
dβdγ
= log p (y | X) − ELBO
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Mean Field VI
Fully factorized approximation to reduce complexity
q (β, γ | θ) =
p
k=1
q (βk , γk ; θk )
Choose approximating distributions from same family as
prior distributions, to exploit conjugacy.
Combine with importance sampling schemes on other
model parameters (Zhang et al., 2016 AOAS)
Extend to GLM models with logit and log links by
incorporating data augmentation schemes of Polson et al.
(2013) and Zhu et al. (2012) for logistic and Negative
Binomial regression models.
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
Summary and Conclusions
Spike-and-slab priors for variable selection are well suited
for applications.
Flexible structure for the incorporation of external
information via MRF or probit priors or spiked DP
constructions.
Priors for edge selection in Gaussian graphical models.
Methodologies can be extended beyond Gaussian data.
Variational inference approximations for scalability.
Neuroimaging and genomics as emerging areas for
applications.
Improved performance over competitive approaches.
Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions
THANKS!
Eric Kook, PhD student, Rice University.
Yinsen Miao, PhD student, Rice University.
Nathan Osborne, PhD student, Rice University.
Matthew Koslovsky, postdoc, Rice University.
Elin Shaddox, PhD 2019, postdoc at UC Denver
Qiwei Li, Assistant Professor, UT Dallas.
Duncan Wadsworth, Data Scientist, Microsoft, Seattle.
Terrance Savitsky, Bureau of Labor Statistics, D.C.
Alberto Cassese, Assistant Professor, Maastricht, The Netherland.
Christine Peterson, Assistant Professor, MD Anderson.
Francesco Stingo, Assistant professor, U. Florence, Italy.
Michele Guindani, Professor, UC Irvine.
HAMILL INNOVATION AWARD

More Related Content

What's hot

MNAR
MNARMNAR
On the vexing dilemma of hypothesis testing and the predicted demise of the B...
On the vexing dilemma of hypothesis testing and the predicted demise of the B...On the vexing dilemma of hypothesis testing and the predicted demise of the B...
On the vexing dilemma of hypothesis testing and the predicted demise of the B...
Christian Robert
 
Statistics symposium talk, Harvard University
Statistics symposium talk, Harvard UniversityStatistics symposium talk, Harvard University
Statistics symposium talk, Harvard University
Christian Robert
 
WSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in StatisticsWSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in Statistics
Christian Robert
 
Graph Neural Network for Phenotype Prediction
Graph Neural Network for Phenotype PredictionGraph Neural Network for Phenotype Prediction
Graph Neural Network for Phenotype Prediction
tuxette
 
TS-IASSL2014
TS-IASSL2014TS-IASSL2014
TS-IASSL2014
ychaubey
 
from model uncertainty to ABC
from model uncertainty to ABCfrom model uncertainty to ABC
from model uncertainty to ABC
Christian Robert
 
A quantum-inspired optimization heuristic for the multiple sequence alignment...
A quantum-inspired optimization heuristic for the multiple sequence alignment...A quantum-inspired optimization heuristic for the multiple sequence alignment...
A quantum-inspired optimization heuristic for the multiple sequence alignment...
Konstantinos Giannakis
 
NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015
Christian Robert
 
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
The Statistical and Applied Mathematical Sciences Institute
 
Explanable models for time series with random forest
Explanable models for time series with random forestExplanable models for time series with random forest
Explanable models for time series with random forest
tuxette
 
La statistique et le machine learning pour l'intégration de données de la bio...
La statistique et le machine learning pour l'intégration de données de la bio...La statistique et le machine learning pour l'intégration de données de la bio...
La statistique et le machine learning pour l'intégration de données de la bio...
tuxette
 
Boston talk
Boston talkBoston talk
Boston talk
Christian Robert
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
Christian Robert
 
Reliable ABC model choice via random forests
Reliable ABC model choice via random forestsReliable ABC model choice via random forests
Reliable ABC model choice via random forests
Christian Robert
 
Slides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processingSlides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processing
Frank Nielsen
 
Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach
Jae-kwang Kim
 
A Note on Confidence Bands for Linear Regression Means-07-24-2015
A Note on Confidence Bands for Linear Regression Means-07-24-2015A Note on Confidence Bands for Linear Regression Means-07-24-2015
A Note on Confidence Bands for Linear Regression Means-07-24-2015
Junfeng Liu
 
Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Laplace's Demon: seminar #1
Laplace's Demon: seminar #1
Christian Robert
 
Introduction to Supervised ML Concepts and Algorithms
Introduction to Supervised ML Concepts and AlgorithmsIntroduction to Supervised ML Concepts and Algorithms
Introduction to Supervised ML Concepts and Algorithms
NBER
 

What's hot (20)

MNAR
MNARMNAR
MNAR
 
On the vexing dilemma of hypothesis testing and the predicted demise of the B...
On the vexing dilemma of hypothesis testing and the predicted demise of the B...On the vexing dilemma of hypothesis testing and the predicted demise of the B...
On the vexing dilemma of hypothesis testing and the predicted demise of the B...
 
Statistics symposium talk, Harvard University
Statistics symposium talk, Harvard UniversityStatistics symposium talk, Harvard University
Statistics symposium talk, Harvard University
 
WSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in StatisticsWSC 2011, advanced tutorial on simulation in Statistics
WSC 2011, advanced tutorial on simulation in Statistics
 
Graph Neural Network for Phenotype Prediction
Graph Neural Network for Phenotype PredictionGraph Neural Network for Phenotype Prediction
Graph Neural Network for Phenotype Prediction
 
TS-IASSL2014
TS-IASSL2014TS-IASSL2014
TS-IASSL2014
 
from model uncertainty to ABC
from model uncertainty to ABCfrom model uncertainty to ABC
from model uncertainty to ABC
 
A quantum-inspired optimization heuristic for the multiple sequence alignment...
A quantum-inspired optimization heuristic for the multiple sequence alignment...A quantum-inspired optimization heuristic for the multiple sequence alignment...
A quantum-inspired optimization heuristic for the multiple sequence alignment...
 
NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015NBBC15, Reyjavik, June 08, 2015
NBBC15, Reyjavik, June 08, 2015
 
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
MUMS Opening Workshop - Quantifying Nonparametric Modeling Uncertainty with B...
 
Explanable models for time series with random forest
Explanable models for time series with random forestExplanable models for time series with random forest
Explanable models for time series with random forest
 
La statistique et le machine learning pour l'intégration de données de la bio...
La statistique et le machine learning pour l'intégration de données de la bio...La statistique et le machine learning pour l'intégration de données de la bio...
La statistique et le machine learning pour l'intégration de données de la bio...
 
Boston talk
Boston talkBoston talk
Boston talk
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
Reliable ABC model choice via random forests
Reliable ABC model choice via random forestsReliable ABC model choice via random forests
Reliable ABC model choice via random forests
 
Slides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processingSlides: A glance at information-geometric signal processing
Slides: A glance at information-geometric signal processing
 
Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach
 
A Note on Confidence Bands for Linear Regression Means-07-24-2015
A Note on Confidence Bands for Linear Regression Means-07-24-2015A Note on Confidence Bands for Linear Regression Means-07-24-2015
A Note on Confidence Bands for Linear Regression Means-07-24-2015
 
Laplace's Demon: seminar #1
Laplace's Demon: seminar #1Laplace's Demon: seminar #1
Laplace's Demon: seminar #1
 
Introduction to Supervised ML Concepts and Algorithms
Introduction to Supervised ML Concepts and AlgorithmsIntroduction to Supervised ML Concepts and Algorithms
Introduction to Supervised ML Concepts and Algorithms
 

Similar to MUMS: Bayesian, Fiducial, and Frequentist Conference - Spatially Informed Variable Selection Priors and Applications to Large-scale Data (revised), Marina Vannucci, April 30, 2019

Quantitative Propagation of Chaos for SGD in Wide Neural Networks
Quantitative Propagation of Chaos for SGD in Wide Neural NetworksQuantitative Propagation of Chaos for SGD in Wide Neural Networks
Quantitative Propagation of Chaos for SGD in Wide Neural Networks
Valentin De Bortoli
 
Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...
Frank Nielsen
 
the ABC of ABC
the ABC of ABCthe ABC of ABC
the ABC of ABC
Christian Robert
 
Workshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinWorkshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael Martin
Christian Robert
 
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
The Statistical and Applied Mathematical Sciences Institute
 
Cluster Analysis: Measuring Similarity & Dissimilarity
Cluster Analysis: Measuring Similarity & DissimilarityCluster Analysis: Measuring Similarity & Dissimilarity
Cluster Analysis: Measuring Similarity & Dissimilarity
ShivarkarSandip
 
Combining Data in Species Distribution Models
Combining Data in Species Distribution ModelsCombining Data in Species Distribution Models
Combining Data in Species Distribution Models
Bob O'Hara
 
Jörg Stelzer
Jörg StelzerJörg Stelzer
Jörg Stelzer
butest
 
Methods for High Dimensional Interactions
Methods for High Dimensional InteractionsMethods for High Dimensional Interactions
Methods for High Dimensional Interactions
sahirbhatnagar
 
asymptotics of ABC
asymptotics of ABCasymptotics of ABC
asymptotics of ABC
Christian Robert
 
Thesis seminar
Thesis seminarThesis seminar
Thesis seminar
gvesom
 
Estimating Space-Time Covariance from Finite Sample Sets
Estimating Space-Time Covariance from Finite Sample SetsEstimating Space-Time Covariance from Finite Sample Sets
Estimating Space-Time Covariance from Finite Sample Sets
Förderverein Technische Fakultät
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
Valentin De Bortoli
 
Tree net and_randomforests_2009
Tree net and_randomforests_2009Tree net and_randomforests_2009
Tree net and_randomforests_2009
Matthew Magistrado
 
Refining Bayesian Data Analysis Methods for Use with Longer Waveforms
Refining Bayesian Data Analysis Methods for Use with Longer WaveformsRefining Bayesian Data Analysis Methods for Use with Longer Waveforms
Refining Bayesian Data Analysis Methods for Use with Longer Waveforms
James Bell
 
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
The Statistical and Applied Mathematical Sciences Institute
 
Probability distributions for ml
Probability distributions for mlProbability distributions for ml
Probability distributions for ml
Sung Yub Kim
 
KAUST_talk_short.pdf
KAUST_talk_short.pdfKAUST_talk_short.pdf
KAUST_talk_short.pdf
Chiheb Ben Hammouda
 
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
The Statistical and Applied Mathematical Sciences Institute
 
Bayesian hybrid variable selection under generalized linear models
Bayesian hybrid variable selection under generalized linear modelsBayesian hybrid variable selection under generalized linear models
Bayesian hybrid variable selection under generalized linear models
Caleb (Shiqiang) Jin
 

Similar to MUMS: Bayesian, Fiducial, and Frequentist Conference - Spatially Informed Variable Selection Priors and Applications to Large-scale Data (revised), Marina Vannucci, April 30, 2019 (20)

Quantitative Propagation of Chaos for SGD in Wide Neural Networks
Quantitative Propagation of Chaos for SGD in Wide Neural NetworksQuantitative Propagation of Chaos for SGD in Wide Neural Networks
Quantitative Propagation of Chaos for SGD in Wide Neural Networks
 
Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...
 
the ABC of ABC
the ABC of ABCthe ABC of ABC
the ABC of ABC
 
Workshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael MartinWorkshop in honour of Don Poskitt and Gael Martin
Workshop in honour of Don Poskitt and Gael Martin
 
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
Program on Mathematical and Statistical Methods for Climate and the Earth Sys...
 
Cluster Analysis: Measuring Similarity & Dissimilarity
Cluster Analysis: Measuring Similarity & DissimilarityCluster Analysis: Measuring Similarity & Dissimilarity
Cluster Analysis: Measuring Similarity & Dissimilarity
 
Combining Data in Species Distribution Models
Combining Data in Species Distribution ModelsCombining Data in Species Distribution Models
Combining Data in Species Distribution Models
 
Jörg Stelzer
Jörg StelzerJörg Stelzer
Jörg Stelzer
 
Methods for High Dimensional Interactions
Methods for High Dimensional InteractionsMethods for High Dimensional Interactions
Methods for High Dimensional Interactions
 
asymptotics of ABC
asymptotics of ABCasymptotics of ABC
asymptotics of ABC
 
Thesis seminar
Thesis seminarThesis seminar
Thesis seminar
 
Estimating Space-Time Covariance from Finite Sample Sets
Estimating Space-Time Covariance from Finite Sample SetsEstimating Space-Time Covariance from Finite Sample Sets
Estimating Space-Time Covariance from Finite Sample Sets
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
 
Tree net and_randomforests_2009
Tree net and_randomforests_2009Tree net and_randomforests_2009
Tree net and_randomforests_2009
 
Refining Bayesian Data Analysis Methods for Use with Longer Waveforms
Refining Bayesian Data Analysis Methods for Use with Longer WaveformsRefining Bayesian Data Analysis Methods for Use with Longer Waveforms
Refining Bayesian Data Analysis Methods for Use with Longer Waveforms
 
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
 
Probability distributions for ml
Probability distributions for mlProbability distributions for ml
Probability distributions for ml
 
KAUST_talk_short.pdf
KAUST_talk_short.pdfKAUST_talk_short.pdf
KAUST_talk_short.pdf
 
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
MUMS Opening Workshop - An Overview of Reduced-Order Models and Emulators (ED...
 
Bayesian hybrid variable selection under generalized linear models
Bayesian hybrid variable selection under generalized linear modelsBayesian hybrid variable selection under generalized linear models
Bayesian hybrid variable selection under generalized linear models
 

More from The Statistical and Applied Mathematical Sciences Institute

Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
The Statistical and Applied Mathematical Sciences Institute
 
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
The Statistical and Applied Mathematical Sciences Institute
 
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
The Statistical and Applied Mathematical Sciences Institute
 
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
The Statistical and Applied Mathematical Sciences Institute
 
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
The Statistical and Applied Mathematical Sciences Institute
 
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
The Statistical and Applied Mathematical Sciences Institute
 

More from The Statistical and Applied Mathematical Sciences Institute (20)

Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
 
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
 
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
 
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
 
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
 
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
 
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
 
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
 
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
 
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
 
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
 
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
 
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
 
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
 
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
 
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
 
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
 
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
 
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
 
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
 

Recently uploaded

Liberal Approach to the Study of Indian Politics.pdf
Liberal Approach to the Study of Indian Politics.pdfLiberal Approach to the Study of Indian Politics.pdf
Liberal Approach to the Study of Indian Politics.pdf
WaniBasim
 
MARY JANE WILSON, A “BOA MÃE” .
MARY JANE WILSON, A “BOA MÃE”           .MARY JANE WILSON, A “BOA MÃE”           .
MARY JANE WILSON, A “BOA MÃE” .
Colégio Santa Teresinha
 
Film vocab for eal 3 students: Australia the movie
Film vocab for eal 3 students: Australia the movieFilm vocab for eal 3 students: Australia the movie
Film vocab for eal 3 students: Australia the movie
Nicholas Montgomery
 
writing about opinions about Australia the movie
writing about opinions about Australia the moviewriting about opinions about Australia the movie
writing about opinions about Australia the movie
Nicholas Montgomery
 
Smart-Money for SMC traders good time and ICT
Smart-Money for SMC traders good time and ICTSmart-Money for SMC traders good time and ICT
Smart-Money for SMC traders good time and ICT
simonomuemu
 
Types of Herbal Cosmetics its standardization.
Types of Herbal Cosmetics its standardization.Types of Herbal Cosmetics its standardization.
Types of Herbal Cosmetics its standardization.
Ashokrao Mane college of Pharmacy Peth-Vadgaon
 
Life upper-Intermediate B2 Workbook for student
Life upper-Intermediate B2 Workbook for studentLife upper-Intermediate B2 Workbook for student
Life upper-Intermediate B2 Workbook for student
NgcHiNguyn25
 
Digital Artefact 1 - Tiny Home Environmental Design
Digital Artefact 1 - Tiny Home Environmental DesignDigital Artefact 1 - Tiny Home Environmental Design
Digital Artefact 1 - Tiny Home Environmental Design
amberjdewit93
 
How to Manage Your Lost Opportunities in Odoo 17 CRM
How to Manage Your Lost Opportunities in Odoo 17 CRMHow to Manage Your Lost Opportunities in Odoo 17 CRM
How to Manage Your Lost Opportunities in Odoo 17 CRM
Celine George
 
S1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptxS1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptx
tarandeep35
 
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
IreneSebastianRueco1
 
C1 Rubenstein AP HuG xxxxxxxxxxxxxx.pptx
C1 Rubenstein AP HuG xxxxxxxxxxxxxx.pptxC1 Rubenstein AP HuG xxxxxxxxxxxxxx.pptx
C1 Rubenstein AP HuG xxxxxxxxxxxxxx.pptx
mulvey2
 
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...
Dr. Vinod Kumar Kanvaria
 
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdfবাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
eBook.com.bd (প্রয়োজনীয় বাংলা বই)
 
PIMS Job Advertisement 2024.pdf Islamabad
PIMS Job Advertisement 2024.pdf IslamabadPIMS Job Advertisement 2024.pdf Islamabad
PIMS Job Advertisement 2024.pdf Islamabad
AyyanKhan40
 
Top five deadliest dog breeds in America
Top five deadliest dog breeds in AmericaTop five deadliest dog breeds in America
Top five deadliest dog breeds in America
Bisnar Chase Personal Injury Attorneys
 
Pollock and Snow "DEIA in the Scholarly Landscape, Session One: Setting Expec...
Pollock and Snow "DEIA in the Scholarly Landscape, Session One: Setting Expec...Pollock and Snow "DEIA in the Scholarly Landscape, Session One: Setting Expec...
Pollock and Snow "DEIA in the Scholarly Landscape, Session One: Setting Expec...
National Information Standards Organization (NISO)
 
Chapter 4 - Islamic Financial Institutions in Malaysia.pptx
Chapter 4 - Islamic Financial Institutions in Malaysia.pptxChapter 4 - Islamic Financial Institutions in Malaysia.pptx
Chapter 4 - Islamic Financial Institutions in Malaysia.pptx
Mohd Adib Abd Muin, Senior Lecturer at Universiti Utara Malaysia
 
Pride Month Slides 2024 David Douglas School District
Pride Month Slides 2024 David Douglas School DistrictPride Month Slides 2024 David Douglas School District
Pride Month Slides 2024 David Douglas School District
David Douglas School District
 
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdfANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
Priyankaranawat4
 

Recently uploaded (20)

Liberal Approach to the Study of Indian Politics.pdf
Liberal Approach to the Study of Indian Politics.pdfLiberal Approach to the Study of Indian Politics.pdf
Liberal Approach to the Study of Indian Politics.pdf
 
MARY JANE WILSON, A “BOA MÃE” .
MARY JANE WILSON, A “BOA MÃE”           .MARY JANE WILSON, A “BOA MÃE”           .
MARY JANE WILSON, A “BOA MÃE” .
 
Film vocab for eal 3 students: Australia the movie
Film vocab for eal 3 students: Australia the movieFilm vocab for eal 3 students: Australia the movie
Film vocab for eal 3 students: Australia the movie
 
writing about opinions about Australia the movie
writing about opinions about Australia the moviewriting about opinions about Australia the movie
writing about opinions about Australia the movie
 
Smart-Money for SMC traders good time and ICT
Smart-Money for SMC traders good time and ICTSmart-Money for SMC traders good time and ICT
Smart-Money for SMC traders good time and ICT
 
Types of Herbal Cosmetics its standardization.
Types of Herbal Cosmetics its standardization.Types of Herbal Cosmetics its standardization.
Types of Herbal Cosmetics its standardization.
 
Life upper-Intermediate B2 Workbook for student
Life upper-Intermediate B2 Workbook for studentLife upper-Intermediate B2 Workbook for student
Life upper-Intermediate B2 Workbook for student
 
Digital Artefact 1 - Tiny Home Environmental Design
Digital Artefact 1 - Tiny Home Environmental DesignDigital Artefact 1 - Tiny Home Environmental Design
Digital Artefact 1 - Tiny Home Environmental Design
 
How to Manage Your Lost Opportunities in Odoo 17 CRM
How to Manage Your Lost Opportunities in Odoo 17 CRMHow to Manage Your Lost Opportunities in Odoo 17 CRM
How to Manage Your Lost Opportunities in Odoo 17 CRM
 
S1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptxS1-Introduction-Biopesticides in ICM.pptx
S1-Introduction-Biopesticides in ICM.pptx
 
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
RPMS TEMPLATE FOR SCHOOL YEAR 2023-2024 FOR TEACHER 1 TO TEACHER 3
 
C1 Rubenstein AP HuG xxxxxxxxxxxxxx.pptx
C1 Rubenstein AP HuG xxxxxxxxxxxxxx.pptxC1 Rubenstein AP HuG xxxxxxxxxxxxxx.pptx
C1 Rubenstein AP HuG xxxxxxxxxxxxxx.pptx
 
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...
Exploiting Artificial Intelligence for Empowering Researchers and Faculty, In...
 
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdfবাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
বাংলাদেশ অর্থনৈতিক সমীক্ষা (Economic Review) ২০২৪ UJS App.pdf
 
PIMS Job Advertisement 2024.pdf Islamabad
PIMS Job Advertisement 2024.pdf IslamabadPIMS Job Advertisement 2024.pdf Islamabad
PIMS Job Advertisement 2024.pdf Islamabad
 
Top five deadliest dog breeds in America
Top five deadliest dog breeds in AmericaTop five deadliest dog breeds in America
Top five deadliest dog breeds in America
 
Pollock and Snow "DEIA in the Scholarly Landscape, Session One: Setting Expec...
Pollock and Snow "DEIA in the Scholarly Landscape, Session One: Setting Expec...Pollock and Snow "DEIA in the Scholarly Landscape, Session One: Setting Expec...
Pollock and Snow "DEIA in the Scholarly Landscape, Session One: Setting Expec...
 
Chapter 4 - Islamic Financial Institutions in Malaysia.pptx
Chapter 4 - Islamic Financial Institutions in Malaysia.pptxChapter 4 - Islamic Financial Institutions in Malaysia.pptx
Chapter 4 - Islamic Financial Institutions in Malaysia.pptx
 
Pride Month Slides 2024 David Douglas School District
Pride Month Slides 2024 David Douglas School DistrictPride Month Slides 2024 David Douglas School District
Pride Month Slides 2024 David Douglas School District
 
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdfANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
ANATOMY AND BIOMECHANICS OF HIP JOINT.pdf
 

MUMS: Bayesian, Fiducial, and Frequentist Conference - Spatially Informed Variable Selection Priors and Applications to Large-scale Data (revised), Marina Vannucci, April 30, 2019

  • 1. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Spatially Informed Variable Selection Priors and Applications to Large-Scale Data Marina Vannucci Department of Statistics Rice University Houston, TX USA April 2019 - SAMSI
  • 2. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Outline of the Talk Variable selection via spike-and-slab priors Accounting for spatial information Nonparametric constructions Edge selection in graphical models Single graphs Multiple graphs Beyond Gaussian data - count data Computational issues - Variational Inference Motivating applications: High-throughput genomics; neuroimaging.
  • 3. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Spike-and-slab Variable Selection Prior for Regression Models Yn×1 = Xn×pβp×1 + ε, ε ∼ N(0, σ2 I) Introduce latent variable γ = (γ1, . . . , γp) to select variables γj = 1 if variable j included in model γj = 0 otherwise Specify priors for model parameters: βj ∼ (1 − γj)δ0 + γjN(0, σ2 τ2 j ), σ2 ∼ IG(ν/2, λ/2) p(γ) = p j=1 ωγj (1 − ω)1−γj Combine data and prior information into a posterior distribution ⇒ interest in posterior distribution p(γ|X, Y).
  • 4. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Posterior Inference via MCMC Use Metropolis as stochastic search. At each iteration (i) Add or Delete one variable in γold (ii) Swap a 0 and a 1 in γold The proposed γnew is accepted with probability min p(γnew |X, Y) p(γold |X, Y) , 1 . George & McCulloch (1993,1997); Brown et al. (1998 & 2002) Sample (β, γ) jointly if marginalization not possible Gottardo & Raftery (2008 JCGS); Savitsky et al. (2011 Stat Science) Select variables in the “best” models or from largest marginal posterior probabilities of inclusion (PPIs) via median model or Bayesian FDR thresholds.
  • 5. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Functional Magnetic Resonance Imaging (fMRI) Source:(www.anc.ed.ac.uk(( Source:(photos.uc.wisc.edu(( Indirect measure of brain activity as changes in blood flow. Collected at rest or during a sensorimotor task. Observed data - time series of the blood oxygenation level dependent (BOLD) response, at each voxel in the brain.
  • 6. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Model for Single Subject Yν = Xνβν + εν, ν = 1, . . . , V fMRI = BOLD Signal + Noise Yν = (Yν1, . . . , YνT )T time-series data for voxel ν. Xν the convolved design matrix T × p (p stimuli). Convolution of stimulus function with HRF (e.g., Poisson with parameter λν) to account for time lapse between stimulus and vascular response. βν = (βν1, . . . , βνp)T regression coefficients vector εν ∼ NT (0, Σν) caused by hardware and subjects
  • 7. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Spatial Modeling Spike-and-slab prior on to select activated voxels (p = 1) βν ∼ γνN(0, τ) + (1 − γν)δ0, ν = 1, . . . , V, MRF prior on voxel-dependent parameter γν (Zhang et al., 2014 NeuroImage) P(γν|d, e, γk , k ∈ Nν) ∝ exp(γν(d + e k∈Nν γν)) with Nν the set of neighbors of voxel ν, d ∈ R, e > 0. Probit priors (Chiang et al., 2017 Human Brain Mapping) p(γν = 1) = Φ (α0 + α1Nν) with Nν a set of covariates (e.g., multi-modal image data).
  • 8. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Spatial Modeling - Nonparametric Prior Objective: capture correlation among time series data of possibly distant voxels Proposed prior: spike-and-slab nonparametric prior, βν|γν, G0 ∼ γνG0 + (1 − γν)δ0 with G0 a Dirichlet process prior with N(0, τ) as centering distribution. Achieves clustering of voxels within subject according to similar activation strengths. Particularly useful for multi-subject modeling
  • 9. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Regression Model for Multiple Subjects Extend the single-subject regression model to multiple subjects Yiν = Xiνβiν + εiν, εiν ∼ NT (0, Σiν) Yiν = (Yiν1, . . . , YiνT )T , T × 1 BOLD response data for the νth voxel in the ith subject Xiν, T × p design matrix, with Poisson HRF βiν, p × 1 vector of regression coefficients (spatial prior) Hierarchical Dirichlet process as slab distribution on βiν
  • 10. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Spatial Nonparametric Prior Objective: capture correlation within and across subjects. Proposed prior: spike-and-slab nonparametric prior, βiν|γiν, Gi ∼ γiνGi + (1 − γiν)δ0 Clustering of subjects based on similar activation patterns; clustering of voxels within each subject according to similar activation strengths. Subject(1( Subject(3( Subject(4( Subject(2( Subject(5( (Zhang et al., 2016 Annals of Applied Statistics)
  • 11. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Posterior Inference Standard approach: MCMC MH updates to jointly update (β, γ) with a combination of add-delete-swap steps and sampling-based methods for DP models (based on the Chinese restaurant franchise) Update λ with Metropolis-Hastings (MH) schemes Update other model parameters Faster strategy: Variational Inference (Matlab GUI) Combine VI schemes with importance sampling steps on other model parameters (Carbonetto and Stephens, 2012) Posterior maps on γiν, βiν, λiν’s
  • 12. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Software: NPBayes-fMRI GUI (https://github.com/marinavannucci) 500 520 540 560 580 600 620 640 660 680 700 Cluster 1 (n = 12) Cluster 2 (n = 16) RTtoCorrectTrials(ms) Nonwords Pseudohomophones * ** * p < .05, ** p < .001 -30 -20 -10 0 10 20 30 40 50 60 PseudohomophoneEffect(ms) Cluster 1 Cluster 2 Fischer-Baum et al. (2018). Individual differences in the neural and cognitive mechanisms of single word reading. Frontiers in Human Neuroscience, 12:271. (Kook et al., 2019 Statistics in Biosciences)
  • 13. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Edge Selection in Graphical Models Undirected Graphical Models Use a graph structure to summarize conditional independence between variables Each node represents a variable No edge exists between two nodes if and only if variables are independent given all others Fundamental assumption True conditional dependence structure is reasonably sparse
  • 14. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Gaussian Graphical Model (GGM) Data follow a joint multivariate normal distribution xl ∼ N(µ, Ω−1 ) l = 1 . . . n ωi,j = 0 if no edge exists between nodes i and j in graph G For simplicity, we column-center the data to assume µ = 0
  • 15. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Choices for Edge Selection Priors G-Wishart priors enforce exact zeros but are limited in scalability (Roverato 2002, Dobra et al. 2011) Continuous shrinkage prior of Wang (BA 2015) p(Ωk |G, v0, v1, λ) ∝ i<j N(ωi,j|0, v2 gij ) i Exp(ωi,i| λ 2 ) defined by introducing binary latent variables of edge-inclusion G ≡ (gi,j)i<j ∈ G ≡ {0, 1}p(p−1)/2 Avoid use of priors over the space of positive definite matrices with fixed zeroes Improve computational efficiency
  • 16. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Bayesian Framework for Multiple Graphs Estimation xk,l ∼ N(µk , Ω−1 k ) l = 1, . . . , nk k = 1, . . . , K Graphical model for each group over common set of variables xk,l, l = 1, . . . , nk data for subject l in group k Ωk = Σ−1 k = (ωi,j,k ) is the precision matrix for group k May share common features, or could differ substantially
  • 17. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Linking Networks with a MRF Prior Encourages selection of common edges for similar groups p(gij|νij, Θ) = exp(νij 1T gij+gij T Θgij) C(νij ,Θ) with gij = {g1ij, . . . , gkij} Spike-and-slab prior on off-diagonal elements of K × K matrix Θ Nonzero entries indicate network relatedness Magnitude measures pairwise graph similarity Flexibility of model Shares information when appropriate Does not enforce similarity unless supported by data
  • 18. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Selection Prior for Network Similarity Spike and slab prior on off-diagonal entries of Θ p(θkm|γkm) = (1 − γkm)δ0 + γkm βα Γ(α) θα−1 km exp(−βθkm) Spike-and-slab prior on parameters measuring network relatedness, to allow shared structure to be favored only when supported by the data Fixed hyperparameters α and β Latent indicator variable γkm of graph relatedness Independent Bernoulli priors defined on latent indicators with probability w ∈ [0, 1]
  • 19. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Completing the Model Edge specific prior for edge inclusion probability P(νij) = 1 β(a, b) exp(aνij) (1 + exp(νij))a+b Can encourage sparsity of the graphs G1, . . . , GK Can be used to incorporate prior knowledge of connections between nodes
  • 20. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Diagram of the Joint Graph Model Peterson et al. (2015, JASA); Shaddox et al. (2019, Biostatistics)
  • 21. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Posterior Inference MCMC Sampling Scheme At each iteration t of our algorithm, sample from posterior full conditionals (i) Update graph G (t) k and precision matrix Ω (t) k for k = 1, . . . , K Block Gibbs sampler p-coupled stochastic search variable selection algorithm (ii) Update network relatedness parameters θkm and γkm for 1 ≤ k < m ≤ K using Metropolis-Hastings steps Incorporates both between-model and within-model moves (iii) Update νij for 1 ≤ i < j ≤ p using a standard Metropolis-Hastings step Model Selection Posterior marginal probabilities of edge inclusion (PPIs)
  • 22. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions [Control] 1 2 3 4 5 6 7 8 9 10 1112 13 14 15 16 17 18 19 20 21 22 23 24 25 [Mild] 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 [Moderate] 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 [Severe] 1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 17 18 19 20 21 22 23 24 25 p = 100 TPR(SE) FPR (SE) MCC(SE) AUC(SE) Fused GLasso 0.8658 (.0479) 0.2620 (.2632) 0.2775(.0940) 0.9688(.0016) Group GLasso 0.8496 (.0095) 0.1739 (.0028) 0.3089 (.0042) 0.8982(.0048) Our Method 0.7135 (.2929) 0.0053 (.0037) 0.7453 (.2334) 0.9578 (.0434) ΘM =   · .95 .97 · .95 ·   Pathway Platform Total Pairs 100 110 011 001 Total Disrupted FcγR Metabolites 73 17 5 3 18 43 (Shaddox et al., 2019 Biostatistics)
  • 23. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Beyond Gaussian Data Linear models Earlier work in GLM/GLMM models (Raftery 1996; Chen & Dunson, 2003; Kinney & Dunson, 2007) Data augmentation techniques for logit and log links (Polson et al. (2013) and Zhu et al. (2012)) Recent interest in (zero-inflated) negative binomial and Dirichlet-multinomial regression models for large-scale genomic data (e.g. microbiome) (Wadsworth et al. 2017, Koslovsky and Vannucci 2019)
  • 24. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Dirichlet-Multinomial Regression Multivariate counts modeled as Multinomial(y+, φ), with y+ = q j=1 yj, e.g. q microbial taxa, and φ ∼ Dirichlet(γ) Integrate p covariates as log(γj(xi)) = xiθj Obtain hierarchical Dirichlet-multinomial log-linear regression Yi | φi ind. ∼ Multinomial(y+, φi) φi | α, β1, . . . , βp ind. ∼ Dirichlet   exp(αj + p k=1 βkjxik )) j=1,...,q   with φi taxa-specific proportions.
  • 25. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Variable Selection Prior Spike-and-slab selects relevant taxa/predictor associations βkj | ξkj ∼ ξkj Normal(0 | t2 kj) + (1 − ξkj)δ0(βkj) Stochastic search MCMC for posterior inference 1 Draw auxiliary variables for DM(γj (x)) 2 Update α using Metropolis–Hastings 3 Update (ξ, β) jointly via SSVB MCMC (Savitsky et al., 2011) Positive Associations (MPPI > 0.981) ko00920 ko00513 ko00260 ko04660 ko00626 ko00620 ko01057 ko04512 ko04340 ko00330 f.Prevotellaceae g.Parabacteroides g.Prevotella o.Bacteroidales g.Roseburia g.Subdoligranulum g.Bacteroides g.Ruminococcus Negative Associations (MPPI > 0.981) ko00680 ko00512 ko00904 ko00500 ko00620 ko00010 ko00534 ko05200 ko00053 ko04540 ko00650 ko00983 g.Sutterella g.Prevotella g.Dialister g.Parabacteroides g.Bacteroides g.Ruminococcus o.Bacteroidales Wadsworth et al. (2017, BMC Bioinformatics); Koslovsky and Vannucci (2019, submitted) for Dirichlet-tree multinomial distributions
  • 26. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Variational Inference Variational inference turns inference into an optimization problem. Faster and more scalable than MCMC. Underlying idea: pick family of distributions q (β, γ) ∈ Q, with free variational parameters; use gradient descent to minimize KL divergence between q and posterior f (β, γ) q∗ = arg min q∈Q KL (q f) = q (β, γ) log q (β, γ) f (β, γ) dβdγ = log p (y | X) − ELBO
  • 27. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Mean Field VI Fully factorized approximation to reduce complexity q (β, γ | θ) = p k=1 q (βk , γk ; θk ) Choose approximating distributions from same family as prior distributions, to exploit conjugacy. Combine with importance sampling schemes on other model parameters (Zhang et al., 2016 AOAS) Extend to GLM models with logit and log links by incorporating data augmentation schemes of Polson et al. (2013) and Zhu et al. (2012) for logistic and Negative Binomial regression models.
  • 28. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions Summary and Conclusions Spike-and-slab priors for variable selection are well suited for applications. Flexible structure for the incorporation of external information via MRF or probit priors or spiked DP constructions. Priors for edge selection in Gaussian graphical models. Methodologies can be extended beyond Gaussian data. Variational inference approximations for scalability. Neuroimaging and genomics as emerging areas for applications. Improved performance over competitive approaches.
  • 29. Introduction Variable Selection Priors Edge Selection Beyond Gaussian Data Variational Inference Conclusions THANKS! Eric Kook, PhD student, Rice University. Yinsen Miao, PhD student, Rice University. Nathan Osborne, PhD student, Rice University. Matthew Koslovsky, postdoc, Rice University. Elin Shaddox, PhD 2019, postdoc at UC Denver Qiwei Li, Assistant Professor, UT Dallas. Duncan Wadsworth, Data Scientist, Microsoft, Seattle. Terrance Savitsky, Bureau of Labor Statistics, D.C. Alberto Cassese, Assistant Professor, Maastricht, The Netherland. Christine Peterson, Assistant Professor, MD Anderson. Francesco Stingo, Assistant professor, U. Florence, Italy. Michele Guindani, Professor, UC Irvine. HAMILL INNOVATION AWARD