SlideShare a Scribd company logo
Inference on Treatment Effects after Model Selection
Jingshen Wang
Department of Statistics, University of Michigan
April 30th, 2019
1 / 22
Estimation of treatment effects/structural parameters
Yerushalmy (1971) and Wilcox and Russell (1983)
2 / 22
Estimation of treatment effects/structural parameters
Engle et al. (1986)
2 / 22
Estimation of treatment effects/structural parameters
Pashkevich et al. (2012) and Chaudhuri et al. (2017)
2 / 22
Estimation of treatment effects/structural parameters
2 / 22
Estimation of treatment effects/structural parameters
2 / 22
Inference on α when p is large
2 / 22
Literature review
Post selection inference
Uniform Inference
Berk et al. (2013), Bachoc et al. (2016), Kuchibhotla et al. (2018)
Data Splitting
Rinaldo et al. (2016), Fithian et al. (2014)
Selective (conditional) Inference
Lee et al. (2016), Zhao et al. (2017), Tian and Taylor (2018)
3 / 22
Literature review
Post selection inference
Uniform Inference
Berk et al. (2013), Bachoc et al. (2016), Kuchibhotla et al. (2018)
Data Splitting
Rinaldo et al. (2016), Fithian et al. (2014)
Selective (conditional) Inference
Lee et al. (2016), Zhao et al. (2017), Tian and Taylor (2018)
Commonality of these different approaches: a data dependent target βM
.
In this talk: structural parameter α as target.
3 / 22
Inference on treatment effects after model selection
4 / 22
Inference on treatment effects after model selection
4 / 22
Key points of the talk
Refitting approach
αrefit is biased: over-fitting and under-fitting.
Provide statistical insight in the bias.
Develop repeated data splitting procedure to remove the bias.
Cross-fitting is not as efficient as the repeated data splitting.
5 / 22
High-dimensional approximately linear model
Model setup
Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0.
α − parameter of interest
6 / 22
High-dimensional approximately linear model
Model setup
Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0.
α − parameter of interest
D − treatment or variable of interest
6 / 22
High-dimensional approximately linear model
Model setup
Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0.
α − parameter of interest
D − treatment or variable of interest
X − high dimensional covariates (e.g. basis functions for nonparametric
regression functions)
6 / 22
High-dimensional approximately linear model
Model setup
Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0.
α − parameter of interest
D − treatment or variable of interest
X − high dimensional covariates (e.g. basis functions for nonparametric
regression functions)
ε − noise
6 / 22
High-dimensional approximately linear model
Model setup
Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0.
α − parameter of interest
D − treatment or variable of interest
X − high dimensional covariates (e.g. basis functions for nonparametric
regression functions)
ε − noise
β − sparse vector of coefficients, i.e.
M0 = {j : βj = 0, j = 1, · · · , p}, |M0| = s0 p.
6 / 22
High-dimensional approximately linear model
Model setup
Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0.
α − parameter of interest
D − treatment or variable of interest
X − high dimensional covariates (e.g. basis functions for nonparametric
regression functions)
ε − noise
β − sparse vector of coefficients, i.e.
M0 = {j : βj = 0, j = 1, · · · , p}, |M0| = s0 p.
Rn − approximation error
6 / 22
High-dimensional approximately linear model
Model setup
Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0.
α − parameter of interest
D − treatment or variable of interest
X − high dimensional covariates (e.g. basis functions for nonparametric
regression functions)
ε − noise
β − sparse vector of coefficients, i.e.
M0 = {j : βj = 0, j = 1, · · · , p}, |M0| = s0 p.
Rn − approximation error
Under Neyman-Robin causal model and the unconfoundedness assumption, α is
the causal effect.
6 / 22
Common perception and challenges for inference after refitting
A common perception
Inference after refitting is valid, because many model selection methods satisfy
the “oracle property” (Fan and Li, 2001)
lim
n→∞
P(M = M0) = 1.
Challenges
“Oracle property” requires strong stringent assumptions.
Perfect model selection does not happen with high probability in finite
samples.
7 / 22
Common perception and challenges for inference after refitting
A common perception
Inference after refitting is valid, because many model selection methods satisfy
the “oracle property” (Fan and Li, 2001)
lim
n→∞
P(M = M0) = 1.
Challenges
“Oracle property” requires strong stringent assumptions.
Perfect model selection does not happen with high probability in finite
samples.
7 / 22
Common perception and challenges for inference after refitting
A common perception
Inference after refitting is valid, because many model selection methods satisfy
the “oracle property” (Fan and Li, 2001)
lim
n→∞
P(M = M0) = 1.
Challenges
“Oracle property” requires strong stringent assumptions.
Perfect model selection does not happen with high probability in finite
samples.
7 / 22
Refitting bias if M = M0
8 / 22
Refitting bias if M = M0
8 / 22
Refitting bias if M = M0
8 / 22
Refitting bias if M = M0
8 / 22
Refitting bias based on Lasso: illustrative example
Simulation study
α = 3, β = (1, 1, 0.5, 0.5, 0, . . . , 0) ∈ Rp
(n, p) = (100, 500)
Σij = 0.9|i−j|
# Monte Carlo samples: 1000
Model selection via adaptive Lasso:
M = j ∈ {1, . . . , p} : βj = 0 ,
where
(α, β ) = arg min
α,β
1
n
n
i=1
(Yi − αDi − Xiβ)2
+ λ
p
j=1
|βj|
wj
.
9 / 22
Refitting bias based on Lasso: illustrative example
Simulation study
α = 3, β = (1, 1, 0.5, 0.5, 0, . . . , 0) ∈ Rp
(n, p) = (100, 500)
Σij = 0.9|i−j|
# Monte Carlo samples: 1000
Model selection via adaptive Lasso:
M = j ∈ {1, . . . , p} : βj = 0 ,
where
(α, β ) = arg min
α,β
1
n
n
i=1
(Yi − αDi − Xiβ)2
+ λ
p
j=1
|βj|
wj
.
Selected model size |M| is parametrized by λ.
9 / 22
Refitting bias: illustrative example
Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑
10 / 22
Refitting bias: illustrative example
Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑.
10 / 22
Refitting bias: illustrative example
Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑.
10 / 22
Refitting bias: illustrative example
Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑.
10 / 22
Refitting bias: illustrative example
Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑.
10 / 22
Refitting bias: illustrative example
Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑.
10 / 22
Refitting bias: illustrative example
Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑.
10 / 22
Refitting bias: illustrative example
Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑.
10 / 22
Refitting bias: illustrative example
Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑.
10 / 22
Percentage of under-fitted models vs. model size
Refitting bias of large models cannot be due to under-fitting.
11 / 22
Percentage of over-fitted models vs. model size
11 / 22
Percentage of over-fitted models vs. model size
Refitting bias of large models is due to over-fitting.
11 / 22
Percentage of perfect model selection vs. model size
11 / 22
Percentage of perfect model selection vs. model size
Perfect model selection never happens with high probability.
11 / 22
Summary: Refitting bias of M
αrefit − α = e1(Z
M
ZM
)−1
Z
M
ε
over-fitting
+ (D (I − PM
)D)−1
D (I − PM
)Xβ
under-fitting
,
where ZM
= (D, XM
), and PM
= XM
(X
M
XM
)−1
XM
12 / 22
Summary: Refitting bias of M
αrefit − α = e1(Z
M
ZM
)−1
Z
M
ε
over-fitting
+ (D (I − PM
)D)−1
D (I − PM
)Xβ
under-fitting
,
where ZM
= (D, XM
), and PM
= XM
(X
M
XM
)−1
XM
Over-fitting and under-fitting bias
If M ⊂ M0, αrefit has under-fitting bias (omitted variable bias).
12 / 22
Summary: Refitting bias of M
αrefit − α = e1(Z
M
ZM
)−1
Z
M
ε
over-fitting
+ (D (I − PM
)D)−1
D (I − PM
)Xβ
under-fitting
,
where ZM
= (D, XM
), and PM
= XM
(X
M
XM
)−1
XM
Over-fitting and under-fitting bias
If M ⊂ M0, αrefit has under-fitting bias (omitted variable bias).
If M0 ⊂ M, αrefit has over-fitting bias due to spurious correlation (fan)
E(αrefit − α) = E e1 Z
M
ZM
−1
Z
M
E(ε|ZM
) .
12 / 22
Summary: Refitting bias of M
αrefit − α = e1(Z
M
ZM
)−1
Z
M
ε
over-fitting
+ (D (I − PM
)D)−1
D (I − PM
)Xβ
under-fitting
,
where ZM
= (D, XM
), and PM
= XM
(X
M
XM
)−1
XM
Over-fitting and under-fitting bias
If M ⊂ M0, αrefit has under-fitting bias (omitted variable bias).
If M0 ⊂ M, αrefit has over-fitting bias due to spurious correlation (fan)
E(αrefit − α) = E e1 Z
M
ZM
−1
Z
M
E(ε|ZM
) .
Over- and under-fitting bias may occur simultaneously.
Hong et al. (2018) and Chernozhukov et al. (2018) discussed a similar bias issue.
12 / 22
Removing the over-fitting bias by data splitting
Suppose that M0 ⊂ M. Then the refitted estimator simplifies to
αrefit − α = e1(Z
M
ZM
)−1
Z
M
ε.
13 / 22
Removing the over-fitting bias by data splitting
Suppose that M0 ⊂ M. Then the refitted estimator simplifies to
αrefit − α = e1(Z
M
ZM
)−1
Z
M
ε.
Remove the over-fitting bias by data splitting (Mosteller and Tukey, 1977):
13 / 22
Removing the over-fitting bias by data splitting
Suppose that M0 ⊂ M. Then the refitted estimator simplifies to
αrefit − α = e1(Z
M
ZM
)−1
Z
M
ε.
Remove the over-fitting bias by data splitting (Mosteller and Tukey, 1977):
13 / 22
Removing the over-fitting bias by data splitting
Suppose that M0 ⊂ M. Then the refitted estimator simplifies to
αrefit − α = e1(Z
M
ZM
)−1
Z
M
ε.
Remove the over-fitting bias by data splitting (Mosteller and Tukey, 1977):
On T2, the over-fitting bias vanishes since
E(εT2 |ZM
) = 0.
13 / 22
Removing the over-fitting bias by data splitting
Suppose that M0 ⊂ M. Then the refitted estimator simplifies to
αrefit − α = e1(Z
M
ZM
)−1
Z
M
ε.
Remove the over-fitting bias by data splitting (Mosteller and Tukey, 1977):
On T2, the over-fitting bias vanishes since
E(εT2 |ZM
) = 0.
Data-splitting removes the over-fitting bias, but it increases the estimation
variability.
13 / 22
R-Split: Repeated Data Splitting
14 / 22
R-Split: Repeated Data Splitting
14 / 22
R-Split: Repeated Data Splitting
14 / 22
R-Split: Repeated Data Splitting
14 / 22
R-Split: Repeated Data Splitting
On each split, αk depends on the data and random subsample indices.
14 / 22
R-Split: Repeated Data Splitting
In theory, B → ∞ and
α = E (αk|Data) .
In practice, B is a large number, e.g. B = 1000.
15 / 22
R-Split: Repeated Data Splitting
In theory, B → ∞ and
α = E (αk|Data) .
In practice, B is a large number, e.g. B = 1000.
R-Split is similar to Bagging (Breiman, 1996).
15 / 22
R-Split: Repeated Data Splitting
In theory, B → ∞ and
α = E (αk|Data) .
In practice, B is a large number, e.g. B = 1000.
R-Split is similar to Bagging (Breiman, 1996).
Sub-samples for both estimation and model selection are random and can
overlap.
15 / 22
Why not cross-fitting?
16 / 22
Why not cross-fitting?
16 / 22
Why not cross-fitting?
16 / 22
Cross-fitting vs. R-Split
αcv − α =
1
2
Σ−1
M1
IM1
+ Σ−1
M2
IM2
1
n
n
i=1
εiZi + op(1/
√
n)
17 / 22
Cross-fitting vs. R-Split
αcv − α =
1
2
Σ−1
M1
IM1
+ Σ−1
M2
IM2
1
n
n
i=1
εiZi + op(1/
√
n)
Variance decomposition of αCV
Var(αcv − α) = E Var
1
2
Σ−1
M1
IM1
+ Σ−1
M2
IM2
1
n
n
i=1
εiZi Data
+ Var E
1
2
Σ−1
M1
IM1
+ Σ−1
M2
IM2
1
n
n
i=1
εiZi Data
variance of R-Split
≥Var(α − α)
17 / 22
Cross-fitting vs. R-Split
αcv − α =
1
2
Σ−1
M1
IM1
+ Σ−1
M2
IM2
1
n
n
i=1
εiZi + op(1/
√
n)
Variance decomposition of αCV
Var(αcv − α) = E Var
1
2
Σ−1
M1
IM1
+ Σ−1
M2
IM2
1
n
n
i=1
εiZi Data
+ Var(α − α)
≥Var(α − α)
17 / 22
Cross-fitting vs. R-Split
αcv − α =
1
2
Σ−1
M1
IM1
+ Σ−1
M2
IM2
1
n
n
i=1
εiZi + op(1/
√
n)
Variance decomposition of αCV
Var(αcv − α) = E Var
1
2
Σ−1
M1
IM1
+ Σ−1
M2
IM2
1
n
n
i=1
εiZi Data
+ Var(α − α)
≥Var(α − α)
If M1 = M2 = M0, then Var(αcv − α) = Var(α − α).
R-Split reduces the variance by aggregating over all possible random
models.
17 / 22
R-Split: Asymptotic Normality
Theorem (R-Split)
Under certain assumptions, the R-Split estimator has the following linear
representation
α − α = ηn
1
n
n
i=1
εiZi + op(1/
√
n),
and thus
σ−1
n
√
n(α − α) ; N(0, 1),
with σn = σε ηnΣnηn
1
2
, Σ = Z Z/n, and Z = (D, X).
18 / 22
R-Split: Regularity assumptions
Assumption 1. Characterization of ηn
There exists a random vector ηn ∈ Rp+1
which is independent of ε and satisfies
E P e1Σ−1
M
Data − ηn
1
= op 1/ log p ,
where P : R|M|
→ Rp+1
is an embedding that sparsifies a vector.
19 / 22
R-Split: Regularity assumptions
Assumption 1. Characterization of ηn
There exists a random vector ηn ∈ Rp+1
which is independent of ε and satisfies
E P e1Σ−1
M
Data − ηn
1
= op 1/ log p ,
where P : R|M|
→ Rp+1
is an embedding that sparsifies a vector.
Suppose M = M0 for all splits,
ηn,j =
(e1Σ−1
M0
)j if j ∈ M0,
0 otherwise,
and therefore
α − α = e1Σ−1
M0
1
n
n
i=1
εiZi,M0 + op(1/
√
n).
19 / 22
R-Split: Regularity assumptions
Assumption 1. Characterization of ηn
There exists a random vector ηn ∈ Rp+1
which is independent of ε and satisfies
E P e1Σ−1
M
Data − ηn
1
= op 1/ log p ,
where P : R|M|
→ Rp+1
is an embedding that sparsifies a vector.
Suppose M = M0 for all splits,
ηn,j =
(e1Σ−1
M0
)j if j ∈ M0,
0 otherwise,
and therefore
α − α = e1Σ−1
M0
1
n
n
i=1
εiZi,M0 + op(1/
√
n).
For fixed model M0, α reduces to OLS based on the full sample.
Our theory generalizes OLS based on fixed to random models.
19 / 22
R-Split: Regularity assumptions
Assumption 1. Characterization of ηn
There exists a random vector ηn ∈ Rp+1
which is independent of ε and satisfies
E P e1Σ−1
M
Data − ηn
1
= op 1/ log p ,
where P : R|M|
→ Rp+1
is an embedding that sparsifies a vector.
Assumption 2. (Negligible under-fitting bias)
The under-fitting bias is negligible after averaging over all splits.
19 / 22
R-Split: Regularity assumptions
Assumption 1. Characterization of ηn
There exists a random vector ηn ∈ Rp+1
which is independent of ε and satisfies
E P e1Σ−1
M
Data − ηn
1
= op 1/ log p ,
where P : R|M|
→ Rp+1
is an embedding that sparsifies a vector.
Assumption 2. (Negligible under-fitting bias)
The under-fitting bias is negligible after averaging over all splits.
Assumption 3. (“Robust” model selection procedure)
The distribution of M remains stable if only one out of n observations changes.
19 / 22
R-Split: Regularity assumptions
Assumption 1. Characterization of ηn
There exists a random vector ηn ∈ Rp+1
which is independent of ε and satisfies
E P e1Σ−1
M
Data − ηn
1
= op 1/ log p ,
where P : R|M|
→ Rp+1
is an embedding that sparsifies a vector.
Assumption 2. (Negligible under-fitting bias)
The under-fitting bias is negligible after averaging over all splits.
Assumption 3. (“Robust” model selection procedure)
The distribution of M remains stable if only one out of n observations changes.
Assumption 4. (Sparsity level)
The selected model sizes are of the same order as s0 and s0 = o(n).
19 / 22
Conclusion
Refitting approach
The bias of αrefit is composed of two parts: under-fitting and over-fitting.
R-Split (Repeated data Splitting) removes the over-fitting bias without
much sacrifice of efficiency.
R-Split is more efficient than cross-fitting.
20 / 22
Conclusion
Refitting approach
The bias of αrefit is composed of two parts: under-fitting and over-fitting.
R-Split (Repeated data Splitting) removes the over-fitting bias without
much sacrifice of efficiency.
R-Split is more efficient than cross-fitting.
Jingshen Wang, Xuming He, and Gongjun Xu. Debiased Inference on Treatment Effect
in a High Dimensional Model. Journal of the American Statistical Association, 2019.
20 / 22
Reference I
Franc¸ois Bachoc, David Preinerstorfer, and Lukas Steinberger. Uniformly valid confidence intervals post-model-selection. arXiv preprint
arXiv:1611.01043, 2016.
Richard Berk, Lawrence Brown, Andreas Buja, Kai Zhang, and Linda Zhao. Valid post-selection inference. The Annals of Statistics, 41(2):802–837,
2013.
Leo Breiman. Bagging predictors. Machine learning, 24(2):123–140, 1996.
Sougata Chaudhuri, Abraham Bagherjeiran, and James Liu. Ranking and calibrating click-attributed purchases in performance display advertising. In
Proceedings of the ADKDD’17, page 7. ACM, 2017.
Victor Chernozhukov, Denis Chetverikov, Mert Demirer, Esther Duflo, Christian Hansen, Whitney Newey, and James Robins. Double/debiased machine
learning for treatment and structural parameters. The Econometrics Journal, 21(1):C1–C68, 2018.
Bradley Efron. Estimation and accuracy after model selection. Journal of the American Statistical Association, 109(507):991–1007, 2014.
Robert F Engle, Clive WJ Granger, John Rice, and Andrew Weiss. Semiparametric estimates of the relation between weather and electricity sales.
Journal of the American statistical Association, 81(394):310–320, 1986.
Jianqing Fan and Runze Li. Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American statistical
Association, 96(456):1348–1360, 2001.
William Fithian, Dennis Sun, and Jonathan Taylor. Optimal inference after model selection. arXiv preprint arXiv:1410.2597, 2014.
Liang Hong, Todd A Kuffner, and Ryan Martin. On overfitting and post-selection uncertainty assessments. Biometrika, 105(1):221–224, 2018.
Arun Kumar Kuchibhotla, Lawrence D Brown, Andreas Buja, Edward I George, and Linda Zhao. A model free perspective for linear regression:
Uniform-in-model bounds for post selection inference. arXiv preprint arXiv:1802.05801, 2018.
Jason D Lee, Dennis L Sun, Yuekai Sun, and Jonathan E Taylor. Exact post-selection inference, with application to the lasso. The Annals of Statistics,
44(3):907–927, 2016.
Frederick Mosteller and John Wilder Tukey. Data analysis and regression: a second course in statistics. Addison-Wesley Series in Behavioral Science:
Quantitative Methods, 1977.
Max Pashkevich, Sundar Dorai-Raj, Melanie Kellar, and Dan Zigmond. Empowering online advertisements by empowering viewers with the right to
choose: the relative effectiveness of skippable video advertisements on youtube. Journal of Advertising Research, 52(4):451–457, 2012.
Alessandro Rinaldo, Larry Wasserman, Max G’Sell, Jing Lei, and Ryan Tibshirani. Bootstrapping and sample splitting for high-dimensional,
assumption-free inference. arXiv preprint arXiv:1611.05401, 2016.
Xiaoying Tian and Jonathan Taylor. Selective inference with a randomized response. The Annals of Statistics, 46(2):679–710, 2018.
Allen J Wilcox and Ian T Russell. Birthweight and perinatal mortality: I. On the frequency distribution of birthweight. International Journal of
Epidemiology, 12(3):314–318, 1983.
J Yerushalmy. The relationship of parents’ cigarette smoking to outcome of pregnancy–implications as to the problem of inferring causation from
observed associations. American Journal of Epidemiology, 93(6):443–443, 1971.
Qingyuan Zhao, Dylan S Small, and Ashkan Ertefaie. Selective inference for effect modification via the lasso. arXiv preprint arXiv:1705.08020, 2017.
21 / 22
Thank you!
22 / 22
R-Split: estimation of the variance
Estimator of the variance of α
By the non-parametric delta method, we have
σ2
n =n
n
j=1
n − 1
n − n2
B−1
B
b=1
(vbj − B−1
B
k=1
vkj)αb
2
approx. of the squared influence function
−
n2n
B2(n − n2)
B
b=1
(αb − α)2
finite “B”-bias correction
.
B : the number of the repeated data splitting
n2 : the size of the sample used for refitting
vbj =
1 if jth obs. is used for refitting in bth sub-sample
0 otherwise.
Note: this is a generalization of the nonparametric delta method for
bootstrapping in Efron (2014).
22 / 22

More Related Content

What's hot

MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...
MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...
MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...
The Statistical and Applied Mathematical Sciences Institute
 
Some sampling techniques for big data analysis
Some sampling techniques for big data analysisSome sampling techniques for big data analysis
Some sampling techniques for big data analysis
Jae-kwang Kim
 
Collaborative filtering using orthogonal nonnegative matrix
Collaborative filtering using orthogonal nonnegative matrixCollaborative filtering using orthogonal nonnegative matrix
Collaborative filtering using orthogonal nonnegative matrix
AllenWu
 
Fi review5
Fi review5Fi review5
Fi review5
Jae-kwang Kim
 
Lecture2 xing
Lecture2 xingLecture2 xing
Lecture2 xing
Tianlu Wang
 
Introductory maths analysis chapter 16 official
Introductory maths analysis   chapter 16 officialIntroductory maths analysis   chapter 16 official
Introductory maths analysis chapter 16 official
Evert Sandye Taasiringan
 
Convergence of ABC methods
Convergence of ABC methodsConvergence of ABC methods
Convergence of ABC methods
Christian Robert
 
Propensity albert
Propensity albertPropensity albert
Propensity albert
Jae-kwang Kim
 
Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach
Jae-kwang Kim
 
ABC short course: model choice chapter
ABC short course: model choice chapterABC short course: model choice chapter
ABC short course: model choice chapter
Christian Robert
 
Interpretable Sparse Sliced Inverse Regression for digitized functional data
Interpretable Sparse Sliced Inverse Regression for digitized functional dataInterpretable Sparse Sliced Inverse Regression for digitized functional data
Interpretable Sparse Sliced Inverse Regression for digitized functional data
tuxette
 
ABC short course: survey chapter
ABC short course: survey chapterABC short course: survey chapter
ABC short course: survey chapter
Christian Robert
 
Seattle.Slides.7
Seattle.Slides.7Seattle.Slides.7
Seattle.Slides.7
Andrey Skripnikov
 
ABC with data cloning for MLE in state space models
ABC with data cloning for MLE in state space modelsABC with data cloning for MLE in state space models
ABC with data cloning for MLE in state space models
Umberto Picchini
 
Approximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsApproximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forests
Christian Robert
 
ABC workshop: 17w5025
ABC workshop: 17w5025ABC workshop: 17w5025
ABC workshop: 17w5025
Christian Robert
 
PMED Transition Workshop - Non-parametric Techniques for Estimating Tumor Het...
PMED Transition Workshop - Non-parametric Techniques for Estimating Tumor Het...PMED Transition Workshop - Non-parametric Techniques for Estimating Tumor Het...
PMED Transition Workshop - Non-parametric Techniques for Estimating Tumor Het...
The Statistical and Applied Mathematical Sciences Institute
 
seminar at Princeton University
seminar at Princeton Universityseminar at Princeton University
seminar at Princeton University
Christian Robert
 
March12 natarajan
March12 natarajanMarch12 natarajan
March12 natarajan
BBKuhn
 
ABC short course: final chapters
ABC short course: final chaptersABC short course: final chapters
ABC short course: final chapters
Christian Robert
 

What's hot (20)

MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...
MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...
MUMS Opening Workshop - Model Uncertainty and Uncertain Quantification - Merl...
 
Some sampling techniques for big data analysis
Some sampling techniques for big data analysisSome sampling techniques for big data analysis
Some sampling techniques for big data analysis
 
Collaborative filtering using orthogonal nonnegative matrix
Collaborative filtering using orthogonal nonnegative matrixCollaborative filtering using orthogonal nonnegative matrix
Collaborative filtering using orthogonal nonnegative matrix
 
Fi review5
Fi review5Fi review5
Fi review5
 
Lecture2 xing
Lecture2 xingLecture2 xing
Lecture2 xing
 
Introductory maths analysis chapter 16 official
Introductory maths analysis   chapter 16 officialIntroductory maths analysis   chapter 16 official
Introductory maths analysis chapter 16 official
 
Convergence of ABC methods
Convergence of ABC methodsConvergence of ABC methods
Convergence of ABC methods
 
Propensity albert
Propensity albertPropensity albert
Propensity albert
 
Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach Chapter2: Likelihood-based approach
Chapter2: Likelihood-based approach
 
ABC short course: model choice chapter
ABC short course: model choice chapterABC short course: model choice chapter
ABC short course: model choice chapter
 
Interpretable Sparse Sliced Inverse Regression for digitized functional data
Interpretable Sparse Sliced Inverse Regression for digitized functional dataInterpretable Sparse Sliced Inverse Regression for digitized functional data
Interpretable Sparse Sliced Inverse Regression for digitized functional data
 
ABC short course: survey chapter
ABC short course: survey chapterABC short course: survey chapter
ABC short course: survey chapter
 
Seattle.Slides.7
Seattle.Slides.7Seattle.Slides.7
Seattle.Slides.7
 
ABC with data cloning for MLE in state space models
ABC with data cloning for MLE in state space modelsABC with data cloning for MLE in state space models
ABC with data cloning for MLE in state space models
 
Approximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forestsApproximate Bayesian model choice via random forests
Approximate Bayesian model choice via random forests
 
ABC workshop: 17w5025
ABC workshop: 17w5025ABC workshop: 17w5025
ABC workshop: 17w5025
 
PMED Transition Workshop - Non-parametric Techniques for Estimating Tumor Het...
PMED Transition Workshop - Non-parametric Techniques for Estimating Tumor Het...PMED Transition Workshop - Non-parametric Techniques for Estimating Tumor Het...
PMED Transition Workshop - Non-parametric Techniques for Estimating Tumor Het...
 
seminar at Princeton University
seminar at Princeton Universityseminar at Princeton University
seminar at Princeton University
 
March12 natarajan
March12 natarajanMarch12 natarajan
March12 natarajan
 
ABC short course: final chapters
ABC short course: final chaptersABC short course: final chapters
ABC short course: final chapters
 

Similar to MUMS: Bayesian, Fiducial, and Frequentist Conference - Inference on Treatment Effects after Model Selection, Jingshen Wang, April 30, 2019

unit 3 regression.pptx
unit 3 regression.pptxunit 3 regression.pptx
unit 3 regression.pptx
ssuser5c580e1
 
MUMS: Bayesian, Fiducial, and Frequentist Conference - Including Factors in B...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Including Factors in B...MUMS: Bayesian, Fiducial, and Frequentist Conference - Including Factors in B...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Including Factors in B...
The Statistical and Applied Mathematical Sciences Institute
 
Bayesian Variable Selection in Linear Regression and A Comparison
Bayesian Variable Selection in Linear Regression and A ComparisonBayesian Variable Selection in Linear Regression and A Comparison
Bayesian Variable Selection in Linear Regression and A Comparison
Atilla YARDIMCI
 
Matlab:Regression
Matlab:RegressionMatlab:Regression
Matlab:Regression
DataminingTools Inc
 
Matlab: Regression
Matlab: RegressionMatlab: Regression
Matlab: Regression
matlab Content
 
LINEAR REGRESSION.pptx
LINEAR REGRESSION.pptxLINEAR REGRESSION.pptx
LINEAR REGRESSION.pptx
neelamsanjeevkumar
 
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
The Statistical and Applied Mathematical Sciences Institute
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
Valentin De Bortoli
 
JISA_Paper
JISA_PaperJISA_Paper
Get Multiple Regression Assignment Help
Get Multiple Regression Assignment Help Get Multiple Regression Assignment Help
Get Multiple Regression Assignment Help
HelpWithAssignment.com
 
Regression Analysis
Regression AnalysisRegression Analysis
Regression Analysis
Muhammad Fazeel
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
Christian Robert
 
article_mdimagh_haddar_2012
article_mdimagh_haddar_2012article_mdimagh_haddar_2012
article_mdimagh_haddar_2012
Mdimagh Ridha
 
Asymptotic properties of bayes factor in one way repeated measurements model
Asymptotic properties of bayes factor in one  way repeated measurements modelAsymptotic properties of bayes factor in one  way repeated measurements model
Asymptotic properties of bayes factor in one way repeated measurements model
Alexander Decker
 
Asymptotic properties of bayes factor in one way repeated measurements model
Asymptotic properties of bayes factor in one  way repeated measurements modelAsymptotic properties of bayes factor in one  way repeated measurements model
Asymptotic properties of bayes factor in one way repeated measurements model
Alexander Decker
 
Ijmet 10 01_192
Ijmet 10 01_192Ijmet 10 01_192
Ijmet 10 01_192
IAEME Publication
 
Control Synthesis by Sum of Squares Optimization
Control Synthesis by Sum of Squares OptimizationControl Synthesis by Sum of Squares Optimization
Control Synthesis by Sum of Squares Optimization
Behzad Samadi
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
theijes
 
Data Analysis Assignment Help
Data Analysis Assignment Help Data Analysis Assignment Help
Data Analysis Assignment Help
Excel Homework Help
 

Similar to MUMS: Bayesian, Fiducial, and Frequentist Conference - Inference on Treatment Effects after Model Selection, Jingshen Wang, April 30, 2019 (20)

unit 3 regression.pptx
unit 3 regression.pptxunit 3 regression.pptx
unit 3 regression.pptx
 
MUMS: Bayesian, Fiducial, and Frequentist Conference - Including Factors in B...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Including Factors in B...MUMS: Bayesian, Fiducial, and Frequentist Conference - Including Factors in B...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Including Factors in B...
 
Bayesian Variable Selection in Linear Regression and A Comparison
Bayesian Variable Selection in Linear Regression and A ComparisonBayesian Variable Selection in Linear Regression and A Comparison
Bayesian Variable Selection in Linear Regression and A Comparison
 
Matlab:Regression
Matlab:RegressionMatlab:Regression
Matlab:Regression
 
Matlab: Regression
Matlab: RegressionMatlab: Regression
Matlab: Regression
 
LINEAR REGRESSION.pptx
LINEAR REGRESSION.pptxLINEAR REGRESSION.pptx
LINEAR REGRESSION.pptx
 
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
MUMS: Bayesian, Fiducial, and Frequentist Conference - Model Selection in the...
 
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
Causal Inference Opening Workshop - Bayesian Nonparametric Models for Treatme...
 
Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...Maximum likelihood estimation of regularisation parameters in inverse problem...
Maximum likelihood estimation of regularisation parameters in inverse problem...
 
JISA_Paper
JISA_PaperJISA_Paper
JISA_Paper
 
Get Multiple Regression Assignment Help
Get Multiple Regression Assignment Help Get Multiple Regression Assignment Help
Get Multiple Regression Assignment Help
 
Regression Analysis
Regression AnalysisRegression Analysis
Regression Analysis
 
ABC-Gibbs
ABC-GibbsABC-Gibbs
ABC-Gibbs
 
article_mdimagh_haddar_2012
article_mdimagh_haddar_2012article_mdimagh_haddar_2012
article_mdimagh_haddar_2012
 
Asymptotic properties of bayes factor in one way repeated measurements model
Asymptotic properties of bayes factor in one  way repeated measurements modelAsymptotic properties of bayes factor in one  way repeated measurements model
Asymptotic properties of bayes factor in one way repeated measurements model
 
Asymptotic properties of bayes factor in one way repeated measurements model
Asymptotic properties of bayes factor in one  way repeated measurements modelAsymptotic properties of bayes factor in one  way repeated measurements model
Asymptotic properties of bayes factor in one way repeated measurements model
 
Ijmet 10 01_192
Ijmet 10 01_192Ijmet 10 01_192
Ijmet 10 01_192
 
Control Synthesis by Sum of Squares Optimization
Control Synthesis by Sum of Squares OptimizationControl Synthesis by Sum of Squares Optimization
Control Synthesis by Sum of Squares Optimization
 
The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)The International Journal of Engineering and Science (The IJES)
The International Journal of Engineering and Science (The IJES)
 
Data Analysis Assignment Help
Data Analysis Assignment Help Data Analysis Assignment Help
Data Analysis Assignment Help
 

More from The Statistical and Applied Mathematical Sciences Institute

Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
The Statistical and Applied Mathematical Sciences Institute
 
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
The Statistical and Applied Mathematical Sciences Institute
 
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
The Statistical and Applied Mathematical Sciences Institute
 
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
The Statistical and Applied Mathematical Sciences Institute
 
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
The Statistical and Applied Mathematical Sciences Institute
 
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
The Statistical and Applied Mathematical Sciences Institute
 
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
The Statistical and Applied Mathematical Sciences Institute
 
2019 GDRR: Blockchain Data Analytics - Modeling Cryptocurrency Markets with T...
2019 GDRR: Blockchain Data Analytics - Modeling Cryptocurrency Markets with T...2019 GDRR: Blockchain Data Analytics - Modeling Cryptocurrency Markets with T...
2019 GDRR: Blockchain Data Analytics - Modeling Cryptocurrency Markets with T...
The Statistical and Applied Mathematical Sciences Institute
 

More from The Statistical and Applied Mathematical Sciences Institute (20)

Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
Causal Inference Opening Workshop - Latent Variable Models, Causal Inference,...
 
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
2019 Fall Series: Special Guest Lecture - 0-1 Phase Transitions in High Dimen...
 
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
Causal Inference Opening Workshop - Causal Discovery in Neuroimaging Data - F...
 
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
Causal Inference Opening Workshop - Smooth Extensions to BART for Heterogeneo...
 
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
Causal Inference Opening Workshop - A Bracketing Relationship between Differe...
 
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
Causal Inference Opening Workshop - Testing Weak Nulls in Matched Observation...
 
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...Causal Inference Opening Workshop - Difference-in-differences: more than meet...
Causal Inference Opening Workshop - Difference-in-differences: more than meet...
 
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
Causal Inference Opening Workshop - New Statistical Learning Methods for Esti...
 
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
Causal Inference Opening Workshop - Bipartite Causal Inference with Interfere...
 
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
Causal Inference Opening Workshop - Bridging the Gap Between Causal Literatur...
 
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
Causal Inference Opening Workshop - Some Applications of Reinforcement Learni...
 
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
Causal Inference Opening Workshop - Bracketing Bounds for Differences-in-Diff...
 
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
Causal Inference Opening Workshop - Assisting the Impact of State Polcies: Br...
 
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
Causal Inference Opening Workshop - Experimenting in Equilibrium - Stefan Wag...
 
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
Causal Inference Opening Workshop - Targeted Learning for Causal Inference Ba...
 
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
2019 Fall Series: Special Guest Lecture - Adversarial Risk Analysis of the Ge...
 
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
2019 Fall Series: Professional Development, Writing Academic Papers…What Work...
 
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
2019 GDRR: Blockchain Data Analytics - Machine Learning in/for Blockchain: Fu...
 
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
2019 GDRR: Blockchain Data Analytics - QuTrack: Model Life Cycle Management f...
 
2019 GDRR: Blockchain Data Analytics - Modeling Cryptocurrency Markets with T...
2019 GDRR: Blockchain Data Analytics - Modeling Cryptocurrency Markets with T...2019 GDRR: Blockchain Data Analytics - Modeling Cryptocurrency Markets with T...
2019 GDRR: Blockchain Data Analytics - Modeling Cryptocurrency Markets with T...
 

Recently uploaded

How to Setup Default Value for a Field in Odoo 17
How to Setup Default Value for a Field in Odoo 17How to Setup Default Value for a Field in Odoo 17
How to Setup Default Value for a Field in Odoo 17
Celine George
 
How to Download & Install Module From the Odoo App Store in Odoo 17
How to Download & Install Module From the Odoo App Store in Odoo 17How to Download & Install Module From the Odoo App Store in Odoo 17
How to Download & Install Module From the Odoo App Store in Odoo 17
Celine George
 
RESULTS OF THE EVALUATION QUESTIONNAIRE.pptx
RESULTS OF THE EVALUATION QUESTIONNAIRE.pptxRESULTS OF THE EVALUATION QUESTIONNAIRE.pptx
RESULTS OF THE EVALUATION QUESTIONNAIRE.pptx
zuzanka
 
Level 3 NCEA - NZ: A Nation In the Making 1872 - 1900 SML.ppt
Level 3 NCEA - NZ: A  Nation In the Making 1872 - 1900 SML.pptLevel 3 NCEA - NZ: A  Nation In the Making 1872 - 1900 SML.ppt
Level 3 NCEA - NZ: A Nation In the Making 1872 - 1900 SML.ppt
Henry Hollis
 
Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...
Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...
Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...
TechSoup
 
78 Microsoft-Publisher - Sirin Sultana Bora.pptx
78 Microsoft-Publisher - Sirin Sultana Bora.pptx78 Microsoft-Publisher - Sirin Sultana Bora.pptx
78 Microsoft-Publisher - Sirin Sultana Bora.pptx
Kalna College
 
Skimbleshanks-The-Railway-Cat by T S Eliot
Skimbleshanks-The-Railway-Cat by T S EliotSkimbleshanks-The-Railway-Cat by T S Eliot
Skimbleshanks-The-Railway-Cat by T S Eliot
nitinpv4ai
 
Educational Technology in the Health Sciences
Educational Technology in the Health SciencesEducational Technology in the Health Sciences
Educational Technology in the Health Sciences
Iris Thiele Isip-Tan
 
Geography as a Discipline Chapter 1 __ Class 11 Geography NCERT _ Class Notes...
Geography as a Discipline Chapter 1 __ Class 11 Geography NCERT _ Class Notes...Geography as a Discipline Chapter 1 __ Class 11 Geography NCERT _ Class Notes...
Geography as a Discipline Chapter 1 __ Class 11 Geography NCERT _ Class Notes...
ImMuslim
 
220711130100 udita Chakraborty Aims and objectives of national policy on inf...
220711130100 udita Chakraborty  Aims and objectives of national policy on inf...220711130100 udita Chakraborty  Aims and objectives of national policy on inf...
220711130100 udita Chakraborty Aims and objectives of national policy on inf...
Kalna College
 
Information and Communication Technology in Education
Information and Communication Technology in EducationInformation and Communication Technology in Education
Information and Communication Technology in Education
MJDuyan
 
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 8 - CẢ NĂM - FRIENDS PLUS - NĂM HỌC 2023-2024 (B...
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 8 - CẢ NĂM - FRIENDS PLUS - NĂM HỌC 2023-2024 (B...BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 8 - CẢ NĂM - FRIENDS PLUS - NĂM HỌC 2023-2024 (B...
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 8 - CẢ NĂM - FRIENDS PLUS - NĂM HỌC 2023-2024 (B...
Nguyen Thanh Tu Collection
 
Ch-4 Forest Society and colonialism 2.pdf
Ch-4 Forest Society and colonialism 2.pdfCh-4 Forest Society and colonialism 2.pdf
Ch-4 Forest Society and colonialism 2.pdf
lakshayrojroj
 
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
EduSkills OECD
 
CHUYÊN ĐỀ ÔN TẬP VÀ PHÁT TRIỂN CÂU HỎI TRONG ĐỀ MINH HỌA THI TỐT NGHIỆP THPT ...
CHUYÊN ĐỀ ÔN TẬP VÀ PHÁT TRIỂN CÂU HỎI TRONG ĐỀ MINH HỌA THI TỐT NGHIỆP THPT ...CHUYÊN ĐỀ ÔN TẬP VÀ PHÁT TRIỂN CÂU HỎI TRONG ĐỀ MINH HỌA THI TỐT NGHIỆP THPT ...
CHUYÊN ĐỀ ÔN TẬP VÀ PHÁT TRIỂN CÂU HỎI TRONG ĐỀ MINH HỌA THI TỐT NGHIỆP THPT ...
Nguyen Thanh Tu Collection
 
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) Curriculum
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) CurriculumPhilippine Edukasyong Pantahanan at Pangkabuhayan (EPP) Curriculum
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) Curriculum
MJDuyan
 
220711130083 SUBHASHREE RAKSHIT Internet resources for social science
220711130083 SUBHASHREE RAKSHIT  Internet resources for social science220711130083 SUBHASHREE RAKSHIT  Internet resources for social science
220711130083 SUBHASHREE RAKSHIT Internet resources for social science
Kalna College
 
Pharmaceutics Pharmaceuticals best of brub
Pharmaceutics Pharmaceuticals best of brubPharmaceutics Pharmaceuticals best of brub
Pharmaceutics Pharmaceuticals best of brub
danielkiash986
 
NIPER 2024 MEMORY BASED QUESTIONS.ANSWERS TO NIPER 2024 QUESTIONS.NIPER JEE 2...
NIPER 2024 MEMORY BASED QUESTIONS.ANSWERS TO NIPER 2024 QUESTIONS.NIPER JEE 2...NIPER 2024 MEMORY BASED QUESTIONS.ANSWERS TO NIPER 2024 QUESTIONS.NIPER JEE 2...
NIPER 2024 MEMORY BASED QUESTIONS.ANSWERS TO NIPER 2024 QUESTIONS.NIPER JEE 2...
Payaamvohra1
 
HYPERTENSION - SLIDE SHARE PRESENTATION.
HYPERTENSION - SLIDE SHARE PRESENTATION.HYPERTENSION - SLIDE SHARE PRESENTATION.
HYPERTENSION - SLIDE SHARE PRESENTATION.
deepaannamalai16
 

Recently uploaded (20)

How to Setup Default Value for a Field in Odoo 17
How to Setup Default Value for a Field in Odoo 17How to Setup Default Value for a Field in Odoo 17
How to Setup Default Value for a Field in Odoo 17
 
How to Download & Install Module From the Odoo App Store in Odoo 17
How to Download & Install Module From the Odoo App Store in Odoo 17How to Download & Install Module From the Odoo App Store in Odoo 17
How to Download & Install Module From the Odoo App Store in Odoo 17
 
RESULTS OF THE EVALUATION QUESTIONNAIRE.pptx
RESULTS OF THE EVALUATION QUESTIONNAIRE.pptxRESULTS OF THE EVALUATION QUESTIONNAIRE.pptx
RESULTS OF THE EVALUATION QUESTIONNAIRE.pptx
 
Level 3 NCEA - NZ: A Nation In the Making 1872 - 1900 SML.ppt
Level 3 NCEA - NZ: A  Nation In the Making 1872 - 1900 SML.pptLevel 3 NCEA - NZ: A  Nation In the Making 1872 - 1900 SML.ppt
Level 3 NCEA - NZ: A Nation In the Making 1872 - 1900 SML.ppt
 
Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...
Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...
Elevate Your Nonprofit's Online Presence_ A Guide to Effective SEO Strategies...
 
78 Microsoft-Publisher - Sirin Sultana Bora.pptx
78 Microsoft-Publisher - Sirin Sultana Bora.pptx78 Microsoft-Publisher - Sirin Sultana Bora.pptx
78 Microsoft-Publisher - Sirin Sultana Bora.pptx
 
Skimbleshanks-The-Railway-Cat by T S Eliot
Skimbleshanks-The-Railway-Cat by T S EliotSkimbleshanks-The-Railway-Cat by T S Eliot
Skimbleshanks-The-Railway-Cat by T S Eliot
 
Educational Technology in the Health Sciences
Educational Technology in the Health SciencesEducational Technology in the Health Sciences
Educational Technology in the Health Sciences
 
Geography as a Discipline Chapter 1 __ Class 11 Geography NCERT _ Class Notes...
Geography as a Discipline Chapter 1 __ Class 11 Geography NCERT _ Class Notes...Geography as a Discipline Chapter 1 __ Class 11 Geography NCERT _ Class Notes...
Geography as a Discipline Chapter 1 __ Class 11 Geography NCERT _ Class Notes...
 
220711130100 udita Chakraborty Aims and objectives of national policy on inf...
220711130100 udita Chakraborty  Aims and objectives of national policy on inf...220711130100 udita Chakraborty  Aims and objectives of national policy on inf...
220711130100 udita Chakraborty Aims and objectives of national policy on inf...
 
Information and Communication Technology in Education
Information and Communication Technology in EducationInformation and Communication Technology in Education
Information and Communication Technology in Education
 
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 8 - CẢ NĂM - FRIENDS PLUS - NĂM HỌC 2023-2024 (B...
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 8 - CẢ NĂM - FRIENDS PLUS - NĂM HỌC 2023-2024 (B...BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 8 - CẢ NĂM - FRIENDS PLUS - NĂM HỌC 2023-2024 (B...
BÀI TẬP BỔ TRỢ TIẾNG ANH LỚP 8 - CẢ NĂM - FRIENDS PLUS - NĂM HỌC 2023-2024 (B...
 
Ch-4 Forest Society and colonialism 2.pdf
Ch-4 Forest Society and colonialism 2.pdfCh-4 Forest Society and colonialism 2.pdf
Ch-4 Forest Society and colonialism 2.pdf
 
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
Andreas Schleicher presents PISA 2022 Volume III - Creative Thinking - 18 Jun...
 
CHUYÊN ĐỀ ÔN TẬP VÀ PHÁT TRIỂN CÂU HỎI TRONG ĐỀ MINH HỌA THI TỐT NGHIỆP THPT ...
CHUYÊN ĐỀ ÔN TẬP VÀ PHÁT TRIỂN CÂU HỎI TRONG ĐỀ MINH HỌA THI TỐT NGHIỆP THPT ...CHUYÊN ĐỀ ÔN TẬP VÀ PHÁT TRIỂN CÂU HỎI TRONG ĐỀ MINH HỌA THI TỐT NGHIỆP THPT ...
CHUYÊN ĐỀ ÔN TẬP VÀ PHÁT TRIỂN CÂU HỎI TRONG ĐỀ MINH HỌA THI TỐT NGHIỆP THPT ...
 
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) Curriculum
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) CurriculumPhilippine Edukasyong Pantahanan at Pangkabuhayan (EPP) Curriculum
Philippine Edukasyong Pantahanan at Pangkabuhayan (EPP) Curriculum
 
220711130083 SUBHASHREE RAKSHIT Internet resources for social science
220711130083 SUBHASHREE RAKSHIT  Internet resources for social science220711130083 SUBHASHREE RAKSHIT  Internet resources for social science
220711130083 SUBHASHREE RAKSHIT Internet resources for social science
 
Pharmaceutics Pharmaceuticals best of brub
Pharmaceutics Pharmaceuticals best of brubPharmaceutics Pharmaceuticals best of brub
Pharmaceutics Pharmaceuticals best of brub
 
NIPER 2024 MEMORY BASED QUESTIONS.ANSWERS TO NIPER 2024 QUESTIONS.NIPER JEE 2...
NIPER 2024 MEMORY BASED QUESTIONS.ANSWERS TO NIPER 2024 QUESTIONS.NIPER JEE 2...NIPER 2024 MEMORY BASED QUESTIONS.ANSWERS TO NIPER 2024 QUESTIONS.NIPER JEE 2...
NIPER 2024 MEMORY BASED QUESTIONS.ANSWERS TO NIPER 2024 QUESTIONS.NIPER JEE 2...
 
HYPERTENSION - SLIDE SHARE PRESENTATION.
HYPERTENSION - SLIDE SHARE PRESENTATION.HYPERTENSION - SLIDE SHARE PRESENTATION.
HYPERTENSION - SLIDE SHARE PRESENTATION.
 

MUMS: Bayesian, Fiducial, and Frequentist Conference - Inference on Treatment Effects after Model Selection, Jingshen Wang, April 30, 2019

  • 1. Inference on Treatment Effects after Model Selection Jingshen Wang Department of Statistics, University of Michigan April 30th, 2019 1 / 22
  • 2. Estimation of treatment effects/structural parameters Yerushalmy (1971) and Wilcox and Russell (1983) 2 / 22
  • 3. Estimation of treatment effects/structural parameters Engle et al. (1986) 2 / 22
  • 4. Estimation of treatment effects/structural parameters Pashkevich et al. (2012) and Chaudhuri et al. (2017) 2 / 22
  • 5. Estimation of treatment effects/structural parameters 2 / 22
  • 6. Estimation of treatment effects/structural parameters 2 / 22
  • 7. Inference on α when p is large 2 / 22
  • 8. Literature review Post selection inference Uniform Inference Berk et al. (2013), Bachoc et al. (2016), Kuchibhotla et al. (2018) Data Splitting Rinaldo et al. (2016), Fithian et al. (2014) Selective (conditional) Inference Lee et al. (2016), Zhao et al. (2017), Tian and Taylor (2018) 3 / 22
  • 9. Literature review Post selection inference Uniform Inference Berk et al. (2013), Bachoc et al. (2016), Kuchibhotla et al. (2018) Data Splitting Rinaldo et al. (2016), Fithian et al. (2014) Selective (conditional) Inference Lee et al. (2016), Zhao et al. (2017), Tian and Taylor (2018) Commonality of these different approaches: a data dependent target βM . In this talk: structural parameter α as target. 3 / 22
  • 10. Inference on treatment effects after model selection 4 / 22
  • 11. Inference on treatment effects after model selection 4 / 22
  • 12. Key points of the talk Refitting approach αrefit is biased: over-fitting and under-fitting. Provide statistical insight in the bias. Develop repeated data splitting procedure to remove the bias. Cross-fitting is not as efficient as the repeated data splitting. 5 / 22
  • 13. High-dimensional approximately linear model Model setup Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0. α − parameter of interest 6 / 22
  • 14. High-dimensional approximately linear model Model setup Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0. α − parameter of interest D − treatment or variable of interest 6 / 22
  • 15. High-dimensional approximately linear model Model setup Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0. α − parameter of interest D − treatment or variable of interest X − high dimensional covariates (e.g. basis functions for nonparametric regression functions) 6 / 22
  • 16. High-dimensional approximately linear model Model setup Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0. α − parameter of interest D − treatment or variable of interest X − high dimensional covariates (e.g. basis functions for nonparametric regression functions) ε − noise 6 / 22
  • 17. High-dimensional approximately linear model Model setup Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0. α − parameter of interest D − treatment or variable of interest X − high dimensional covariates (e.g. basis functions for nonparametric regression functions) ε − noise β − sparse vector of coefficients, i.e. M0 = {j : βj = 0, j = 1, · · · , p}, |M0| = s0 p. 6 / 22
  • 18. High-dimensional approximately linear model Model setup Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0. α − parameter of interest D − treatment or variable of interest X − high dimensional covariates (e.g. basis functions for nonparametric regression functions) ε − noise β − sparse vector of coefficients, i.e. M0 = {j : βj = 0, j = 1, · · · , p}, |M0| = s0 p. Rn − approximation error 6 / 22
  • 19. High-dimensional approximately linear model Model setup Y = αD + Xβ + Rn + ε, E(ε|D, X) = 0. α − parameter of interest D − treatment or variable of interest X − high dimensional covariates (e.g. basis functions for nonparametric regression functions) ε − noise β − sparse vector of coefficients, i.e. M0 = {j : βj = 0, j = 1, · · · , p}, |M0| = s0 p. Rn − approximation error Under Neyman-Robin causal model and the unconfoundedness assumption, α is the causal effect. 6 / 22
  • 20. Common perception and challenges for inference after refitting A common perception Inference after refitting is valid, because many model selection methods satisfy the “oracle property” (Fan and Li, 2001) lim n→∞ P(M = M0) = 1. Challenges “Oracle property” requires strong stringent assumptions. Perfect model selection does not happen with high probability in finite samples. 7 / 22
  • 21. Common perception and challenges for inference after refitting A common perception Inference after refitting is valid, because many model selection methods satisfy the “oracle property” (Fan and Li, 2001) lim n→∞ P(M = M0) = 1. Challenges “Oracle property” requires strong stringent assumptions. Perfect model selection does not happen with high probability in finite samples. 7 / 22
  • 22. Common perception and challenges for inference after refitting A common perception Inference after refitting is valid, because many model selection methods satisfy the “oracle property” (Fan and Li, 2001) lim n→∞ P(M = M0) = 1. Challenges “Oracle property” requires strong stringent assumptions. Perfect model selection does not happen with high probability in finite samples. 7 / 22
  • 23. Refitting bias if M = M0 8 / 22
  • 24. Refitting bias if M = M0 8 / 22
  • 25. Refitting bias if M = M0 8 / 22
  • 26. Refitting bias if M = M0 8 / 22
  • 27. Refitting bias based on Lasso: illustrative example Simulation study α = 3, β = (1, 1, 0.5, 0.5, 0, . . . , 0) ∈ Rp (n, p) = (100, 500) Σij = 0.9|i−j| # Monte Carlo samples: 1000 Model selection via adaptive Lasso: M = j ∈ {1, . . . , p} : βj = 0 , where (α, β ) = arg min α,β 1 n n i=1 (Yi − αDi − Xiβ)2 + λ p j=1 |βj| wj . 9 / 22
  • 28. Refitting bias based on Lasso: illustrative example Simulation study α = 3, β = (1, 1, 0.5, 0.5, 0, . . . , 0) ∈ Rp (n, p) = (100, 500) Σij = 0.9|i−j| # Monte Carlo samples: 1000 Model selection via adaptive Lasso: M = j ∈ {1, . . . , p} : βj = 0 , where (α, β ) = arg min α,β 1 n n i=1 (Yi − αDi − Xiβ)2 + λ p j=1 |βj| wj . Selected model size |M| is parametrized by λ. 9 / 22
  • 29. Refitting bias: illustrative example Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑ 10 / 22
  • 30. Refitting bias: illustrative example Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑. 10 / 22
  • 31. Refitting bias: illustrative example Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑. 10 / 22
  • 32. Refitting bias: illustrative example Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑. 10 / 22
  • 33. Refitting bias: illustrative example Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑. 10 / 22
  • 34. Refitting bias: illustrative example Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑. 10 / 22
  • 35. Refitting bias: illustrative example Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑. 10 / 22
  • 36. Refitting bias: illustrative example Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑. 10 / 22
  • 37. Refitting bias: illustrative example Note: a smaller λ yields a larger model, i.e. (− log λ) ↑ ⇒ |M| ↑. 10 / 22
  • 38. Percentage of under-fitted models vs. model size Refitting bias of large models cannot be due to under-fitting. 11 / 22
  • 39. Percentage of over-fitted models vs. model size 11 / 22
  • 40. Percentage of over-fitted models vs. model size Refitting bias of large models is due to over-fitting. 11 / 22
  • 41. Percentage of perfect model selection vs. model size 11 / 22
  • 42. Percentage of perfect model selection vs. model size Perfect model selection never happens with high probability. 11 / 22
  • 43. Summary: Refitting bias of M αrefit − α = e1(Z M ZM )−1 Z M ε over-fitting + (D (I − PM )D)−1 D (I − PM )Xβ under-fitting , where ZM = (D, XM ), and PM = XM (X M XM )−1 XM 12 / 22
  • 44. Summary: Refitting bias of M αrefit − α = e1(Z M ZM )−1 Z M ε over-fitting + (D (I − PM )D)−1 D (I − PM )Xβ under-fitting , where ZM = (D, XM ), and PM = XM (X M XM )−1 XM Over-fitting and under-fitting bias If M ⊂ M0, αrefit has under-fitting bias (omitted variable bias). 12 / 22
  • 45. Summary: Refitting bias of M αrefit − α = e1(Z M ZM )−1 Z M ε over-fitting + (D (I − PM )D)−1 D (I − PM )Xβ under-fitting , where ZM = (D, XM ), and PM = XM (X M XM )−1 XM Over-fitting and under-fitting bias If M ⊂ M0, αrefit has under-fitting bias (omitted variable bias). If M0 ⊂ M, αrefit has over-fitting bias due to spurious correlation (fan) E(αrefit − α) = E e1 Z M ZM −1 Z M E(ε|ZM ) . 12 / 22
  • 46. Summary: Refitting bias of M αrefit − α = e1(Z M ZM )−1 Z M ε over-fitting + (D (I − PM )D)−1 D (I − PM )Xβ under-fitting , where ZM = (D, XM ), and PM = XM (X M XM )−1 XM Over-fitting and under-fitting bias If M ⊂ M0, αrefit has under-fitting bias (omitted variable bias). If M0 ⊂ M, αrefit has over-fitting bias due to spurious correlation (fan) E(αrefit − α) = E e1 Z M ZM −1 Z M E(ε|ZM ) . Over- and under-fitting bias may occur simultaneously. Hong et al. (2018) and Chernozhukov et al. (2018) discussed a similar bias issue. 12 / 22
  • 47. Removing the over-fitting bias by data splitting Suppose that M0 ⊂ M. Then the refitted estimator simplifies to αrefit − α = e1(Z M ZM )−1 Z M ε. 13 / 22
  • 48. Removing the over-fitting bias by data splitting Suppose that M0 ⊂ M. Then the refitted estimator simplifies to αrefit − α = e1(Z M ZM )−1 Z M ε. Remove the over-fitting bias by data splitting (Mosteller and Tukey, 1977): 13 / 22
  • 49. Removing the over-fitting bias by data splitting Suppose that M0 ⊂ M. Then the refitted estimator simplifies to αrefit − α = e1(Z M ZM )−1 Z M ε. Remove the over-fitting bias by data splitting (Mosteller and Tukey, 1977): 13 / 22
  • 50. Removing the over-fitting bias by data splitting Suppose that M0 ⊂ M. Then the refitted estimator simplifies to αrefit − α = e1(Z M ZM )−1 Z M ε. Remove the over-fitting bias by data splitting (Mosteller and Tukey, 1977): On T2, the over-fitting bias vanishes since E(εT2 |ZM ) = 0. 13 / 22
  • 51. Removing the over-fitting bias by data splitting Suppose that M0 ⊂ M. Then the refitted estimator simplifies to αrefit − α = e1(Z M ZM )−1 Z M ε. Remove the over-fitting bias by data splitting (Mosteller and Tukey, 1977): On T2, the over-fitting bias vanishes since E(εT2 |ZM ) = 0. Data-splitting removes the over-fitting bias, but it increases the estimation variability. 13 / 22
  • 52. R-Split: Repeated Data Splitting 14 / 22
  • 53. R-Split: Repeated Data Splitting 14 / 22
  • 54. R-Split: Repeated Data Splitting 14 / 22
  • 55. R-Split: Repeated Data Splitting 14 / 22
  • 56. R-Split: Repeated Data Splitting On each split, αk depends on the data and random subsample indices. 14 / 22
  • 57. R-Split: Repeated Data Splitting In theory, B → ∞ and α = E (αk|Data) . In practice, B is a large number, e.g. B = 1000. 15 / 22
  • 58. R-Split: Repeated Data Splitting In theory, B → ∞ and α = E (αk|Data) . In practice, B is a large number, e.g. B = 1000. R-Split is similar to Bagging (Breiman, 1996). 15 / 22
  • 59. R-Split: Repeated Data Splitting In theory, B → ∞ and α = E (αk|Data) . In practice, B is a large number, e.g. B = 1000. R-Split is similar to Bagging (Breiman, 1996). Sub-samples for both estimation and model selection are random and can overlap. 15 / 22
  • 63. Cross-fitting vs. R-Split αcv − α = 1 2 Σ−1 M1 IM1 + Σ−1 M2 IM2 1 n n i=1 εiZi + op(1/ √ n) 17 / 22
  • 64. Cross-fitting vs. R-Split αcv − α = 1 2 Σ−1 M1 IM1 + Σ−1 M2 IM2 1 n n i=1 εiZi + op(1/ √ n) Variance decomposition of αCV Var(αcv − α) = E Var 1 2 Σ−1 M1 IM1 + Σ−1 M2 IM2 1 n n i=1 εiZi Data + Var E 1 2 Σ−1 M1 IM1 + Σ−1 M2 IM2 1 n n i=1 εiZi Data variance of R-Split ≥Var(α − α) 17 / 22
  • 65. Cross-fitting vs. R-Split αcv − α = 1 2 Σ−1 M1 IM1 + Σ−1 M2 IM2 1 n n i=1 εiZi + op(1/ √ n) Variance decomposition of αCV Var(αcv − α) = E Var 1 2 Σ−1 M1 IM1 + Σ−1 M2 IM2 1 n n i=1 εiZi Data + Var(α − α) ≥Var(α − α) 17 / 22
  • 66. Cross-fitting vs. R-Split αcv − α = 1 2 Σ−1 M1 IM1 + Σ−1 M2 IM2 1 n n i=1 εiZi + op(1/ √ n) Variance decomposition of αCV Var(αcv − α) = E Var 1 2 Σ−1 M1 IM1 + Σ−1 M2 IM2 1 n n i=1 εiZi Data + Var(α − α) ≥Var(α − α) If M1 = M2 = M0, then Var(αcv − α) = Var(α − α). R-Split reduces the variance by aggregating over all possible random models. 17 / 22
  • 67. R-Split: Asymptotic Normality Theorem (R-Split) Under certain assumptions, the R-Split estimator has the following linear representation α − α = ηn 1 n n i=1 εiZi + op(1/ √ n), and thus σ−1 n √ n(α − α) ; N(0, 1), with σn = σε ηnΣnηn 1 2 , Σ = Z Z/n, and Z = (D, X). 18 / 22
  • 68. R-Split: Regularity assumptions Assumption 1. Characterization of ηn There exists a random vector ηn ∈ Rp+1 which is independent of ε and satisfies E P e1Σ−1 M Data − ηn 1 = op 1/ log p , where P : R|M| → Rp+1 is an embedding that sparsifies a vector. 19 / 22
  • 69. R-Split: Regularity assumptions Assumption 1. Characterization of ηn There exists a random vector ηn ∈ Rp+1 which is independent of ε and satisfies E P e1Σ−1 M Data − ηn 1 = op 1/ log p , where P : R|M| → Rp+1 is an embedding that sparsifies a vector. Suppose M = M0 for all splits, ηn,j = (e1Σ−1 M0 )j if j ∈ M0, 0 otherwise, and therefore α − α = e1Σ−1 M0 1 n n i=1 εiZi,M0 + op(1/ √ n). 19 / 22
  • 70. R-Split: Regularity assumptions Assumption 1. Characterization of ηn There exists a random vector ηn ∈ Rp+1 which is independent of ε and satisfies E P e1Σ−1 M Data − ηn 1 = op 1/ log p , where P : R|M| → Rp+1 is an embedding that sparsifies a vector. Suppose M = M0 for all splits, ηn,j = (e1Σ−1 M0 )j if j ∈ M0, 0 otherwise, and therefore α − α = e1Σ−1 M0 1 n n i=1 εiZi,M0 + op(1/ √ n). For fixed model M0, α reduces to OLS based on the full sample. Our theory generalizes OLS based on fixed to random models. 19 / 22
  • 71. R-Split: Regularity assumptions Assumption 1. Characterization of ηn There exists a random vector ηn ∈ Rp+1 which is independent of ε and satisfies E P e1Σ−1 M Data − ηn 1 = op 1/ log p , where P : R|M| → Rp+1 is an embedding that sparsifies a vector. Assumption 2. (Negligible under-fitting bias) The under-fitting bias is negligible after averaging over all splits. 19 / 22
  • 72. R-Split: Regularity assumptions Assumption 1. Characterization of ηn There exists a random vector ηn ∈ Rp+1 which is independent of ε and satisfies E P e1Σ−1 M Data − ηn 1 = op 1/ log p , where P : R|M| → Rp+1 is an embedding that sparsifies a vector. Assumption 2. (Negligible under-fitting bias) The under-fitting bias is negligible after averaging over all splits. Assumption 3. (“Robust” model selection procedure) The distribution of M remains stable if only one out of n observations changes. 19 / 22
  • 73. R-Split: Regularity assumptions Assumption 1. Characterization of ηn There exists a random vector ηn ∈ Rp+1 which is independent of ε and satisfies E P e1Σ−1 M Data − ηn 1 = op 1/ log p , where P : R|M| → Rp+1 is an embedding that sparsifies a vector. Assumption 2. (Negligible under-fitting bias) The under-fitting bias is negligible after averaging over all splits. Assumption 3. (“Robust” model selection procedure) The distribution of M remains stable if only one out of n observations changes. Assumption 4. (Sparsity level) The selected model sizes are of the same order as s0 and s0 = o(n). 19 / 22
  • 74. Conclusion Refitting approach The bias of αrefit is composed of two parts: under-fitting and over-fitting. R-Split (Repeated data Splitting) removes the over-fitting bias without much sacrifice of efficiency. R-Split is more efficient than cross-fitting. 20 / 22
  • 75. Conclusion Refitting approach The bias of αrefit is composed of two parts: under-fitting and over-fitting. R-Split (Repeated data Splitting) removes the over-fitting bias without much sacrifice of efficiency. R-Split is more efficient than cross-fitting. Jingshen Wang, Xuming He, and Gongjun Xu. Debiased Inference on Treatment Effect in a High Dimensional Model. Journal of the American Statistical Association, 2019. 20 / 22
  • 76. Reference I Franc¸ois Bachoc, David Preinerstorfer, and Lukas Steinberger. Uniformly valid confidence intervals post-model-selection. arXiv preprint arXiv:1611.01043, 2016. Richard Berk, Lawrence Brown, Andreas Buja, Kai Zhang, and Linda Zhao. Valid post-selection inference. The Annals of Statistics, 41(2):802–837, 2013. Leo Breiman. Bagging predictors. Machine learning, 24(2):123–140, 1996. Sougata Chaudhuri, Abraham Bagherjeiran, and James Liu. Ranking and calibrating click-attributed purchases in performance display advertising. In Proceedings of the ADKDD’17, page 7. ACM, 2017. Victor Chernozhukov, Denis Chetverikov, Mert Demirer, Esther Duflo, Christian Hansen, Whitney Newey, and James Robins. Double/debiased machine learning for treatment and structural parameters. The Econometrics Journal, 21(1):C1–C68, 2018. Bradley Efron. Estimation and accuracy after model selection. Journal of the American Statistical Association, 109(507):991–1007, 2014. Robert F Engle, Clive WJ Granger, John Rice, and Andrew Weiss. Semiparametric estimates of the relation between weather and electricity sales. Journal of the American statistical Association, 81(394):310–320, 1986. Jianqing Fan and Runze Li. Variable selection via nonconcave penalized likelihood and its oracle properties. Journal of the American statistical Association, 96(456):1348–1360, 2001. William Fithian, Dennis Sun, and Jonathan Taylor. Optimal inference after model selection. arXiv preprint arXiv:1410.2597, 2014. Liang Hong, Todd A Kuffner, and Ryan Martin. On overfitting and post-selection uncertainty assessments. Biometrika, 105(1):221–224, 2018. Arun Kumar Kuchibhotla, Lawrence D Brown, Andreas Buja, Edward I George, and Linda Zhao. A model free perspective for linear regression: Uniform-in-model bounds for post selection inference. arXiv preprint arXiv:1802.05801, 2018. Jason D Lee, Dennis L Sun, Yuekai Sun, and Jonathan E Taylor. Exact post-selection inference, with application to the lasso. The Annals of Statistics, 44(3):907–927, 2016. Frederick Mosteller and John Wilder Tukey. Data analysis and regression: a second course in statistics. Addison-Wesley Series in Behavioral Science: Quantitative Methods, 1977. Max Pashkevich, Sundar Dorai-Raj, Melanie Kellar, and Dan Zigmond. Empowering online advertisements by empowering viewers with the right to choose: the relative effectiveness of skippable video advertisements on youtube. Journal of Advertising Research, 52(4):451–457, 2012. Alessandro Rinaldo, Larry Wasserman, Max G’Sell, Jing Lei, and Ryan Tibshirani. Bootstrapping and sample splitting for high-dimensional, assumption-free inference. arXiv preprint arXiv:1611.05401, 2016. Xiaoying Tian and Jonathan Taylor. Selective inference with a randomized response. The Annals of Statistics, 46(2):679–710, 2018. Allen J Wilcox and Ian T Russell. Birthweight and perinatal mortality: I. On the frequency distribution of birthweight. International Journal of Epidemiology, 12(3):314–318, 1983. J Yerushalmy. The relationship of parents’ cigarette smoking to outcome of pregnancy–implications as to the problem of inferring causation from observed associations. American Journal of Epidemiology, 93(6):443–443, 1971. Qingyuan Zhao, Dylan S Small, and Ashkan Ertefaie. Selective inference for effect modification via the lasso. arXiv preprint arXiv:1705.08020, 2017. 21 / 22
  • 78. R-Split: estimation of the variance Estimator of the variance of α By the non-parametric delta method, we have σ2 n =n n j=1 n − 1 n − n2 B−1 B b=1 (vbj − B−1 B k=1 vkj)αb 2 approx. of the squared influence function − n2n B2(n − n2) B b=1 (αb − α)2 finite “B”-bias correction . B : the number of the repeated data splitting n2 : the size of the sample used for refitting vbj = 1 if jth obs. is used for refitting in bth sub-sample 0 otherwise. Note: this is a generalization of the nonparametric delta method for bootstrapping in Efron (2014). 22 / 22