SlideShare a Scribd company logo
1 of 23
Download to read offline
Regression Discontinuity Designs in Stata
Matias D. Cattaneo
University of Michigan
July 30, 2015
Overview
Main goal: learn about treatment e¤ect of policy or intervention.
If treatment randomization available, easy to estimate treatment e¤ects.
If treatment randomization not available, turn to observational studies.
I Instrumental variables.
I Selection on observables.
Regression discontinuity (RD) designs.
I Simple and objective. Requires little information, if design available.
I Might be viewed as a “local” randomized trial.
I Easy to falsify, easy to interpret.
I Careful: very local!
Overview of RD packages
https://sites.google.com/site/rdpackages
rdrobust package: estimation, inference and graphical presentation using local
polynomials, partitioning, and spacings estimators.
I rdrobust: RD inference (point estimation and CI; classic, bias-corrected, robust).
I rdbwselect: bandwidth or window selection (IK, CV, CCT).
I rdplot: plots data (with “optimal” block length).
rddensity package: discontinuity in density test at cuto¤ (a.k.a. manipulation testing)
using novel local polynomial density estimator.
I rddensity: manipulation testing using local polynomial density estimation.
I rdbwdensity: bandwidth or window selection.
rdlocrand package: covariate balance, binomial tests, randomization inference
methods (window selection & inference).
I rdrandinf: inference using randomization inference methods.
I rdwinselect: falsi…cation testing and window selection.
I rdsensitivity: treatment e¤ect models over grid of windows, CI inversion.
I rdrbounds: Rosenbaum bounds.
Randomized Control Trials
Notation: (Yi(0); Yi(1); Xi), i = 1; 2; : : : ; n.
Treatment: Ti 2 f0; 1g, Ti independent of (Yi(0); Yi(1); Xi).
Data: (Yi; Ti; Xi), i = 1; 2; : : : ; n, with
Yi =
(
Yi(0) if Ti = 0
Yi(1) if Ti = 1
Average Treatment E¤ect:
ATE = E[Yi(1) Yi(0)] = E[YijT = 1] E[YijT = 0]
Experimental Design.
Sharp RD design
Notation: (Yi(0); Yi(1); Xi), i = 1; 2; : : : ; n, Xi continuous
Treatment: Ti 2 f0; 1g, Ti = 1(Xi x).
Data: (Yi; Ti; Xi), i = 1; 2; : : : ; n, with
Yi =
(
Yi(0) if Ti = 0
Yi(1) if Ti = 1
Average Treatment E¤ect at the cuto¤:
SRD = E[Yi(1) Yi(0)jXi = x] = lim
x#x
E[YijXi = x] lim
x"x
E[YijXi = x]
Quasi-Experimental Design: “local randomization” (more later)
−0.6 −0.4 −0.2 0.0 0.2 0.4 0.6
−2
−1
0
1
2
3
Assignment variable (R)
Outcome
variable
(Y)
τ0
−0.6 −0.4 −0.2 0.0 0.2 0.4 0.6
−2
−1
0
1
2
3
Assignment variable (R)
Outcome
variable
(Y)
τ0
−0.6 −0.4 −0.2 0.0 0.2 0.4 0.6
−2
−1
0
1
2
3
Assignment variable (R)
Outcome
variable
(Y)
τ0
−0.6 −0.4 −0.2 0.0 0.2 0.4 0.6
−6
−4
−2
0
2
4
6
Assignment variable (R)
Outcome
variable
(Y)
Local Random Assignment
τ0
Empirical Illustration: Cattaneo, Frandsen & Titiunik (2015, JCI)
Problem: incumbency advantage (U.S. senate).
Data:
Yi = election outcome.
Ti = whether incumbent.
Xi = vote share previous election (x = 0).
Zi = covariates (demvoteshlag1, demvoteshlag2, dopen, etc.).
Potential outcomes:
Yi(0) = election outcome if had not been incumbent.
Yi(1) = election outcome if had been incumbent.
Causal Inference:
Yi(0) 6= YijTi = 0 and Yi(1) 6= YijTi = 1
Graphical and Falsi…cation Methods
Always plot data: main advantage of RD designs!
Plot regression functions to assess treatment e¤ect and validity.
Plot density of Xi for assessing validity; test for continuity at cuto¤ and elsewhere.
Important: use also estimators that do not “smooth-out” data.
RD Plots (Calonico, Cattaneo & Titiunik, JASA):
I Two ingredients: (i) Smoothed global polynomial …t & (ii) binned discontinuous
local-means …t.
I Two goals: (i) detention of discontinuities, & (ii) representation of variability.
I Two tuning parameters:
F G lobal p olynom ial degree (kn).
F Location (ES or Q S) and numb er of bins (Jn).
Manipulation Tests & Covariate Balance and Placebo Tests
Density tests near cuto¤:
I Idea: distribution of running variable should be similar at either side of cuto¤.
I Method 1: Histograms & Binomial count test.
I Method 2: Density Estimator at boundary.
F Pre-binned local p olynom ial m ethod – M cCrary (2008).
F New tuning-param eter-free m ethod – C attaneo, Jansson and M a (2015).
Placebo tests on pre-determined/exogenous covariates.
I Idea: zero RD treatment e¤ect for pre-determined/exogenous covariates.
I Methods: global polynomial, local polynomial, randomization-based.
Placebo tests on outcomes.
I Idea: zero RD treatment e¤ect for outcome at values other than cuto¤.
I Methods: global polynomial, local polynomial, randomization-based.
Estimation and Inference Methods
Global polynomial approach (not recommended).
Robust local polynomial inference methods.
I Bandwidth selection.
I Bias-correction.
I Con…dence intervals.
Local randomization and randomization inference methods.
I Window selection.
I Estimation and Inference methods.
I Falsi…cation, sensitivity and related methods
Conventional Local-polynomial Approach
Idea: approximate regression functions for control and treatment units locally.
“Local-linear” estimator (w/ weights K( )):
hn Xi < x : x Xi hn :
Yi = + (Xi x) + " ;i Yi = + + (Xi x) + + "+;i
I Treatment e¤ect (at the cuto¤): ^SRD = ^+ ^
Can be estimated using linear models (w/ weights K( )):
Yi = + SRD Ti + (Xi x) 1 + Ti (Xi x) 1 + "i, hn Xi hn
Once hn chosen, inference is “standard”: weighted linear models.
I Details coming up next.
Conventional Local-polynomial Approach
How to choose hn?
Imbens & Kalyanaraman (2012, ReStud): “optimal” plug-in,
^
hIK = ^
CIK n 1=5
Calonico, Cattaneo & Titiunik (2014, ECMA): re…nement of IK
^
hCCT = ^
CCCT n 1=5
Ludwig & Miller (2007, QJE): cross-validation,
^
hCV = arg min
h
n
X
i=1
w(Xi) (Yi ^1(Xi; h))2
Key idea: trade-o¤ bias and variance of ^SRD(hn). Heuristically:
" Bias(^SRD) =) # ^
h and " Var(^SRD) =) " ^
h
Local-Polynomial Methods: Bandwidth Selection
Two main methods: plug-in & cross-validation. Both MSE-optimal in some sense.
Imbens & Kalyanaraman (2012, ReStud): propose MSE-optimal rule,
hMSE = C
1=5
MSE n 1=5
CMSE = C(K)
Var(^SRD)
Bias(^SRD)2
I IK implementation: …rst-generation plug-in rule.
I CCT implementation: second-generation plug-in rule.
I They di¤er in the way Var(^SRD) and Bias(^SRD) are estimated.
Imbens & Kalyanaraman (2012, ReStud): discuss cross-validation approach,
^
hCV = arg min
h>0
CV (h) , CV (h) =
n
X
i=1
1(X ;[ ] Xi X+;[ ]) (Yi ^(Xi; h))2
,
where
I ^+;p(x; h) and ^ ;p(x; h) are local polynomials estimates.
I 2 (0; 1), X ;[ ] and X+;[ ] denote -th quantile of fXi : Xi < xg and fXi : Xi xg.
I Our implementation uses = 0:5; but this is a tuning parameter!
Conventional Approach to RD
“Local-linear” estimator (w/ weights K( )):
hn Xi < x : x Xi hn :
Yi = + (Xi x) + " ;i Yi = + + (Xi x) + + "+;i
I Treatment e¤ect (at the cuto¤): ^SRD = ^+ ^
Construct usual t-test. For H0 : SRD = 0,
^
T(hn) =
^SRD
p
^
Vn
=
^+ ^
q
^
V+;n + ^
V ;n
d N(0; 1)
95% Con…dence interval:
^
I(hn) = ^SRD 1:96
q
^
Vn
Bias-Correction Approach to RD
Note well: for usual t-test,
^
T(hMSE) =
^SRD
p
^
Vn
d N(B; 1) 6= N(0; 1), B > 0
I Bias B in RD estimator captures “curvature” of regression functions.
Undersmoothing/“Small Bias”Approach: Choose “smaller”hn... Perhaps ^
hn = 0:5 ^
hIK?
=) Not clear guidance & power loss!
Bias-correction Approach:
^
Tbc
(hn; bn) =
^SRD
^
Bn
p
^
Vn
d N(0; 1)
=) 95% Con…dence Interval: ^
Ibc(hn; bn) =
h
^SRD ^
Bn 1:96
p
^
Vn
i
How to choose bn? Same ideas as before... ^
bn = ^
C n 1=7
Robust Bias-Correction Approach to RD
Recall:
^
T (hn) =
^SRD
p
^
Vn
d N (0; 1) and ^
T
bc
(hn; bn) =
^SRD
^
Bn
p
^
Vn
d N (0; 1)
I ^
Bn is constructed to estimate leading bias B.
Robust approach:
^
Tbc
(hn; bn) =
^SRD
^
Bn
p
^
Vn
=
^SRD Bn
p
^
Vn
| {z }
d N (0;1)
+
Bn
^
Bn
p
^
Vn
| {z }
d N (0; )
Robust bias-corrected t-test:
^
Trbc
(hn; bn) =
^SRD
^
Bn
p
^
Vn + ^
Wn
=
^SRD
^
Bn
q
^
Vbc
n
d N(0; 1)
=) 95% Con…dence Interval:
^
Irbc
(hn; bn) = ^SRD
^
Bn 1:96
q
^
Vbc
n , ^
Vbc
n = ^
Vn + ^
Wn
Local-Polynomial Methods: Robust Inference
Approach 1: Undersmoothing/“Small Bias”.
^
I(hn) = ^SRD 1:96
q
^
Vn
Approach 2: Bias correction (not recommended).
^
Ibc
(hn; bn) = ^SRD
^
Bn 1:96
q
^
Vn
Approach 3: Robust Bias correction.
^
Irbc
(hn; bn) = ^SRD
^
Bn 1:96
q
^
Vn + ^
Wn
Local-randomization approach and …nite-sample inference
Popular approach: local-polynomial methods.
I Approximates regression function and relies on continuity assumptions.
I Requires: choosing weights, bandwidth and polynomial order.
Alternative approach: local-randomization + randomization-inference
I Gives an alternative that can be used as a robustness check.
I Key assumption: exists window W = [ hn; hn] around cuto¤ ( hn < x < hn) where
Ti independent of (Yi(0); Yi(1)) (for all Xi 2 W )
I In words: treatment is randomly assigned within W .
I Good news: if plausible, then RCT ideas/methods apply.
I Not-so-good news: most plausible for very small windows (very few observations).
I One solution: employ small window but use randomization-inference methods.
I Requires: choosing randomization rule, window and statistic.
Local-randomization approach and …nite-sample inference
Recall key assumption: exists W = [ hn; hn] around cuto¤ ( hn < x < hn) where
Ti independent of (Yi(0); Yi(1)) (for all Xi 2 W)
How to choose window?
I Use balance tests on pre-determined/exogenous covariates.
I Very intuitive, easy to implement.
How to conduct inference? Use randomization-inference methods.
1 Choose statistic of interest. E.g., t-stat for di¤erence-in-means.
2 Choose randomization rule. E.g., number of treatments and controls given.
3 Compute …nite-sample distribution of statistics by permuting treatment assignments.
Local-randomization approach and …nite-sample inference
Do not forget to validate & falsify the empirical strategy.
1 Plot data to make sure local-randomization is plausible.
2 Conduct placebo tests.
(e.g., use pre-intervention outcomes or other covariates not used select W )
3 Do sensitivity analysis.
See Cattaneo, Frandsen and Titiunik (2015) for introduction.
See Cattaneo, Titiunik and Vazquez-Bare (2015) for further results and
implementation.

More Related Content

Similar to columbus15_cattaneo.pdf

Introduction to Supervised ML Concepts and Algorithms
Introduction to Supervised ML Concepts and AlgorithmsIntroduction to Supervised ML Concepts and Algorithms
Introduction to Supervised ML Concepts and AlgorithmsNBER
 
Nonparametric Density Estimation
Nonparametric Density EstimationNonparametric Density Estimation
Nonparametric Density Estimationjachno
 
Unbiased Bayes for Big Data
Unbiased Bayes for Big DataUnbiased Bayes for Big Data
Unbiased Bayes for Big DataChristian Robert
 
Computational Information Geometry on Matrix Manifolds (ICTP 2013)
Computational Information Geometry on Matrix Manifolds (ICTP 2013)Computational Information Geometry on Matrix Manifolds (ICTP 2013)
Computational Information Geometry on Matrix Manifolds (ICTP 2013)Frank Nielsen
 
Inference for stochastic differential equations via approximate Bayesian comp...
Inference for stochastic differential equations via approximate Bayesian comp...Inference for stochastic differential equations via approximate Bayesian comp...
Inference for stochastic differential equations via approximate Bayesian comp...Umberto Picchini
 
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...Chiheb Ben Hammouda
 
Thesis seminar
Thesis seminarThesis seminar
Thesis seminargvesom
 
Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...Frank Nielsen
 
Intro to Quant Trading Strategies (Lecture 7 of 10)
Intro to Quant Trading Strategies (Lecture 7 of 10)Intro to Quant Trading Strategies (Lecture 7 of 10)
Intro to Quant Trading Strategies (Lecture 7 of 10)Adrian Aley
 
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...Chiheb Ben Hammouda
 
Low Complexity Regularization of Inverse Problems
Low Complexity Regularization of Inverse ProblemsLow Complexity Regularization of Inverse Problems
Low Complexity Regularization of Inverse ProblemsGabriel Peyré
 
Multiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsMultiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsChristian Robert
 
Time-Series Analysis on Multiperiodic Conditional Correlation by Sparse Covar...
Time-Series Analysis on Multiperiodic Conditional Correlation by Sparse Covar...Time-Series Analysis on Multiperiodic Conditional Correlation by Sparse Covar...
Time-Series Analysis on Multiperiodic Conditional Correlation by Sparse Covar...Michael Lie
 
Divergence clustering
Divergence clusteringDivergence clustering
Divergence clusteringFrank Nielsen
 
Lecture: Monte Carlo Methods
Lecture: Monte Carlo MethodsLecture: Monte Carlo Methods
Lecture: Monte Carlo MethodsFrank Kienle
 

Similar to columbus15_cattaneo.pdf (20)

Student t t est
Student t t estStudent t t est
Student t t est
 
Introduction to Supervised ML Concepts and Algorithms
Introduction to Supervised ML Concepts and AlgorithmsIntroduction to Supervised ML Concepts and Algorithms
Introduction to Supervised ML Concepts and Algorithms
 
Nonparametric Density Estimation
Nonparametric Density EstimationNonparametric Density Estimation
Nonparametric Density Estimation
 
Randomized algorithms ver 1.0
Randomized algorithms ver 1.0Randomized algorithms ver 1.0
Randomized algorithms ver 1.0
 
Unbiased Bayes for Big Data
Unbiased Bayes for Big DataUnbiased Bayes for Big Data
Unbiased Bayes for Big Data
 
Computational Information Geometry on Matrix Manifolds (ICTP 2013)
Computational Information Geometry on Matrix Manifolds (ICTP 2013)Computational Information Geometry on Matrix Manifolds (ICTP 2013)
Computational Information Geometry on Matrix Manifolds (ICTP 2013)
 
Inference for stochastic differential equations via approximate Bayesian comp...
Inference for stochastic differential equations via approximate Bayesian comp...Inference for stochastic differential equations via approximate Bayesian comp...
Inference for stochastic differential equations via approximate Bayesian comp...
 
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...
 
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
QMC Program: Trends and Advances in Monte Carlo Sampling Algorithms Workshop,...
 
Thesis seminar
Thesis seminarThesis seminar
Thesis seminar
 
nber_slides.pdf
nber_slides.pdfnber_slides.pdf
nber_slides.pdf
 
Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...Pattern learning and recognition on statistical manifolds: An information-geo...
Pattern learning and recognition on statistical manifolds: An information-geo...
 
Intro to Quant Trading Strategies (Lecture 7 of 10)
Intro to Quant Trading Strategies (Lecture 7 of 10)Intro to Quant Trading Strategies (Lecture 7 of 10)
Intro to Quant Trading Strategies (Lecture 7 of 10)
 
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...
Hierarchical Deterministic Quadrature Methods for Option Pricing under the Ro...
 
Low Complexity Regularization of Inverse Problems
Low Complexity Regularization of Inverse ProblemsLow Complexity Regularization of Inverse Problems
Low Complexity Regularization of Inverse Problems
 
Multiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximationsMultiple estimators for Monte Carlo approximations
Multiple estimators for Monte Carlo approximations
 
Time-Series Analysis on Multiperiodic Conditional Correlation by Sparse Covar...
Time-Series Analysis on Multiperiodic Conditional Correlation by Sparse Covar...Time-Series Analysis on Multiperiodic Conditional Correlation by Sparse Covar...
Time-Series Analysis on Multiperiodic Conditional Correlation by Sparse Covar...
 
Divergence clustering
Divergence clusteringDivergence clustering
Divergence clustering
 
Pres metabief2020jmm
Pres metabief2020jmmPres metabief2020jmm
Pres metabief2020jmm
 
Lecture: Monte Carlo Methods
Lecture: Monte Carlo MethodsLecture: Monte Carlo Methods
Lecture: Monte Carlo Methods
 

Recently uploaded

Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)eniolaolutunde
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfsanyamsingh5019
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Krashi Coaching
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdfQucHHunhnh
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxiammrhaywood
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeThiyagu K
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room servicediscovermytutordmt
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptxVS Mahajan Coaching Centre
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactdawncurless
 
Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3JemimahLaneBuaron
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...Sapna Thakur
 
Russian Call Girls in Andheri Airport Mumbai WhatsApp 9167673311 💞 Full Nigh...
Russian Call Girls in Andheri Airport Mumbai WhatsApp  9167673311 💞 Full Nigh...Russian Call Girls in Andheri Airport Mumbai WhatsApp  9167673311 💞 Full Nigh...
Russian Call Girls in Andheri Airport Mumbai WhatsApp 9167673311 💞 Full Nigh...Pooja Nehwal
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactPECB
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxheathfieldcps1
 
Separation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesSeparation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesFatimaKhan178732
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docxPoojaSen20
 

Recently uploaded (20)

Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)Software Engineering Methodologies (overview)
Software Engineering Methodologies (overview)
 
Sanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdfSanyam Choudhary Chemistry practical.pdf
Sanyam Choudhary Chemistry practical.pdf
 
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
Kisan Call Centre - To harness potential of ICT in Agriculture by answer farm...
 
1029-Danh muc Sach Giao Khoa khoi 6.pdf
1029-Danh muc Sach Giao Khoa khoi  6.pdf1029-Danh muc Sach Giao Khoa khoi  6.pdf
1029-Danh muc Sach Giao Khoa khoi 6.pdf
 
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptxSOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
SOCIAL AND HISTORICAL CONTEXT - LFTVD.pptx
 
Measures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and ModeMeasures of Central Tendency: Mean, Median and Mode
Measures of Central Tendency: Mean, Median and Mode
 
9548086042 for call girls in Indira Nagar with room service
9548086042  for call girls in Indira Nagar  with room service9548086042  for call girls in Indira Nagar  with room service
9548086042 for call girls in Indira Nagar with room service
 
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions  for the students and aspirants of Chemistry12th.pptxOrganic Name Reactions  for the students and aspirants of Chemistry12th.pptx
Organic Name Reactions for the students and aspirants of Chemistry12th.pptx
 
Accessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impactAccessible design: Minimum effort, maximum impact
Accessible design: Minimum effort, maximum impact
 
Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3Q4-W6-Restating Informational Text Grade 3
Q4-W6-Restating Informational Text Grade 3
 
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
Mattingly "AI & Prompt Design: Structured Data, Assistants, & RAG"
 
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptxINDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
INDIA QUIZ 2024 RLAC DELHI UNIVERSITY.pptx
 
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
BAG TECHNIQUE Bag technique-a tool making use of public health bag through wh...
 
Russian Call Girls in Andheri Airport Mumbai WhatsApp 9167673311 💞 Full Nigh...
Russian Call Girls in Andheri Airport Mumbai WhatsApp  9167673311 💞 Full Nigh...Russian Call Girls in Andheri Airport Mumbai WhatsApp  9167673311 💞 Full Nigh...
Russian Call Girls in Andheri Airport Mumbai WhatsApp 9167673311 💞 Full Nigh...
 
Beyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global ImpactBeyond the EU: DORA and NIS 2 Directive's Global Impact
Beyond the EU: DORA and NIS 2 Directive's Global Impact
 
The basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptxThe basics of sentences session 2pptx copy.pptx
The basics of sentences session 2pptx copy.pptx
 
Separation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and ActinidesSeparation of Lanthanides/ Lanthanides and Actinides
Separation of Lanthanides/ Lanthanides and Actinides
 
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"Mattingly "AI & Prompt Design: The Basics of Prompt Design"
Mattingly "AI & Prompt Design: The Basics of Prompt Design"
 
Advance Mobile Application Development class 07
Advance Mobile Application Development class 07Advance Mobile Application Development class 07
Advance Mobile Application Development class 07
 
mini mental status format.docx
mini    mental       status     format.docxmini    mental       status     format.docx
mini mental status format.docx
 

columbus15_cattaneo.pdf

  • 1. Regression Discontinuity Designs in Stata Matias D. Cattaneo University of Michigan July 30, 2015
  • 2. Overview Main goal: learn about treatment e¤ect of policy or intervention. If treatment randomization available, easy to estimate treatment e¤ects. If treatment randomization not available, turn to observational studies. I Instrumental variables. I Selection on observables. Regression discontinuity (RD) designs. I Simple and objective. Requires little information, if design available. I Might be viewed as a “local” randomized trial. I Easy to falsify, easy to interpret. I Careful: very local!
  • 3. Overview of RD packages https://sites.google.com/site/rdpackages rdrobust package: estimation, inference and graphical presentation using local polynomials, partitioning, and spacings estimators. I rdrobust: RD inference (point estimation and CI; classic, bias-corrected, robust). I rdbwselect: bandwidth or window selection (IK, CV, CCT). I rdplot: plots data (with “optimal” block length). rddensity package: discontinuity in density test at cuto¤ (a.k.a. manipulation testing) using novel local polynomial density estimator. I rddensity: manipulation testing using local polynomial density estimation. I rdbwdensity: bandwidth or window selection. rdlocrand package: covariate balance, binomial tests, randomization inference methods (window selection & inference). I rdrandinf: inference using randomization inference methods. I rdwinselect: falsi…cation testing and window selection. I rdsensitivity: treatment e¤ect models over grid of windows, CI inversion. I rdrbounds: Rosenbaum bounds.
  • 4. Randomized Control Trials Notation: (Yi(0); Yi(1); Xi), i = 1; 2; : : : ; n. Treatment: Ti 2 f0; 1g, Ti independent of (Yi(0); Yi(1); Xi). Data: (Yi; Ti; Xi), i = 1; 2; : : : ; n, with Yi = ( Yi(0) if Ti = 0 Yi(1) if Ti = 1 Average Treatment E¤ect: ATE = E[Yi(1) Yi(0)] = E[YijT = 1] E[YijT = 0] Experimental Design.
  • 5. Sharp RD design Notation: (Yi(0); Yi(1); Xi), i = 1; 2; : : : ; n, Xi continuous Treatment: Ti 2 f0; 1g, Ti = 1(Xi x). Data: (Yi; Ti; Xi), i = 1; 2; : : : ; n, with Yi = ( Yi(0) if Ti = 0 Yi(1) if Ti = 1 Average Treatment E¤ect at the cuto¤: SRD = E[Yi(1) Yi(0)jXi = x] = lim x#x E[YijXi = x] lim x"x E[YijXi = x] Quasi-Experimental Design: “local randomization” (more later)
  • 6. −0.6 −0.4 −0.2 0.0 0.2 0.4 0.6 −2 −1 0 1 2 3 Assignment variable (R) Outcome variable (Y) τ0
  • 7. −0.6 −0.4 −0.2 0.0 0.2 0.4 0.6 −2 −1 0 1 2 3 Assignment variable (R) Outcome variable (Y) τ0
  • 8. −0.6 −0.4 −0.2 0.0 0.2 0.4 0.6 −2 −1 0 1 2 3 Assignment variable (R) Outcome variable (Y) τ0
  • 9. −0.6 −0.4 −0.2 0.0 0.2 0.4 0.6 −6 −4 −2 0 2 4 6 Assignment variable (R) Outcome variable (Y) Local Random Assignment τ0
  • 10. Empirical Illustration: Cattaneo, Frandsen & Titiunik (2015, JCI) Problem: incumbency advantage (U.S. senate). Data: Yi = election outcome. Ti = whether incumbent. Xi = vote share previous election (x = 0). Zi = covariates (demvoteshlag1, demvoteshlag2, dopen, etc.). Potential outcomes: Yi(0) = election outcome if had not been incumbent. Yi(1) = election outcome if had been incumbent. Causal Inference: Yi(0) 6= YijTi = 0 and Yi(1) 6= YijTi = 1
  • 11. Graphical and Falsi…cation Methods Always plot data: main advantage of RD designs! Plot regression functions to assess treatment e¤ect and validity. Plot density of Xi for assessing validity; test for continuity at cuto¤ and elsewhere. Important: use also estimators that do not “smooth-out” data. RD Plots (Calonico, Cattaneo & Titiunik, JASA): I Two ingredients: (i) Smoothed global polynomial …t & (ii) binned discontinuous local-means …t. I Two goals: (i) detention of discontinuities, & (ii) representation of variability. I Two tuning parameters: F G lobal p olynom ial degree (kn). F Location (ES or Q S) and numb er of bins (Jn).
  • 12. Manipulation Tests & Covariate Balance and Placebo Tests Density tests near cuto¤: I Idea: distribution of running variable should be similar at either side of cuto¤. I Method 1: Histograms & Binomial count test. I Method 2: Density Estimator at boundary. F Pre-binned local p olynom ial m ethod – M cCrary (2008). F New tuning-param eter-free m ethod – C attaneo, Jansson and M a (2015). Placebo tests on pre-determined/exogenous covariates. I Idea: zero RD treatment e¤ect for pre-determined/exogenous covariates. I Methods: global polynomial, local polynomial, randomization-based. Placebo tests on outcomes. I Idea: zero RD treatment e¤ect for outcome at values other than cuto¤. I Methods: global polynomial, local polynomial, randomization-based.
  • 13. Estimation and Inference Methods Global polynomial approach (not recommended). Robust local polynomial inference methods. I Bandwidth selection. I Bias-correction. I Con…dence intervals. Local randomization and randomization inference methods. I Window selection. I Estimation and Inference methods. I Falsi…cation, sensitivity and related methods
  • 14. Conventional Local-polynomial Approach Idea: approximate regression functions for control and treatment units locally. “Local-linear” estimator (w/ weights K( )): hn Xi < x : x Xi hn : Yi = + (Xi x) + " ;i Yi = + + (Xi x) + + "+;i I Treatment e¤ect (at the cuto¤): ^SRD = ^+ ^ Can be estimated using linear models (w/ weights K( )): Yi = + SRD Ti + (Xi x) 1 + Ti (Xi x) 1 + "i, hn Xi hn Once hn chosen, inference is “standard”: weighted linear models. I Details coming up next.
  • 15. Conventional Local-polynomial Approach How to choose hn? Imbens & Kalyanaraman (2012, ReStud): “optimal” plug-in, ^ hIK = ^ CIK n 1=5 Calonico, Cattaneo & Titiunik (2014, ECMA): re…nement of IK ^ hCCT = ^ CCCT n 1=5 Ludwig & Miller (2007, QJE): cross-validation, ^ hCV = arg min h n X i=1 w(Xi) (Yi ^1(Xi; h))2 Key idea: trade-o¤ bias and variance of ^SRD(hn). Heuristically: " Bias(^SRD) =) # ^ h and " Var(^SRD) =) " ^ h
  • 16. Local-Polynomial Methods: Bandwidth Selection Two main methods: plug-in & cross-validation. Both MSE-optimal in some sense. Imbens & Kalyanaraman (2012, ReStud): propose MSE-optimal rule, hMSE = C 1=5 MSE n 1=5 CMSE = C(K) Var(^SRD) Bias(^SRD)2 I IK implementation: …rst-generation plug-in rule. I CCT implementation: second-generation plug-in rule. I They di¤er in the way Var(^SRD) and Bias(^SRD) are estimated. Imbens & Kalyanaraman (2012, ReStud): discuss cross-validation approach, ^ hCV = arg min h>0 CV (h) , CV (h) = n X i=1 1(X ;[ ] Xi X+;[ ]) (Yi ^(Xi; h))2 , where I ^+;p(x; h) and ^ ;p(x; h) are local polynomials estimates. I 2 (0; 1), X ;[ ] and X+;[ ] denote -th quantile of fXi : Xi < xg and fXi : Xi xg. I Our implementation uses = 0:5; but this is a tuning parameter!
  • 17. Conventional Approach to RD “Local-linear” estimator (w/ weights K( )): hn Xi < x : x Xi hn : Yi = + (Xi x) + " ;i Yi = + + (Xi x) + + "+;i I Treatment e¤ect (at the cuto¤): ^SRD = ^+ ^ Construct usual t-test. For H0 : SRD = 0, ^ T(hn) = ^SRD p ^ Vn = ^+ ^ q ^ V+;n + ^ V ;n d N(0; 1) 95% Con…dence interval: ^ I(hn) = ^SRD 1:96 q ^ Vn
  • 18. Bias-Correction Approach to RD Note well: for usual t-test, ^ T(hMSE) = ^SRD p ^ Vn d N(B; 1) 6= N(0; 1), B > 0 I Bias B in RD estimator captures “curvature” of regression functions. Undersmoothing/“Small Bias”Approach: Choose “smaller”hn... Perhaps ^ hn = 0:5 ^ hIK? =) Not clear guidance & power loss! Bias-correction Approach: ^ Tbc (hn; bn) = ^SRD ^ Bn p ^ Vn d N(0; 1) =) 95% Con…dence Interval: ^ Ibc(hn; bn) = h ^SRD ^ Bn 1:96 p ^ Vn i How to choose bn? Same ideas as before... ^ bn = ^ C n 1=7
  • 19. Robust Bias-Correction Approach to RD Recall: ^ T (hn) = ^SRD p ^ Vn d N (0; 1) and ^ T bc (hn; bn) = ^SRD ^ Bn p ^ Vn d N (0; 1) I ^ Bn is constructed to estimate leading bias B. Robust approach: ^ Tbc (hn; bn) = ^SRD ^ Bn p ^ Vn = ^SRD Bn p ^ Vn | {z } d N (0;1) + Bn ^ Bn p ^ Vn | {z } d N (0; ) Robust bias-corrected t-test: ^ Trbc (hn; bn) = ^SRD ^ Bn p ^ Vn + ^ Wn = ^SRD ^ Bn q ^ Vbc n d N(0; 1) =) 95% Con…dence Interval: ^ Irbc (hn; bn) = ^SRD ^ Bn 1:96 q ^ Vbc n , ^ Vbc n = ^ Vn + ^ Wn
  • 20. Local-Polynomial Methods: Robust Inference Approach 1: Undersmoothing/“Small Bias”. ^ I(hn) = ^SRD 1:96 q ^ Vn Approach 2: Bias correction (not recommended). ^ Ibc (hn; bn) = ^SRD ^ Bn 1:96 q ^ Vn Approach 3: Robust Bias correction. ^ Irbc (hn; bn) = ^SRD ^ Bn 1:96 q ^ Vn + ^ Wn
  • 21. Local-randomization approach and …nite-sample inference Popular approach: local-polynomial methods. I Approximates regression function and relies on continuity assumptions. I Requires: choosing weights, bandwidth and polynomial order. Alternative approach: local-randomization + randomization-inference I Gives an alternative that can be used as a robustness check. I Key assumption: exists window W = [ hn; hn] around cuto¤ ( hn < x < hn) where Ti independent of (Yi(0); Yi(1)) (for all Xi 2 W ) I In words: treatment is randomly assigned within W . I Good news: if plausible, then RCT ideas/methods apply. I Not-so-good news: most plausible for very small windows (very few observations). I One solution: employ small window but use randomization-inference methods. I Requires: choosing randomization rule, window and statistic.
  • 22. Local-randomization approach and …nite-sample inference Recall key assumption: exists W = [ hn; hn] around cuto¤ ( hn < x < hn) where Ti independent of (Yi(0); Yi(1)) (for all Xi 2 W) How to choose window? I Use balance tests on pre-determined/exogenous covariates. I Very intuitive, easy to implement. How to conduct inference? Use randomization-inference methods. 1 Choose statistic of interest. E.g., t-stat for di¤erence-in-means. 2 Choose randomization rule. E.g., number of treatments and controls given. 3 Compute …nite-sample distribution of statistics by permuting treatment assignments.
  • 23. Local-randomization approach and …nite-sample inference Do not forget to validate & falsify the empirical strategy. 1 Plot data to make sure local-randomization is plausible. 2 Conduct placebo tests. (e.g., use pre-intervention outcomes or other covariates not used select W ) 3 Do sensitivity analysis. See Cattaneo, Frandsen and Titiunik (2015) for introduction. See Cattaneo, Titiunik and Vazquez-Bare (2015) for further results and implementation.