why propensity scores should not be used for …why propensity scores should not be used for...
TRANSCRIPT
-
Why Propensity ScoresShould Not Be Used For Matching
Gary King1 Richard Nielsen2
Institute for Quantitative Social Science MITHarvard University
MacMillan-CSAP Workshop on Quantitative Research Methods,Yale University, 3/10/2016
1GaryKing.org2www.mit.edu/∼rnielsen
1/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores:
E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method
• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores:
E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)
• Maybe even “the most developed and popular strategy forcausal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores:
E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores:
E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores:
E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching
• as used in practiceNot implicated by our results:
• Other uses of propensity scores:
E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores:
E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores:
E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores:
E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores: E.g., regression adjustment,
inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores: E.g., regression adjustment,inverse weighting,
stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores: E.g., regression adjustment,inverse weighting, stratification,
pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores: E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores: E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores:
Correct,but inadequate
2/23
-
The Scholarly Influence of Propensity Score Matching
• The most commonly used matching method• In 53,200 articles! (according to Google Scholar)• Maybe even “the most developed and popular strategy for
causal analysis in observational studies” (Pearl, 2010)
This paper is about:
• propensity score matching• as used in practice
Not implicated by our results:
• Other uses of propensity scores: E.g., regression adjustment,inverse weighting, stratification, pscores used in other methods
• The mathematical theorems about propensity scores: Correct,but inadequate
2/23
-
Matching to Reduce Model Dependence
3/23
-
Matching to Reduce Model Dependence(Ho, Imai, King, Stuart, 2007: fig.1, Political Analysis)
3/23
-
Matching to Reduce Model Dependence(Ho, Imai, King, Stuart, 2007: fig.1, Political Analysis)
Education (years)
Out
com
e
12 14 16 18 20 22 24 26 28
0
2
4
6
8
10
12
3/23
-
Matching to Reduce Model Dependence(Ho, Imai, King, Stuart, 2007: fig.1, Political Analysis)
Education (years)
Out
com
e
12 14 16 18 20 22 24 26 28
0
2
4
6
8
10
12
T
T
T
T T
T
T
TTT
TT
T TT T
T
T
T
T
3/23
-
Matching to Reduce Model Dependence(Ho, Imai, King, Stuart, 2007: fig.1, Political Analysis)
Education (years)
Out
com
e
12 14 16 18 20 22 24 26 28
0
2
4
6
8
10
12
T
T
T
T T
T
T
TTT
TT
T TT T
T
T
T
T
CC
C
CC
C
C
C
C
C
C
C
C
C
C
C
C
CC C
C
C
C
C
C
C
C
C
C
C
C
CCC
CC
CC
C
C
3/23
-
Matching to Reduce Model Dependence(Ho, Imai, King, Stuart, 2007: fig.1, Political Analysis)
Education (years)
Out
com
e
12 14 16 18 20 22 24 26 28
0
2
4
6
8
10
12
T
T
T
T T
T
T
TTT
TT
T TT T
T
T
T
T
CC
C
CC
C
C
C
C
C
C
C
C
C
C
C
C
CC C
C
C
C
C
C
C
C
C
C
C
C
CCC
CC
CC
C
C
3/23
-
Matching to Reduce Model Dependence(Ho, Imai, King, Stuart, 2007: fig.1, Political Analysis)
Education (years)
Out
com
e
12 14 16 18 20 22 24 26 28
0
2
4
6
8
10
12
T
T
T
T T
T
T
TTT
TT
T TT T
T
T
T
T
CC
C
CC
C
C
C
C
C
C
C
C
C
C
C
C
CC C
C
C
C
C
C
C
C
C
C
C
C
CCC
CC
CC
C
C
3/23
-
Matching to Reduce Model Dependence(Ho, Imai, King, Stuart, 2007: fig.1, Political Analysis)
Education (years)
Out
com
e
12 14 16 18 20 22 24 26 28
0
2
4
6
8
10
12
T
T
T
T T
T
T
TTT
TT
T TT T
T
T
T
T
CC
C
CC
C
C
C
C
C
C
C
C
C
C
C
C
CC C
C
C
C
C
C
C
C
C
C
C
C
CCC
CC
CC
C
C
3/23
-
Matching to Reduce Model Dependence(Ho, Imai, King, Stuart, 2007: fig.1, Political Analysis)
Education (years)
Out
com
e
12 14 16 18 20 22 24 26 28
0
2
4
6
8
10
12
T
T
T
T T
T
T
TTT
TT
T TT T
T
T
T
TC
C
C
C
C
CC
C
C
CC
C CC
C
C
CCCC
C
CC
C
CC
CC
CC
C
C
C
C
CC
CCCC
3/23
-
Matching to Reduce Model Dependence(Ho, Imai, King, Stuart, 2007: fig.1, Political Analysis)
Education (years)
Out
com
e
12 14 16 18 20 22 24 26 28
0
2
4
6
8
10
12
T
T
T
T T
T
T
TTT
TT
T TT T
T
T
T
TC
C
C
C
C
CC
C
C
CC
C CC
C
C
CCCC
C
CC
C
CC
CC
CC
C
C
C
C
CC
CCCC
3/23
-
The Problems Matching Solves
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance Model Dependence
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance Model Dependence Researcher discretion
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance Model Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance Model Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance Model Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator• e.g., Choosing from results of 50 randomized experiments
• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance Model Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance Model Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]
• People do not have easy access to their own mental processesor feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance Model Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance Model Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
Without Matching:
Imbalance Model Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
With��HHout Matching:
Imbalance Model Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
With��HHout Matching:
��ZZImbalance Model Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
With��HHout Matching:
��ZZImbalance ((((((((
(hhhhhhhhhModel Dependence Researcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
With��HHout Matching:
��ZZImbalance ((((((((
(hhhhhhhhhModel Dependence ((((((((
((hhhhhhhhhhResearcher discretion Bias
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
With��HHout Matching:
��ZZImbalance ((((((((
(hhhhhhhhhModel Dependence ((((((((
((hhhhhhhhhhResearcher discretion ���XXXBias
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
The Problems Matching Solves
With��HHout Matching:
��ZZImbalance ((((((((
(hhhhhhhhhModel Dependence ((((((((
((hhhhhhhhhhResearcher discretion ���XXXBias
A central project of statistics: Automating away human discretion
• Qualitative choice from unbiased estimates = biased estimator
• e.g., Choosing from results of 50 randomized experiments• Choosing based on “plausibility” is probably worse[eff]
• conscientious effort doesn’t avoid biases (Banaji 2013)[acc]• People do not have easy access to their own mental processes
or feedback to avoid the problem (Wilson and Brekke1994)[exprt]
• Experts overestimate their ability to control personal biasesmore than nonexperts, and more prominent experts are themost overconfident (Tetlock 2005)[tch]
• “Teaching psychology is mostly a waste of time” (Kahneman2011)
4/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi − Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders
• Treatment Effect for treated observation i :
TEi = Yi − Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi − Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi (1)− Yi (0)
= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi (1)− Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi − Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi − Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control
• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi − Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi − Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi − Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi − Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
What’s Matching?
• Yi dep var, Ti (1=treated, 0=control), Xi confounders• Treatment Effect for treated observation i :
TEi = Yi − Yi (0)= observed− unobserved
• Estimate Yi (0) with Yj with a matched (Xi ≈ Xj) control• Quantities of Interest:
1. SATT: Sample Average Treatment effect on the Treated:
SATT = Meani∈{Ti=1}
(TEi )
2. FSATT: Feasible SATT (prune badly matched treateds too)
• Big convenience: Follow preprocessing with whateverstatistical method you’d have used without matching
• Pruning nonmatches makes control vars matter less: reducesimbalance, model dependence, researcher discretion, & bias
5/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed
On average Exact
Unobserved
On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average
Exact
Unobserved
On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average
Exact
Unobserved On average
On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average
On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization
for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:
imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance,
model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence,
power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power,
efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency,
bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias,
researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts,
robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness.
E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization
• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked
• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM
(wait, it gets worse)
6/23
-
Matching: Finding Hidden Randomized Experiments
Types of Experiments
BalanceCovariates:
CompleteRandomization
FullyBlocked
Observed On average ExactUnobserved On average On average
Fully blocked dominates complete randomization for:imbalance, model dependence, power, efficiency, bias, researchcosts, robustness. E.g., Imai, King, Nall 2009: SEs 600% smaller!
Goal of Each Matching Method (in Observational Data)
• PSM: complete randomization• Other methods: fully blocked• Other matching methods dominate PSM (wait, it gets worse)
6/23
-
Method 1: Mahalanobis Distance Matching
1. Preprocess (Matching)
• Distance(Xc ,Xt) =√
(Xc − Xt)′S−1(Xc − Xt)• (Mahalanobis is for methodologists; in applications, use
Euclidean!)• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
7/23
-
Method 1: Mahalanobis Distance Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)
• Distance(Xc ,Xt) =√
(Xc − Xt)′S−1(Xc − Xt)• (Mahalanobis is for methodologists; in applications, use
Euclidean!)• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
7/23
-
Method 1: Mahalanobis Distance Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)
• Distance(Xc ,Xt) =√
(Xc − Xt)′S−1(Xc − Xt)• (Mahalanobis is for methodologists; in applications, use
Euclidean!)• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
7/23
-
Method 1: Mahalanobis Distance Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)
• Distance(Xc ,Xt) =√
(Xc − Xt)′S−1(Xc − Xt)
• (Mahalanobis is for methodologists; in applications, useEuclidean!)
• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
7/23
-
Method 1: Mahalanobis Distance Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)
• Distance(Xc ,Xt) =√
(Xc − Xt)′S−1(Xc − Xt)• (Mahalanobis is for methodologists; in applications, use
Euclidean!)
• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
7/23
-
Method 1: Mahalanobis Distance Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)
• Distance(Xc ,Xt) =√
(Xc − Xt)′S−1(Xc − Xt)• (Mahalanobis is for methodologists; in applications, use
Euclidean!)• Match each treated unit to the nearest control unit
• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
7/23
-
Method 1: Mahalanobis Distance Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)
• Distance(Xc ,Xt) =√
(Xc − Xt)′S−1(Xc − Xt)• (Mahalanobis is for methodologists; in applications, use
Euclidean!)• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused
• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
7/23
-
Method 1: Mahalanobis Distance Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)
• Distance(Xc ,Xt) =√
(Xc − Xt)′S−1(Xc − Xt)• (Mahalanobis is for methodologists; in applications, use
Euclidean!)• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper
• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
7/23
-
Method 1: Mahalanobis Distance Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)
• Distance(Xc ,Xt) =√
(Xc − Xt)′S−1(Xc − Xt)• (Mahalanobis is for methodologists; in applications, use
Euclidean!)• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
7/23
-
Mahalanobis Distance Matching
Education (years)
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
8/23
-
Mahalanobis Distance Matching
Education (years)
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
TTTT
T
T
T
T
T
T
T
T
T
TT
T
T
T
T
T
8/23
-
Mahalanobis Distance Matching
Education (years)
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
C
C
CC
C
C
C
C
C
CC
C
CCC
CC
C
C
C
CC CC
C
C
CC
C
CC
CC
C
C C
CC
C
C
TTTT
T
T
T
T
T
T
T
T
T
TT
T
T
T
T
T
8/23
-
Mahalanobis Distance Matching
Education (years)
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
C
C
CC
C
C
C
C
C
CC
C
CCC
CC
C
C
C
CC CC
C
C
CC
C
CC
CC
C
C C
CC
C
C
TTTT
T
T
T
T
T
T
T
T
T
TT
T
T
T
T
T
8/23
-
Mahalanobis Distance Matching
Education (years)
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
T TT T
TTTT
T TTTT
T TT
TTTT
CCC C
CC
C
C
C CC
C
CC
CCC CC
C
C
CCCCC
CCC CCCCC
C CCCC
C
8/23
-
Mahalanobis Distance Matching
Education (years)
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
T TT T
TTTT
T TTTT
T TT
TTTT
CCC C
CC
C
C
C CC
C
CC
CCC CC
C
8/23
-
Mahalanobis Distance Matching
Education (years)
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
T TT T
TTTT
T TTTT
T TT
TTTT
CCC C
CC
C
C
C CC
C
CC
CCC CC
C
8/23
-
Best Case: Mahalanobis Distance Matching
9/23
-
Best Case: Mahalanobis Distance Matching
Education (years)
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
T TTTT T TTTT TT T
T TTTTTTT T
TTTTT T T
T TTTT T
TT TTTTT
TTTT
TTTT
C CCCC C CCCC CC C
C CCCCCCC C
CCCCC C C
C CCCC C
CC CCCCC
CCCC
CCCC
CC CCC CCCCC CCCCCC CCC CC C C CCCCC CCC CC CC CC CC CC CC CC CC C CCC
CC CC CC C CCCC C CCC CC CC C CCC CC CC CC C CCC CC CC CCC C CC CCCCC
CC CCCC C CCC CCC CC C C CCCCCC CCCC C CCCC C CC CCC CC CCC C
C CC CC CC C CCCC C CC C C CC CCC CC CCC CCCCC CCCC CCC CC CCC C C CC CC C
C CCCC CC CC CCC C C CCC C CC CCC CC C CCC C C CCC CC CC C CC C CC
CCCCC CCCC C C CC C CCCC CC CCC CCC C CCC CC CC CC CC CC C CC C C
C CC CCC CC C C CCC CCC C CC CC CCC CCC CC CCC CC C CCC CC C C
C CCCC C CC C CC CC C CC C CCC C C CCC CC C CCCCCC CC CCCC CC CC C CC C CC CC C CC CC
C CC C CCC CCCC C CC CC C CCC CC CC CC C CCCCC CCC C C CC C CC CCC
CCC CC CCC CC CCCC C CCC CC CCC
9/23
-
Best Case: Mahalanobis Distance Matching
Education (years)
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
T TTTT T TTTT TT T
T TTTTTTT T
TTTTT T T
T TTTT T
TT TTTTT
TTTT
TTTT
C CCCC C CCCC CC C
C CCCCCCC C
CCCCC C C
C CCCC C
CC CCCCC
CCCC
CCCC
9/23
-
Method 2: Coarsened Exact Matching
1. Preprocess (Matching)
• Temporarily coarsen X as much as you’re willing
• e.g., Education (grade school, high school, college, graduate)
• Apply exact matching to the coarsened X , C (X )
• Sort observations into strata, each with unique values of C(X )• Prune any stratum with 0 treated or 0 control units
• Pass on original (uncoarsened) units except those pruned
2. Estimation Difference in means or a model
• Weight controls in each stratum to equal treateds
10/23
-
Method 2: Coarsened Exact Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)
• Temporarily coarsen X as much as you’re willing
• e.g., Education (grade school, high school, college, graduate)
• Apply exact matching to the coarsened X , C (X )
• Sort observations into strata, each with unique values of C(X )• Prune any stratum with 0 treated or 0 control units
• Pass on original (uncoarsened) units except those pruned
2. Estimation Difference in means or a model
• Weight controls in each stratum to equal treateds
10/23
-
Method 2: Coarsened Exact Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)
• Temporarily coarsen X as much as you’re willing
• e.g., Education (grade school, high school, college, graduate)
• Apply exact matching to the coarsened X , C (X )
• Sort observations into strata, each with unique values of C(X )• Prune any stratum with 0 treated or 0 control units
• Pass on original (uncoarsened) units except those pruned
2. Estimation Difference in means or a model
• Weight controls in each stratum to equal treateds
10/23
-
Method 2: Coarsened Exact Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)• Temporarily coarsen X as much as you’re willing
• e.g., Education (grade school, high school, college, graduate)• Apply exact matching to the coarsened X , C (X )
• Sort observations into strata, each with unique values of C(X )• Prune any stratum with 0 treated or 0 control units
• Pass on original (uncoarsened) units except those pruned
2. Estimation Difference in means or a model
• Weight controls in each stratum to equal treateds
10/23
-
Method 2: Coarsened Exact Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)• Temporarily coarsen X as much as you’re willing
• e.g., Education (grade school, high school, college, graduate)
• Apply exact matching to the coarsened X , C (X )
• Sort observations into strata, each with unique values of C(X )• Prune any stratum with 0 treated or 0 control units
• Pass on original (uncoarsened) units except those pruned
2. Estimation Difference in means or a model
• Weight controls in each stratum to equal treateds
10/23
-
Method 2: Coarsened Exact Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)• Temporarily coarsen X as much as you’re willing
• e.g., Education (grade school, high school, college, graduate)• Apply exact matching to the coarsened X , C (X )
• Sort observations into strata, each with unique values of C(X )• Prune any stratum with 0 treated or 0 control units
• Pass on original (uncoarsened) units except those pruned
2. Estimation Difference in means or a model
• Weight controls in each stratum to equal treateds
10/23
-
Method 2: Coarsened Exact Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)• Temporarily coarsen X as much as you’re willing
• e.g., Education (grade school, high school, college, graduate)• Apply exact matching to the coarsened X , C (X )
• Sort observations into strata, each with unique values of C(X )
• Prune any stratum with 0 treated or 0 control units• Pass on original (uncoarsened) units except those pruned
2. Estimation Difference in means or a model
• Weight controls in each stratum to equal treateds
10/23
-
Method 2: Coarsened Exact Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)• Temporarily coarsen X as much as you’re willing
• e.g., Education (grade school, high school, college, graduate)• Apply exact matching to the coarsened X , C (X )
• Sort observations into strata, each with unique values of C(X )• Prune any stratum with 0 treated or 0 control units
• Pass on original (uncoarsened) units except those pruned
2. Estimation Difference in means or a model
• Weight controls in each stratum to equal treateds
10/23
-
Method 2: Coarsened Exact Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)• Temporarily coarsen X as much as you’re willing
• e.g., Education (grade school, high school, college, graduate)• Apply exact matching to the coarsened X , C (X )
• Sort observations into strata, each with unique values of C(X )• Prune any stratum with 0 treated or 0 control units
• Pass on original (uncoarsened) units except those pruned
2. Estimation Difference in means or a model
• Weight controls in each stratum to equal treateds
10/23
-
Method 2: Coarsened Exact Matching(Approximates Fully Blocked Experiment)
1. Preprocess (Matching)• Temporarily coarsen X as much as you’re willing
• e.g., Education (grade school, high school, college, graduate)• Apply exact matching to the coarsened X , C (X )
• Sort observations into strata, each with unique values of C(X )• Prune any stratum with 0 treated or 0 control units
• Pass on original (uncoarsened) units except those pruned
2. Estimation Difference in means or a model• Weight controls in each stratum to equal treateds
10/23
-
Coarsened Exact Matching
11/23
-
Coarsened Exact Matching
Education
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
CCC C
CC CC
C CC C CCC CCC
CCCC CC CC
C CCCCCC
C CCC CC
C
T TT T
TTTT
T TTTT
T TT
TTTT
11/23
-
Coarsened Exact Matching
Education
HS BA MA PhD 2nd PhD
Drinking age
Don't trust anyoneover 30
The Big 40
Senior Discounts
Retirement
Old
CCC C
CC CC
C CC C CCC CCC
CCCC CC CC
C CCCCCC
C CCC CC
C
T TT T
TTTT
T TTTT
T TT
TTTT
11/23
-
Coarsened Exact Matching
Education
HS BA MA PhD 2nd PhD
Drinking age
Don't trust anyoneover 30
The Big 40
Senior Discounts
Retirement
Old
CCC C
CC CC
C CC C CCC CCC
CCCC CC CC
C CCCCCC
C CCC CC
C
T TT T
TTTT
T TTTT
T TT
TTTT
11/23
-
Coarsened Exact Matching
Education
HS BA MA PhD 2nd PhD
Drinking age
Don't trust anyoneover 30
The Big 40
Senior Discounts
Retirement
Old
CC C
CC
CC CCCC CC C
CCCC
TTT T T
TTT
T TT
TTTT
11/23
-
Coarsened Exact Matching
Education
HS BA MA PhD 2nd PhD
Drinking age
Don't trust anyoneover 30
The Big 40
Senior Discounts
Retirement
Old
CC C
CCCC C CC
C CC CCC
C
C
TTT T T
TTT
T TT
TTTT
11/23
-
Coarsened Exact Matching
Education
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
CC C
CCCC C CC
C CC CCC
C
C
TTT T T
TTT
T TT
TTTT
11/23
-
Best Case: Coarsened Exact Matching
12/23
-
Best Case: Coarsened Exact Matching
Education
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
CC CCC CCCCC CCCCCC CCC CC C C CCCCC CCC CC CC CC C CCC CCC CCC CC CC C CCC
CCC CC CCC C C CCCC C CCC CC CC C CCCCC CC C CC CC CCC CCC CC CC CCC C C CC C
CC CCC CC CCCC C CC CC CCC CC CC C CCCC C CC CC CCC CC CCCCC C CC CCC CC
C C CCC CC CC CC CCC C CC CCC C CC C C CC CCC CC CCC C CCCCC CCCC CCC CC CC CCC C CC CC C
C CC CCCCC CC CC CCC C CC CCC CCC CCC CC C CCCC C C CCC CC CC CCC C C
C C CC CCCCC CCCC CC C CCC CCCC CCC CCC CCC C CCC CC CC CC CCC CC C
C CC C C CC CC CC CC CCC C C CCC CCC C CCCCC CCC CCC CC CCC CC C CCC C C
C C CC CCCCC C CC C CC CC C CC C CCC C C CCC CC C CCC CCC CC CCCC CC CC C CC C CC CC CC CC
CCC CC C CCC CCCC C CC CC C CCC C CC CCC CC C CCC CCCC CCCC C C CC C CC CC
C CCC CC CCC CC CCCC CCCC CC CCC
T TTTT T TTTT TT T
T TTTTTTT T
TTTTT T T
T TTTT T
TT TTTTT
TTTT
TTTT
12/23
-
Best Case: Coarsened Exact Matching
Education
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
C CCCC CCCCC C
C CCCCCC
CCCCC C C
CCCC C
C CCC C
C
C CCCC
T TTTT TTTTT T
T TTTTTT
TTTTT T T
TTTT T
T TTT T
T
T TTTT
12/23
-
Best Case: Coarsened Exact Matching
Education
Age
12 14 16 18 20 22 24 26 28
20
30
40
50
60
70
80
C CCCC CCCCC C
C CCCCCC
CCCCC C C
CCCC C
C CCC C
C
C CCCC
T TTTT TTTTT T
T TTTTTT
TTTTT T T
TTTT T
T TTT T
T
T TTTT
12/23
-
Method 3: Propensity Score Matching
1. Preprocess (Matching)
• Reduce k elements of X to scalarπi ≡ Pr(Ti = 1|X ) = 11+e−Xiβ
• Distance(Xc ,Xt) = |πc − πt |• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
13/23
-
Method 3: Propensity Score Matching(Approximates Completely Randomized Experiment)
1. Preprocess (Matching)
• Reduce k elements of X to scalarπi ≡ Pr(Ti = 1|X ) = 11+e−Xiβ
• Distance(Xc ,Xt) = |πc − πt |• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
13/23
-
Method 3: Propensity Score Matching(Approximates Completely Randomized Experiment)
1. Preprocess (Matching)
• Reduce k elements of X to scalarπi ≡ Pr(Ti = 1|X ) = 11+e−Xiβ
• Distance(Xc ,Xt) = |πc − πt |• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
13/23
-
Method 3: Propensity Score Matching(Approximates Completely Randomized Experiment)
1. Preprocess (Matching)• Reduce k elements of X to scalarπi ≡ Pr(Ti = 1|X ) = 11+e−Xiβ
• Distance(Xc ,Xt) = |πc − πt |• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
13/23
-
Method 3: Propensity Score Matching(Approximates Completely Randomized Experiment)
1. Preprocess (Matching)• Reduce k elements of X to scalarπi ≡ Pr(Ti = 1|X ) = 11+e−Xiβ
• Distance(Xc ,Xt) = |πc − πt |
• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
13/23
-
Method 3: Propensity Score Matching(Approximates Completely Randomized Experiment)
1. Preprocess (Matching)• Reduce k elements of X to scalarπi ≡ Pr(Ti = 1|X ) = 11+e−Xiβ
• Distance(Xc ,Xt) = |πc − πt |• Match each treated unit to the nearest control unit
• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
13/23
-
Method 3: Propensity Score Matching(Approximates Completely Randomized Experiment)
1. Preprocess (Matching)• Reduce k elements of X to scalarπi ≡ Pr(Ti = 1|X ) = 11+e−Xiβ
• Distance(Xc ,Xt) = |πc − πt |• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused
• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
13/23
-
Method 3: Propensity Score Matching(Approximates Completely Randomized Experiment)
1. Preprocess (Matching)• Reduce k elements of X to scalarπi ≡ Pr(Ti = 1|X ) = 11+e−Xiβ
• Distance(Xc ,Xt) = |πc − πt |• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper
• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
13/23
-
Method 3: Propensity Score Matching(Approximates Completely Randomized Experiment)
1. Preprocess (Matching)• Reduce k elements of X to scalarπi ≡ Pr(Ti = 1|X ) = 11+e−Xiβ
• Distance(Xc ,Xt) = |πc − πt |• Match each treated unit to the nearest control unit• Control units: not reused; pruned if unused• Prune matches if Distance>caliper• (Many adjustments available to this basic method)
2. Estimation Difference in means or a model
13/23
-
Propensity Score Matching
Education (years)
Age
12 16 20 24 28
20
30
40
50
60
70
80
C
C
CC
C
C
C
C
C
CC
C
CCC
CC
C
C
C
CCCC
C
C
CC
C
CC
CC
C
C C
CC
C
C
TTTT
T
T
T
T
T
T
T
T
T
TT
T
T
T
T
T
14/23
-
Propensity Score Matching
Education (years)
Age
12 16 20 24 28
20
30
40
50
60
70
80
C
C
CC
C
C
C
C
C
CC
C
CCC
CC
C
C
C
CCCC
C
C
CC
C
CC
CC
C
C C
CC
C
C
TTTT
T
T
T
T
T
T
T
T
T
TT
T
T
T
T
T
1
0
PropensityScore 14/23
-
Propensity Score Matching
Education (years)
Age
12 16 20 24 28
20
30
40
50
60
70
80
C
C
CC
C
C
C
C
C
CC
C
CCC
CC
C
C
C
CCCC
C
C
CC
C
CC
CC
C
C C
CC
C
C
TTTT
T
T
T
T
T
T
T
T
T
TT
T
T
T
T
T
1
0
PropensityScore
C
C
CC
CCC
C
C
C
CC
C
C
C
C
C
C
C
CCCCC
C
CCCCCCCCC
C
C
C
C
CC
T
TTT
T
TT
T
T
T
T
TT
T
T
T
T
T
T
T
14/23
-
Propensity Score Matching
Education (years)
Age
12 16 20 24 28
20
30
40
50
60
70
80
1
0
PropensityScore
C
C
CC
CCC
C
C
C
CC
C
C
C
C
C
C
C
CCCCC
C
CCCCCCCCC
C
C
C
C
CC
T
TTT
T
TT
T
T
T
T
TT
T
T
T
T
T
T
T
14/23
-
Propensity Score Matching
Education (years)
Age
12 16 20 24 28
20
30
40
50
60
70
80
1
0
PropensityScore
C
C
CC
CCC
C
C
C
CC
C
C
C
C
C
C
C
CCCCC
C
CCCCCCCCC
C
C
C
C
CC
CCC
C
CCC
C
CCC
CCCC
T
TTT
T
TT
T
T
T
T
TT
T
T
T
T
T
T
T
14/23
-
Propensity Score Matching
Education (years)
Age
12 16 20 24 28
20
30
40
50
60
70
80
1
0
PropensityScore
CCC
C
CCC
C
CCC
CCCC
T
TTT
T
TT
T
T
T
T
TT
T
T
T
T
T
T
T
14/23
-
Propensity Score Matching
Education (years)
Age
12 16 20 24 28
20
30
40
50
60
70
80
C
C
CC
C
CC
C
C
C
C
C
C
C
C
C
C
CC
C TTTT
T
T
T
T
T
T
T
T
T
TT
T
T
T
T
T
1
0
PropensityScore
CCC
C
CCC
C
CCC
CCCC
T
TTT
T
TT
T
T
T
T
TT
T
T
T
T
T
T
T
14/23
-
Propensity Score Matching
Education (years)
Age
12 16 20 24 28
20
30
40
50
60
70
80
C
C
CC
C
CC
C
C
C
C
C
C
C
C
C
C
CC
C TTTT
T
T
T
T
T
T
T
T
T
TT
T
T
T
T
T
14/23
-
Best Case: Propensity Score Matching
15/23
-
Best Case: Propensity Score Matching
Education (years)
Age
12 16 20 24 28
20
30
40
50
60
70
80
C
CCC
C
C
C
C
CC
C
C
C
CC
C
C
C
C
C
CCCC
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
CC
C
C
CCC
CC
C
C
C
CC
C
C
C
C
CC
C
CC
C
C
C C
C
C
C
CC
CC
C
CC
C
C
CC
C
C
C
C
CC
CC
C
C
C
CC
C
C
C
C
C
C
C
C
C
CC
C
CC
C
C
C
CC
CC
C
C
CC
C
C
C
C
C
C
C
C
C
CC
C
C
C CC
C
C
C
CC
C
C
C
C
C
C
CC
C
C
C
C
C
C
C
C
C
C
CCC
C
C
C
CC
CC
C
C
CC
C
C
CC
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C C
C
C C
CC
CC
C
C
C
C
CC
C
C
C C
CC
CC
C
C
C
C C
CC
C
C
C
C
C
C
C C
CC
C
CC
C
C
C
C
C
C
C
C
CC
CC
C
C
CC
CC
C
CC
C
C
C
C
CC
C
CC
C
C
C
CC
C
C
CC
C
C
C
C
C
C
C C
C
CC C
C
CC
C C
C
C
C
C
C
CC
CC
C
C
C
C
CC
C
C
C
C
C
C
C
C
C
C
C
CC
CC C
C
C
C
C
C
C C
C
C
C
CC
C
C
C
CC
CC
C
CC
C
C
C
C
C
C
C
CC
C
C
CC
C
C
C
C
C
C
C
C
C
CC
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
CC
C
C
C
C
C
CC
CC
C
CC
C
C
C
C
C
C
C
C
C
CC
C
C
CC
C
CC
CC C
CCC
C
C
C
C
C
C
C
C
CC
C
C
C
C
C
C
C
C
C
CC
C
C C
C
C
C
C
C
C CC
C
C
C
C CC
C
C
C
C
C
C
C
C
C
CC
C
C
C
C
C
C
C
C
CC
C
C
CC
C
C
C
C
CC
CC
C
C
T
TTT
T
T
T
T
TT
T
T
T
TT
T
T
T
T
T
TT TT
T
T
T
T
T
T
T
T
T
T
T
T
T
T
T
T
T
TT
T
T
TTT
TT
1
0
PropensityScore
15/23
-
Best Case: Propensity Score Matching
Education (years)
Age
12 16 20 24 28
20
30
40
50
60
70
80
C
CCC
C
C
C
C
CC
C
C
C
CC
C
C
C
C
C
CCCC
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
CC
C
C
CCC
CC
C
C
C
CC
C
C
C
C
CC
C
CC
C
C
C C
C
C
C
CC
CC
C
CC
C
C
CC
C
C
C
C
CC
CC
C
C
C
CC
C
C
C
C
C
C
C
C
C
CC
C
CC
C
C
C
CC
CC
C
C
CC
C
C
C
C
C
C
C
C
C
CC
C
C
C CC
C
C
C
CC
C
C
C
C
C
C
CC
C
C
C
C
C
C
C
C
C
C
CCC
C
C
C
CC
CC
C
C
CC
C
C
CC
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C
C C
C
C C
CC
CC
C
C
C
C
CC
C
C
C C
CC
CC
C
C
C
C C
CC
C
C
C
C
C
C
C C
CC
C
CC
C
C
C
C
C
C
C
C
CC
CC
C
C
CC
CC
C
CC
C
C
C
C
CC
C
CC
C
C
C
CC
C
C
CC
C
C
C
C
C
C
C C
C
CC C
C
CC
C C
C
C
C
C
C
CC
CC
C
C
C
C
CC
C
C
C
C
C
C
C
C
C
C
C
CC
CC C
C
C
C
C
C
C C
C
C
C
CC
C
C
C
CC
CC
C
CC
C
C
C
C
C
C
C
CC
C
C
CC
C
C
C
C
C
C
C