correcting for self-selection bias using the heckman selection correction in a valuation survey...
TRANSCRIPT
Correcting for Self-Selection Bias Using the Heckman Selection
Correction in a Valuation Survey Using Knowledge Networks
Presented at the 2005 Annual Meeting of the American Association of Public Opinion
Research
Trudy Cameron, University of Oregon
J. R. DeShazo, UCLA
Mike Dennis, Knowledge Networks
Motivating Insights
It is possible to have 20% response rate, yet still have a representative sample
It is possible to have 80% response rate, yet still have very non-representative sample
Need to know: What factors affect response propensity Whether response propensity is correlated with the survey outcome of
interest
Research Questions
Can the inferences from our two samples of respondents from the Knowledge Networks Inc (KN) consumer panel be generalized to the population?
Do observed/unobserved factors that affect the odds of a respondent being in the sample also affect the answers that he/she gives on the survey?
What we find:
Insignificant selectivity using one method
Significant, but very tiny, selectivity using an alternative method
A bit disappointing for us (no sensational results, reduced publication potential)
Probably reassuring for Knowledge Networks, since our samples appear to be reasonably representative
Heckman Selectivity Correction Intuition: Example 1
Suppose your sample matches the US population on observables: age, gender, income
Survey is about government regulation
Suppose liberals are more likely to fill out surveys, but no data on political ideology for non-respondents (unobserved heterogeneity)
Sample will have disproportionate number of liberals
Your sample is likely to overstate sympathy for government regulation (sample selection bias)
Heckman Selectivity Correction Intuition: Example 2
Suppose your sample matches the US population on observables: age, gender income
Survey is about WTP for health programs
People who are fearful about their future health are more likely to respond, but have no data for anyone on “fearfulness” (the salience of health programs)
Sample will tend to overestimate WTP for health programs
Level curves, 0
Level curves, 0
The selection “process”
Several phases of attrition (transitions) in KN samples. Assume RDD=“random”
1. RDD → recruited
2. Recruited → profiled
3. Profiled → active at time of sampling
4. Active → drawn for survey sample
5. Drawn → member of estimating sample
Can explore a different selection process for each of these transitions
Exploit RDD telephone numbers
Recruits can be asked their addresses
Some non-recruit numbers matched to addresses using reverse directories
Phone numbers with no street address? Matched (approximately) to best census tract using the geographic extent of the telephone exchange
Link to other geocoded data
Panel protection during geocoding
Using dummy identifiers, match street addresses to relevant census tract (or telephone exchange to tract, courtesy Dale Kulp at MSG), return data to KN
Get back from KN the pool of initial RDD contacts, minus all confidential addresses, with our respondent case IDs restored
Merge with auxiliary data about census tract attributes and voting behaviors
Cameron and Crawford (2004)
Data for each of the 65,000 census tracts in the year 2000 U.S. census
~ 95 count variables for different categories
Convert to proportions of population (or households, or dwellings)
Factor analysis: 15 orthogonal factors that together account for ~ 88% of variation in sociodemographic characteristics across tracts
Categories of Census VariablesPopulation density
Ethnicity; Gender; Age distribution
Family structure
Housing occupancy status; Housing characteristics
Urbanization; Residential mobility
Linguistic isolation
Educational Attainment
Disabilities
Employment Status
Industry; Occupation; Type of income
Labels: 15 orthogonal factors
"well-to-do prime" "elderly disabled"
"well-to-do seniors" "rural farm., self-employ.
"single renter twenties" "low mobil., stable neigh."
"unemployed" "Native American"
"minority single moms" "female"
"thirty-somethings" "health-care workers"
"working-age disabled" "asian-hispanic-multi, language isolation"
"some college, no grad"
2000 Presidential Voting Data
Leip (2004) Atlas of U.S. Presidential Elections: vote counts for each county
Use: % of county votes for “Gore,” “Nader,” versus “Bush and others” (omitted category)
will not be orthogonal to our 15 census factors
Empirical Illustrations
1. Analysis of “government” question in “public interventions” sample--by naïve OLS, and via preferred Heckman model
2. Analysis of selection processes leading to “private interventions” sample
Marginal selection probabilities Conditional selection probabilities Allow marginal utilities to depend on propensity to respond to
survey
Analysis 1:Heckman Selectivity Model
Public intervention sample
Find an outcome variable that is Measures an attitude that may be relevant to other research questions Can be treated as cardinal and continuous (although it is actually
discrete and ordinal) Can be modeled (naively) by OLS methods Can be generalized into a two-equation FIML selectivity model
Government Involvement in Regulating Env., Health, Safety?
Rating Frequency Percent
1 – minimally involved 134 4.60 2 90 3.09 3 199 6.84 4 493 16.94 5 609 20.92 6 591 20.30 7 – heavily involved 795 27.31
Total 2,911 100.00
Heckman correction model:bias not statistically significant
Fail to reject at 5% level, at 10% level (but close)
Point estimate of error correlation = +0.10 May be more likely to respond if approve of govt reg
Interpretation: Insufficient signal-to-noise to conclude that there is non-random selection
Reassuring, but could stem from noise due to: – Census tract factors, county votes rather than individual characteristics
– Treating ordinal ratings as cardinal and continuous
Implications for “govt” variable
1
OLS 2
Heckmana 3
simulated 4
simulated (fitted) (fitted) (index) (prob)
Observations 2911 2911 2911 2911
Mean 5.17 4.69 5.18 5.17
5% 4.80 4.34 1.99 2.00 25% 4.97 4.50 4.02 4.00 50% 5.12 4.64 5.03 5.00 75% 5.33 4.85 6.97 7.00 95% 5.67 5.18 7.03 7.00
Log L -5549.30 -23061.82 -5548.07 -5548.63
Analysis 2Conditional Logit Choice Models
Very attractive properties for analyzing multiple discrete choices,…but
No established methods for joint modeling of selection propensity and outcomes in the form of 3-way choices
Testable Hypothesis:
Do the marginal utilities of key attributes depend upon the fitted selection index (or the fitted selection probability)?
If yes: then observable heterogeneity in the odds of being in the sample contributes to heterogeneity in the apparent preferences in the estimating sample
If no: greater confidence in representativeness of the estimating sample (although still no certainty)
RDD contact dispositions
Variable Proportion
Disposition (n=525,190) Recruited 0.354 Profiled 0.183 Active panel at sample time 0.075 (Eligible at sample time…>24 years) (0.068) Drawn for sample 0.006 Estimating sample 0.003
Results: Response propensity models
Use response propensity as a shifter on the parameters of conditional logit choice models
Only one marginal utility parameter (related to the disutility of a sick-year) appears robustly sensitive to selection propensity
Baseline coefficient is -50 units, average shift coefficient is on the order of 3 units, times a deviation in fitted response probabilities that averages about 0.004
Conclusions 1
For our samples: hard to find convincing and robust evidence of substantial sample selection bias in models for outcome variables in two KN samples
Good news for Knowledge Networks; not so good for us, as researchers, in terms of the publication prospects for this dimension of our work
Conclusions 2
Analysis 1: Insignificant point estimate of bias in distribution of attitudes toward regulation on the order of “10% too much in favor”
Analysis 2: Statistically significant (but tiny) heterogeneity in key parameters across response propensities in systematically varying parameters models
Guidance High response rates do not necessarily eliminate biases in survey samples
Weights help (so sample matches population on observable dimensions), but weights are not necessarily a fix if unobservables are correlated
Cannot tell if correlated unobservables are a problem without doing this type of analysis
Need to model the selection process explicitly Need to explain differences in response propensities Need analogous data for respondents and for people who do not
respond– E.g. Census tract factors and county voting percentages– Anything else that might capture salience of survey topic (e.g. county
mortality rates from same diseases covered in survey, hospital densities)