daniel stein economist, dime istanbul, may 12, 2015 measuring impact: experiments

26
Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

Upload: della-potter

Post on 28-Dec-2015

215 views

Category:

Documents


1 download

TRANSCRIPT

Page 1: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

Daniel SteinEconomist, DIME

Istanbul, May 12, 2015

Measuring Impact: Experiments

Page 2: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

2

Practical Challenge

• Find a very good counterfactual to tell us what would have happened…

• Non-experimental methods require many assumptions and very good data

• Is there an easier way to go?

Page 3: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

3

Experiments• Other names: Randomized Control Trials

(RCTs), Randomization or random assignment

• Assignment to treatment and control is based on chance, it is random (like flipping a coin)

• This is the best way of recovering the counterfactual

Page 4: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

Randomization? Thats Not For Me!

• Randomization cant offer all questions, but it can be a useful tool in many types of projects

• Maybe you cant randomize trade laws, but maybe you can experiment with:– Information– Incentives to export– Loans– Etc

• Randomization allows clear answers to YOUR questions!

Page 5: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

5

Experiments: plan

Purpose of randomizing and how it works: Intuition

Important Concepts

Set Up

Different ways of randomizing

Real Threats to Your Experiment

Page 6: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

6

Purpose… • Identify the causal impact of an

intervention on some outcomes of interest• To do so, you need:– Identify what is your targeted population (or

eligible group)– Select two groups: treatment and control– Assure the groups are, on average, identical in

observed and unobserved characteristics

• How to do it? – You need a random assignment (or

experiment)

Page 7: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

Basic intuition: random sampling ≠ random assignment

1. Population

External Validity

2. Evaluation sample

3. Randomization

Internal Validity

Comparison

Treatment

Page 8: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

More intuition…

Automatically excluded -- ineligibles

Automatically included

May or may not enter – eligibles to enter

“A program is targeted to the most needed firms so they can catch-up”

Page 9: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

More intuition…

Automatically excluded -- ineligibles

Automatically included -- eligibles

May or may not enter – randomize here!

“A program is targeted to the most needed firms so they can catch-up”

Page 10: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

10

Set Upo Example: The government of Eurasia wants to test the effect of

a training program. It can be evaluated within or between firms. You’re hired to do an impact evaluation

o Case 1: You first consider to work within firmso A firm with 800 employees is selected and it’s believed that some of

them could perform better if received some trainingo What can be done to see if the training is actually improving workers’

productivity on the job?

Page 11: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

11

1. Pure Randomizationo We don’t know beforehand if the program

works or not. It’s then decided to test a pilot:o Sampling frame: 400 employeeso Give to all workers the same chance to

participate into the piloto A random number is assigned to each

worker, the workers are ordered in an ascending order based on the number they were given, and the first 200 are selected to the program

o This is an example of pure randomization in practice

o Challenge you may face: unselected workers get unhappy and respond to the selection process changing behavior

o Why would that be a problem?

Page 12: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

1. Pure Randomization

The globe is your firm

Randomize treatment

Comparison: 200

Treatment: 200

Average performance

outcome in the end of the month (YC)

Average performance

outcome in the end of the month (YT)

Effect = (YT) - (YC)

Page 13: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

13

• If the performance of the trained workers affects the behavior of the untrained in the same team, then…

Example of Spillover Effects

Page 14: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

14

• If the performance of the trained workers affects the behavior of the untrained in the same team, then…

This is an example of spillover effects

Example of Spillover Effects

Page 15: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

15

1. Pure Randomization

T

T

T

C

C

C

C

TCo Suppose that workers are teamed up

o Work teams are randomly assigned to get trained instead

o Advantages of working at cluster levelo Reduce risk of spillovers within

groups

Page 16: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

16

Important Concepts

• Pure randomization: unit of intervention = unit of analysis – you randomize workers and assess the workers’

performance individually– you randomize team and assess team’s performance

• Cluster randomization: unit of intervention ≠ unit of analysis– you randomize firms and assess workers’ performance

individually

Page 17: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

Important Concepts

Unit of Randomization: choose according to type of programo Individual/Household/Workerso Firm/School/Health Clinic

(Clusters)o Block/Village/Community

(Clusters)o Ward/District/Region (Clusters)

As a rule of thumb, randomize at the smallest viable unit of implementation.

Page 18: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

18

• Case 2: You decide to explore the intervention across firms but you’re worried about some practical issues such as limited capacity, low participation and inexistence of pure control

• What can you do?

Set Up

Page 19: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

Opportunities for Randomization

• Budget constraints prevent full coverage– Random assignment (lottery) is fair and transparent

• Limited implementation capacity– Randomized phase-in gives all the same chance to

go first

• No evidence on which alternative is best– Random variation in treatment with equal ex ante

chance of success

• Take up of existing program is not complete– Encouragement design: Randomly provide

information or incentive for some to sign up

19

Page 20: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

20

Scenario

• The government of Eurasia has identified 200 firms who have export potential but are not accessing markets

• They want to implement a program to give them export advisory services.

• What types of randomized designs might be possible?

Page 21: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

21

Budget Constraints->Pure Randomization

• Imagine that the government only has the budget to deliver the service to 100 out of 200 firms.

• How to choose?• This is an opportunity for a pure

randomized selection• Randomization is fair, transparent,

and allow a rigorous IE!

Page 22: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

22

Capacity Constraints: Phase-in Design

• The government only has the capacity to deliver 100 out of the 200 eligible in the first year of the program.

• There is a waiting list and the government wants to give firms the same chance to participate first– You randomly assign 100 to participate now and treat the other

100 later – say, one year.

– This solved your problem, but notice that as soon as the control firms enter the program you are no longer able to identify the impact of it – no long-term effects!

Page 23: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

Not sure to work best: variation in treatment

• Let’s say the government is not sure what to offer the firms: export subsidies or training.

• You can test which works better: randomly assign 100 firms to receive training and 100 to receive subsidies.

• Figure out which is more effective!

Page 24: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

24

Lack of Interest->Encouragement design

• Now suppose that the government does not want to exclude anyone from the program who really wants it.

• This is the opportunity for an Encouragement Design• You can randomly select 100 of the 200 firms to get

special encouragement (visits, phone calls, tax breaks) if they take advantage of the program. The other 100 can still sign up.

• If your encouragement is effective, this still allows you to evaluate the program.

Page 25: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

25

Why randomize?

• Randomization is the “gold standard”• Compared to other techniques,

results from randomization will be treated differently by donors, researchers, policy makers.

• Go for the gold!

Page 26: Daniel Stein Economist, DIME Istanbul, May 12, 2015 Measuring Impact: Experiments

Thank you! facebook.com/ieKnow

#impacteval

blogs.worldbank.org/impactevaluations

microdata.worldbank.org/index.php/catalog/impact_evaluation

http://dime.worldbank.orgWEB