ie slide04

Upload: leo-kaligis

Post on 07-Jan-2016

11 views

Category:

Documents


0 download

DESCRIPTION

economics econ 2206

TRANSCRIPT

  • Introductory Econometrics

    ECON2206/ECON3209

    Slides04

    Lecturer: Minxian Yang

    ie_Slides04 my, School of Economics, UNSW 1

    Assignment -1 is due in Week 5. Please submit it

    to your tutor at the beginning of your tutorial.

    See Course Outline for more information.Staple your pages together. Do not submit loose pages.

    Do not use plastic sheets or binders.

  • 4. Multiple Regression Model: Inference (Ch4)

    4. Multiple Regression Model: Inference

    Lecture plan

    Classical linear model assumptions

    Sampling distribution of OLS estimators under CLM

    Testing hypothesis about one population parameter

    p-values

    Confidence intervals

    Test hypothesis with CI

    ie_Slides04 my, School of Economics, UNSW 2

  • 4. Multiple Regression Model: Inference (Ch4)

    Motivation: y = 0 + 1x1 +...+ kxk + u

    Goal is to gain knowledge about the population parameters (s) in the model.

    OLS provides the point estimates of the parameters.

    OLS will get it right on average (being unbiased).

    Knowing the mean and variance of is not enough.

    how to decide if a hypothesis is supported or not?

    what we can say about the true values?

    We need the sampling distribution of the OLS

    estimators to answer these questions.

    To simplify, we use a strong assumption here (and will relax it for large-sample cases).

    ie_Slides04 my, School of Economics, UNSW 3

    j

  • 4. Multiple Regression Model: Inference (Ch4)

    Normality assumption

    6. (MLR6, normality) The disturbance u is independent

    of all explanatory variables and normally distributed

    with mean zero and variance 2:

    u ~ Normal(0, 2).

    This is a very strong assumption. It implies both MLR4 (ZCM) and MLR5 (homoskedasticity).

    MLR1-6 together are known as the classical linear model (CLM) assumptions.

    Under CLM, the OLS produces the minimum variance unbiased estimators.

    They are the best of unbiased estimators (not just the best of linear unbiased estimators).

    ie_Slides04 my, School of Economics, UNSW 4

  • 4. Multiple Regression Model: Inference (Ch4)

    Normality assumption

    CLM implies

    y | x ~ Normal( 0 + 1x1 +...+ kxk , 2).

    CLM also implies is normally distributed.

    Whether or not MLR6 is a reasonable assumption depends on data

    Is it reasonable for wage model, given that no wage can be negative?

    Empirically, it is reasonable for log(wage) model.

    MLR6 is restrictive. But the results here will be useful for large-sample cases (Ch5) without MLR6.

    ie_Slides04 my, School of Economics, UNSW 5

    j

    It is completely characterised

    by the mean and variance.

  • 4. Multiple Regression Model: Inference (Ch4)

    Sampling distribution of OLS

    Theorem 4.1 (normal sampling distribution)

    Under CLM, conditional on independent variables,

    where the variance is given in Ch3 (ie_Slides03):

    It implies:

    ie_Slides04 my, School of Economics, UNSW 6

    ,,Normal~ 2jjj

    .,...,1 ,)1(

    )(2

    22 kj

    RSSTVar

    jj

    jj

    .)()( ),,(Normal~)(

    jj

    j

    jjVarsd

    sd

    10

  • 4. Multiple Regression Model: Inference (Ch4)

    Sampling distribution of OLS

    In practice, we have to estimate 2 and

    Theorem 4.2 (t-distribution)

    Under CLM, conditional on independent variables,

    where tn-k-1 is the t-distribution with n-k-1 df.

    This is a basis for statistical inference.

    ie_Slides04 my, School of Economics, UNSW 7

    .,..., ,)(

    )( kj

    RSSTarV

    jj

    j 11 2

    2

    ,)()( ,~)(

    jjkn

    j

    jjarVset

    se

    1

    Known as

    standard error

    )( jVar

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing simple null hypothesis

    Some questions of interest may be formulated as a simple null hypothesis and an alternative

    hypothesis about a population parameter,

    H0: j = aj , H1: j aj (or > aj or < aj ),

    where aj is a know value (often zero).

    eg. In the log wage model

    log(wage) = 0 + 1educ + 2 exper + 3 tenure + u,

    H0: 1 = 0, is economically interesting. It says that, holding exper and tenure fixed, a persons education level has no effect on wage.

    ie_Slides04 my, School of Economics, UNSW 8

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing simple null hypothesis

    To test a simple null hypothesis, the test statistic is usually the t-statistic

    We will call the t-statistic the t-ratio when aj = 0. The STATA output includes the t-ratios.

    By Theorem 4.2, the t-statistic has the t-distribution with n-k-1 df, under the null H0.

    When df or n is large, the t-distribution approaches to the standard normal distribution. See Table G.2.

    The decision rule depends on the alternative H1.

    ie_Slides04 my, School of Economics, UNSW 9

    .)()( ,)(

    jj

    j

    jjarVse

    se

    at

    j

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing simple null hypothesis

    When H0 is true, we can choose a critical value csuch that the probability of the t-statistic exceeding c

    is a small number, known as significant level.

    c depends on df and the significant level. (Use normal critical values when df >120.)

    If we reject H0 whenever the t-statistic exceeds c, the probability that we make a (Type I) error is small.

    Hence, reject H0 whenever the t-statistic exceeds c is a reasonable decision rule.

    What do we mean by the t-statistic exceeds c?

    It depends on the alternative hypothesis H1.

    ie_Slides04 my, School of Economics, UNSW 10

    .05.0)|(| eg. ctPj

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing simple null hypothesis

    Decision rule :

    eg. 5% significant level, df = 19.

    For one tail H1, c = 1.729.

    For two tail H1, c = 2.093.

    Table G.2 of Wooldridge

    -c 0 c

    ie_Slides04 my, School of Economics, UNSW 11

    H0: j = ajLower

    tail H1

    Upper

    tail H1

    Two

    tail H1

    Alternative H1: j < aj j > aj j aj

    Reject H0 when < -c > c | | > cjt jt jt

    f(t)

    t

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing simple null hypothesis

    Example 4.1

    log wage model (standard errors are in brackets):

    log(wage) = .284 + .092educ + .0041expr + .022tenure

    (.104) (.007) (.0017) (.003)

    n = 526, R2 = .316

    Q. Is return to education statistically significant at the

    1% level, after controlling for experience and tenure?

    Hypotheses: H0: educ = 0 vs H1: educ 0

    Test statistic and decision rule: reject H0 if

    Critical value (large df, normal): c = 2.576

    Conclusion: reject H0 at the 1% level because

    ie_Slides04 my, School of Economics, UNSW 12

    cteduc

    ||

    .../. || cteduc

    14913007092

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing simple null hypothesis

    Example 4.5. Housing Prices and Air Pollution:

    log(price) = 11.08 .954log(nox) .134log(dist )

    (.32) (.117) (.043)

    + .255rooms .052stratio

    (.019) (.006)

    n = 506, R2 = .581

    Hypotheses: H0: nox = 1 vs H1: nox > 1

    Test statistic:

    Decision rule: reject H0 if

    5% Critical value: c = 1.645

    Conclusion: do not reject H0 at the 5% level because

    ie_Slides04 my, School of Economics, UNSW 13

    )(/)( noxnox setnox

    1

    .../).( ctnox

    3931171954

    ctnox

    Is the price elasticity

    w.r.t. nox equal to 1? All coefficients are significantly different

    from 0 at the 5% level.

    After controlling for

    dist, rooms and

    stratio, there is little

    evidence against H0.

  • 4. Multiple Regression Model: Inference (Ch4)

    Terminology

    In the conclusion, we need to be explicit about the hypotheses and the significant level, eg,

    H0 is (is not) rejected in favour of H1 at the 5% level of significance.

    In the case of H0: j = 0 and H1: j 0, we say

    xj is (is not) statistically significant at the 5% level of significance. or

    xj is (is not) statistically different from 0 at the 5% level of significance.

    ie_Slides04 my, School of Economics, UNSW 14

  • 4. Multiple Regression Model: Inference (Ch4)

    p-values

    The choice of significant level is somewhat arbitrary. Different researchers may have different choices

    (We tend to use high levels with small samples).

    It is more informative to measure the data evidence for H0. The p-value serves this role.

    The p-value is the probability of the null distribution beyond the observed test statistic:

    reported routinely by software.

    Smaller the p-value, stronger

    the evidence against H0.

    observed tie_Slides04 my, School of Economics, UNSW 15

    f(t)

    t

    |)| |(|value j

    tTPp

  • 4. Multiple Regression Model: Inference (Ch4)

    Economic/statistical significance

    An explanatory variable is statistically significant when the size of the t-ratio is sufficiently large

    (beyond the critical value c).

    An explanatory variable is economically (or practically) significant when the size of the estimate

    is sufficiently large (in comparison to the size of y).

    An important x should be both statistically and economically significant.

    Over-emphasising statistical significance may lead to false conclusion about the importance of an x.

    See the guidelines on p137, and Example 4.6-4.7.

    ie_Slides04 my, School of Economics, UNSW 16

    j

    t

    j

  • 4. Multiple Regression Model: Inference (Ch4)

    Confidence intervals

    The confidence interval (CI) for j is based on Theorem 4.2, namely,

    which directly leads to the CI

    Here, c is the tn-k-1 critical value, dependent on l.o.c..

    Example 4.5. The 95%CI for nox:

    n-k-1 = 501, c = 1.96 (large sample, use normal cv),

    ie_Slides04 my, School of Economics, UNSW 17

    ].. ,.[)(... 7251831117961954

    ]. ,[)]( ),([)( ULsecsecsec jjjjjj

    ,~)(

    1

    kn

    j

    jjt

    se

    The probability that [L, U]

    covers the parameter is

    the level of confidence.

  • 4. Multiple Regression Model: Inference (Ch4)

    Confidence intervals

    To construct CI, we need and c. For c, we need df and the confidence level.

    When df is large (> 120), the tn-k-1 distribution is very close to the normal distribution and we use N(0,1)

    critical values.

    eg. For large df, the 95% CI is about

    The interpretation of 95% CI

    If many random samples are drawn and [L, U] is

    computed for each sample, then 95% of these [L, U]

    will cover the true population parameter j.

    ie_Slides04 my, School of Economics, UNSW 18

    ] ,[)]( ),([)( ULsecsecsec jjjjjj

    , j )( jse

    ).( jj se 2

  • 4. Multiple Regression Model: Inference (Ch4)

    Confidence intervals

    The width of CI depends on the standard error and the critical value c.

    high confidence level large c wide CI,

    large standard error wide CI.

    CI and two-tailed test

    test H0: j = aj against H1: j aj .

    reject H0 at the 5% significant level if (and only if) the 95% CI does not contain aj.

    Example 4.5

    95% CI = [-1.183, -0.725].

    Do not reject H0: nox = -1 in favour of H1: nox -1 at the 5% significant level.

    ie_Slides04 my, School of Economics, UNSW 19

    )( jse

  • 4. Multiple Regression Model: Inference (Ch4)

    Summary so far CLM assumptions

    Sampling distribution of OLS estimators

    Standard error, t-statistic, t-distribution and df

    Testing hypothesis about a single

    p-Values

    Confidence interval for a single

    CI and two-tailed test

    To do: the F test

    hypotheses about a linear combination of parameters

    multiple linear restrictions on parameters

    STATA

    ie_Slides04 my, School of Economics, UNSW 20

  • 4. Multiple Regression Model: Inference (Ch4)

    Hypotheses about a linear combination of parameters

    In the log wage model

    log(wage) = 0 + 1educ + 2exper + u,

    we wish to see whether or not educ has the same

    causal effect on log(wage) as exper, ie, to test

    H0: 1 2 = 0 vs H1: 1 2 0,

    which involve a combination of 2 parameters.

    If we had , we could use

    But is not usually reported by software.

    ie_Slides04 my, School of Economics, UNSW 21

    )( 21 se

    .)(

    21

    21

    21

    se

    t

    )( 21 se

    21

    2121

    21

    2 /)],cov()var()[var(

    )(

    se

  • 4. Multiple Regression Model: Inference (Ch4)

    Hypotheses about a linear combination of parameters

    We re-parameterise the log wage model

    log(wage) = 0 + educ + 2(exper+educ) + u,

    where = 1 2 and 1 is replaced by + 2.

    The hypotheses become

    H0: = 0 vs H1: 0,

    which can easily be tested by regressing log(wage)

    on educ and (exper+educ).

    The main idea here is to isolate the parameter of interest = 1 2 by re-parameterisation. The OLS output provides both and .

    ie_Slides04 my, School of Economics, UNSW 22

    )(se

  • 4. Multiple Regression Model: Inference (Ch4)

    Hypotheses about a linear combination of parameters

    eg. log wage model (standard errors are in brackets):

    log(wage) = .284 + .0920educ + .0041expr + .022tenure

    (.104) (.0073) (.0017) (.003)

    n = 526, R2 = .316

    Hypotheses H0: educexpr= 0 vs H1: educexpr 0.

    Re-parameterised model with exed=exper+educ

    log(wage) = .284 + .0879educ + .0041exed + .022tenure

    (.104) (.0070) (.0017) (.003)

    n = 526, R2 = .316

    Hypotheses H0: = 0 vs H1: 0.

    Test statistic t = .0879/.0070 = 12.59.

    ie_Slides04 my, School of Economics, UNSW 23

  • 4. Multiple Regression Model: Inference (Ch4)

    Hypotheses about a linear combination of parameters

    ie_Slides04 my, School of Economics, UNSW 24

    How do you test

    H0: educ2expr= 0 by re-parameterisation?

    t-ratios

    p-values

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    Exclusion restrictions

    It is of interest to check whether or not a group of xvariables has a joint effect on y (with the rest of x

    variables as controls).

    This question is formulated as the null (H0) that all coefficients of the group is zero, called exclusion

    restrictions. The model under H0 is known as the

    restricted model.

    The alternative (H1) is simply that the null is false. The model under H1 is known as the unrestricted model.

    In general, multiple linear restrictions involve more than one linear restrictions on parameters.

    ie_Slides04 my, School of Economics, UNSW 25

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    Example

    Child birth weight and parents education

    bwght = 0 + 1cigs + 2parity + 3faminc

    + 4motheduc + 5fatheduc + u

    bwght : birth weight

    cigs : average cigarettes per day by mother

    parity : birth order

    faminc : family income

    motheduc : years of education for mother

    fatheduc : years of education for father

    H0: 4 = 0 and 5 = 0. vs H1: H0 is false.

    We need the F statistic to test the joint hypotheses.

    ie_Slides04 my, School of Economics, UNSW 26

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    Example

    Unrestricted model (ur)

    bwght = 0 + 1cigs + 2parity + 3faminc

    + 4motheduc + 5fatheduc + u

    SSRur.

    Restricted model (r)

    bwght = 0 + 1cigs + 2parity + 3faminc + u(r)

    SSRr.

    If SSRr is much too greater than SSRur, we reject the

    restrictions (ie. H0).

    But how greater is much too greater?

    ie_Slides04 my, School of Economics, UNSW 27

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    Example

    The F statistic is the relative difference between

    SSRr and SSRur :

    Under H0, F has the F-distribution

    with (2, n-6) degrees of freedom.

    Decision rule: reject if F > c,

    where c is the F2,n-6 critical value.

    ie_Slides04 my, School of Economics, UNSW 28

    .)/()(

    /)(

    )/(

    /)(

    61

    2

    6

    22

    22

    nR

    RR

    nSSR

    SSRSSRF

    ur

    rur

    ur

    urr

    626

    2

    n

    ur

    urr FnSSR

    SSRSSRF ,~

    )/(

    /)(# of restrictions

    # of coefficients

    in unrestricted

    model

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    Example

    Use the data in BWGHT.RAW : n = 1191,

    SSRr = 465166.792 and SSRur = 464041.135.

    The 5% F2,n-6 critical value is c = 3.00.

    (see Table G.3b)

    According to the decision rule, H0 is not rejected at

    the 5% level because F < c.

    ie_Slides04 my, School of Economics, UNSW 29

    .43.1)6/(

    2/)(

    nSSR

    SSRSSRF

    ur

    urr

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    Example

    ie_Slides04 my, School of Economics, UNSW 30

    SSR

    You can use STATA

    test to do the test.

    overall significance

    test (see p34)

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    General case with CLM assumptions

    The unrestricted model:

    y = 0 + 1x1 +...+ kxk + u.

    There are q restrictions:

    H0: k-q+1 = 0, ..., k = 0

    which lead to the restricted model:

    y = 0 + 1x1 +...+ k-qxk-q + u(r).

    Test statistic

    Decision rule: reject H0 if F > c (Fq,n-k-1 critical value).

    ie_Slides04 my, School of Economics, UNSW 31

    .H under ~)/(

    /)(0, 1

    1

    knq

    ur

    urr FknSSR

    qSSRSSRF

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    General case with CLM assumptions

    If H0 is rejected, we say that xk-q+1, ..., xk are jointly statistically significant.

    If H0 is not rejected, we say that xk-q+1, ..., xk are jointly statistically insignificant, which justifies dropping them

    from the model.

    It is possible that a group of variables are jointly significant but individually insignificant. This is a

    symptom of a group of highly correlated variables.

    The p-value for F test is the probability of F-distribution beyond observed F-stat

    ie_Slides04 my, School of Economics, UNSW 32

    ).(value , FFPp knq 1

    When

    highly

    correlated,

    it is hard

    to

    precisely

    estimate

    their

    individual

    partial

    effects.

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    The F distribution

    F and t statistics

    When q = 1, H0

    can be tested

    with either t-stat

    or F-stat.

    It turns out

    (t-stat)2 = F-stat

    in the case q = 1.

    ie_Slides04 my, School of Economics, UNSW 33

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    The overall significance of a regression

    When q = k, the null H0: 1 = 0, ..., k = 0 is routinely tested by most regression packages, known as the F

    test for overall significance.

    The null is that none of the explanatory variables has an effect on y. The restricted model is simply

    y = 0 + u.

    The F-stat under the null has an Fk,n-k-1 distribution. As the R-squared is zero under the null, this F-stat is

    where R2 is from the unrestricted model.

    ie_Slides04 my, School of Economics, UNSW 34

    ,)/()(

    /

    11 2

    2

    knR

    kRF

  • 4. Multiple Regression Model: Inference (Ch4)

    Testing multiple linear restrictions: the F test

    Testing general linear restriction

    Based on restricted and unrestricted regression SSRs, the F test is applicable to testing any linear

    restrictions on parameters. For example,

    H0: 1 = 1, 2 = 0, 3 = 24 .

    We only need the SSRr from the restricted model (H0), which may involve reparameterisation, and SSRur from

    the unrestricted model.

    eg. House price (4.47)

    log(price) = 0 + 1log(assess) + 2log(lotsize)

    + 3log(sqrft) + 4bdrms + u

    If the assessment is a rational valuation,

    H0: 1 = 1, 2 = 0, 3 = 0, 4 = 0 should hold.

    ie_Slides04 my, School of Economics, UNSW 35

  • 4. Multiple Regression Model: Inference (Ch4)

    Reporting regression results Good practice (minimum)

    Report estimated coefficients AND standard errors

    Report the mean of dependent variable

    Report sample size

    Report R-squared and SSR

    Report in equation-form if the number of equations is small

    Report in table-form and indicate dependent variable

    eg. log wage model

    ie_Slides04 my, School of Economics, UNSW 36

    Dependent variable: log(WAGE), mean = 1.623

    Model 1 Model 2

    Variable Coeff Stderr Coeff Stderr

    CONSTANT 0.2844 0.1042 0.5014 0.1019

    EDUC 0.0920 0.0073 0.0875 0.0069

    EXPER 0.0041 0.0017 0.0046 0.0016

    TENURE 0.0221 0.0031 0.0174 0.0030

    FEMALE -0.3012 0.0373

    Sample size 526 526

    R-squared 0.316 0.392

    SSR 101.460 90.144

  • 4. Multiple Regression Model: Inference (Ch4)

    Summary The methods covered allow us to infer knowledge about

    population parameters from a random sample.

    CLM assumptions OLS estimators follow normal distribution t-stat and F-stat follow t and F-distributions.

    To test hypotheses: choose a (small) level of significance, find the critical value, compute the test statistic, use the

    decision rule (which depends on the alternative) to draw

    conclusion.

    To construct CI: choose a (large) level of confidence, find the critical value and standard error, use

    In practice, both statistical and economic significance of your results need to be commented on.

    ie_Slides04 my, School of Economics, UNSW 37

    ).( jj sec

  • 4. Multiple Regression Model: Inference (Ch4)

    What we are able to do now We covered statistical inference methods, base on the

    sampling distribution of the OLS estimators under the CLM

    assumptions.

    We are able to test hypotheses about a parameter.

    We are able to construct confidence intervals for parameters.

    We are able to test single or multiple restrictions on parameters.

    Under the CLM assumptions, the above inference methods are exact, in the sense that we know the exact

    distribution of t-stat or F-stat, regardless of the sample

    size.

    But MLR6 in the CLM assumptions is too strong....

    ie_Slides04 my, School of Economics, UNSW 38