statistics and data analysis
DESCRIPTION
Statistics and Data Analysis. Professor William Greene Stern School of Business IOMS Department Department of Economics. Statistics and Data Analysis. Part 24 – Multiple Regression: 4. Hypothesis Tests in Multiple Regression. Simple regression: Test β = 0 - PowerPoint PPT PresentationTRANSCRIPT
![Page 1: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/1.jpg)
Part 24: Multiple Regression – Part 424-1/45
Statistics and Data Analysis
Professor William GreeneStern School of Business
IOMS DepartmentDepartment of Economics
![Page 2: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/2.jpg)
Part 24: Multiple Regression – Part 424-2/45
Statistics and Data Analysis
Part 24 – Multiple Regression: 4
![Page 3: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/3.jpg)
Part 24: Multiple Regression – Part 424-3/45
Hypothesis Tests in Multiple Regression Simple regression: Test β = 0 Testing about individual coefficients
in a multiple regression R2 as the fit measure in a multiple
regression Testing R2 = 0 Testing about sets of coefficients Testing whether two groups have the
same model
![Page 4: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/4.jpg)
Part 24: Multiple Regression – Part 424-4/45
Regression Analysis Investigate: Is the coefficient in a regression model really nonzero? Testing procedure:
Model: y = α + βx + ε Hypothesis: H0: β = 0. Rejection region: Least squares coefficient is far from zero.
Test: α level for the test = 0.05 as usual Compute t = b/StandardError Reject H0 if t is above the critical value
1.96 if large sample Value from t table if small sample.
Reject H0 if reported P value is less than α level
Degrees of Freedom for the t statistic is N-2
![Page 5: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/5.jpg)
Part 24: Multiple Regression – Part 424-5/45
Application: Monet Paintings Does the size of the
painting really explain the sale prices of Monet’s paintings?
Investigate: Compute the regression
Hypothesis: The slope is actually zero.
Rejection region: Slope estimates that are very far from zero.
The hypothesis that β = 0 is rejected
![Page 6: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/6.jpg)
Part 24: Multiple Regression – Part 424-6/45
An Equivalent Test Is there a
relationship? H0: No correlation Rejection region:
Large R2. Test: F= Reject H0 if F > 4 Math result: F = t2.
2
2
(N-2)R1 - R
Degrees of Freedom for the F statistic are 1 and N-2
![Page 7: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/7.jpg)
Part 24: Multiple Regression – Part 424-7/45
Partial Effects in a Multiple Regression Hypothesis: If we include the signature effect, size does not
explain the sale prices of Monet paintings. Test: Compute the multiple regression; then H0: β1 = 0. α level for the test = 0.05 as usual Rejection Region: Large value of b1 (coefficient) Test based on t = b1/StandardError
Regression Analysis: ln (US$) versus ln (SurfaceArea), Signed The regression equation isln (US$) = 4.12 + 1.35 ln (SurfaceArea) + 1.26 SignedPredictor Coef SE Coef T PConstant 4.1222 0.5585 7.38 0.000ln (SurfaceArea) 1.3458 0.08151 16.51 0.000Signed 1.2618 0.1249 10.11 0.000S = 0.992509 R-Sq = 46.2% R-Sq(adj) = 46.0%
Reject H0.
Degrees of Freedom for the t statistic is N-3 = N-number of predictors – 1.
![Page 8: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/8.jpg)
Part 24: Multiple Regression – Part 424-8/45
Use individual “T” statistics.T > +2 or T < -2 suggests the variable is “significant.”T for LogPCMacs = +9.66.This is large.
CoefT =SE Coef
![Page 9: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/9.jpg)
Part 24: Multiple Regression – Part 424-9/45
Women appear to assess health satisfaction differently from men.
![Page 10: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/10.jpg)
Part 24: Multiple Regression – Part 424-10/45
Or do they? Not when other things are held constant
![Page 11: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/11.jpg)
Part 24: Multiple Regression – Part 424-11/45
![Page 12: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/12.jpg)
Part 24: Multiple Regression – Part 424-12/45
Confidence Interval for Regression Coefficient
Coefficient on OwnRent Estimate = +0.040923 Standard error = 0.007141 Confidence interval
0.040923 ± 1.96 X 0.007141 (large sample)= 0.040923 ± 0.013996= 0.02693 to 0.05492
Form a confidence interval for the coefficient on SelfEmpl. (Left for the reader)
![Page 13: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/13.jpg)
Part 24: Multiple Regression – Part 424-13/45
Model Fit How well does the model fit the data? R2 measures fit – the larger the better
Time series: expect .9 or better Cross sections: it depends
Social science data: .1 is good Industry or market data: .5 is routine
Use R2 to compare models and find the right model
![Page 14: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/14.jpg)
Part 24: Multiple Regression – Part 424-14/45
Dear Prof William I hope you are doing great.I have got one of your presentations on Statistics and Data Analysis, particularly on regression modeling. There you said that R squared value could come around .2 and not bad for large scale survey data. Currently, I am working on a large scale survey data set data (1975 samples) and r squared value came as .30 which is low. So, I need to justify this. I thought to consider your presentation in this case. However, do you have any reference book which I can refer while justifying low r squared value of my findings? The purpose is scientific article.
![Page 15: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/15.jpg)
Part 24: Multiple Regression – Part 424-15/45
Pretty Good Fit: R2 = .722
Regression of Fuel Bill on Number of Rooms
![Page 16: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/16.jpg)
Part 24: Multiple Regression – Part 424-16/45
A Huge Theorem
R2 always goes up when you add variables to your model.
Always.
![Page 17: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/17.jpg)
Part 24: Multiple Regression – Part 424-17/45
The Adjusted R Squared Adjusted R2 penalizes your model for
obtaining its fit with lots of variables. Adjusted R2 = 1 – [(N-1)/(N-K-1)]*(1 – R2)
Adjusted R2 is denoted Adjusted R2 is not the mean of anything
and it is not a square. This is just a name.
2R
![Page 18: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/18.jpg)
Part 24: Multiple Regression – Part 424-18/45
The Adjusted R SquaredS = 0.952237 R-Sq = 57.0% R-Sq(adj) = 56.6%
Analysis of Variance
Source DF SS MS F PRegression 20 2617.58 130.88 144.34 0.000Residual Error 2177 1974.01 0.91Total 2197 4591.58
If N is very large, R2 and Adjusted R2 will not differ by very much.2198 is quite large for this purpose.
![Page 19: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/19.jpg)
Part 24: Multiple Regression – Part 424-19/45
Success Measure
Hypothesis: There is no regression. Equivalent Hypothesis: R2 = 0. How to test: For now, rough rule.
Look for F > 2 for multiple regression(Critical F was 4 for simple regression)
F = 144.34 for Movie Madness
![Page 20: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/20.jpg)
Part 24: Multiple Regression – Part 424-20/45
Testing “The Regression”
1 1 2 2 K K
0 1 2 K
1
Model: y = + x + x + ... + x + Hypothesis: The x variables are not relevant to y. H : 0 and 0 and ... 0 H : At least one coefficient is not zero.Set level to 0.05 as us
2
2
ual.Rejection region: In principle, values of coefficients that are far from zero
Rejection region for purposes of the test: Large R . The test is
equivalent to a test of the hypothesis that R = 2
2
0
0.
R / KTest procedure: Compute F = (1 - R )/(N-K-1)
Reject H if F is large. Critical value depends on K and N-K-1(see next page). (F is not the square of any t statistic if K > 1.)
Degrees of Freedom for the F statistic are K and N-K-1
![Page 21: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/21.jpg)
Part 24: Multiple Regression – Part 424-21/45
The F Test for the Model
Determine the appropriate “critical” value from the table.
Is the F from the computed model larger than the theoretical F from the table? Yes: Conclude the relationship is significant No: Conclude R2= 0.
![Page 22: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/22.jpg)
Part 24: Multiple Regression – Part 424-22/45
n1 = Number of predictors n2 = Sample size – number of predictors – 1
![Page 23: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/23.jpg)
Part 24: Multiple Regression – Part 424-23/45
Movie Madness RegressionS = 0.952237 R-Sq = 57.0% R-Sq(adj) = 56.6%
Analysis of Variance
Source DF SS MS F PRegression 20 2617.58 130.88 144.34 0.000Residual Error 2177 1974.01 0.91Total 2197 4591.58
![Page 24: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/24.jpg)
Part 24: Multiple Regression – Part 424-24/45
Compare Sample F to Critical F
F = 144.34 for Movie Madness
Critical value from the table is 1.57.
Reject the hypothesis of no relationship.
![Page 25: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/25.jpg)
Part 24: Multiple Regression – Part 424-25/45
An Equivalent Approach What is the “P Value?” We observed an F of 144.34 (or, whatever it is). If there really were no relationship, how likely is
it that we would have observed an F this large (or larger)? Depends on N and K The probability is reported with the
regression results as the P Value.
![Page 26: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/26.jpg)
Part 24: Multiple Regression – Part 424-26/45
The F TestS = 0.952237 R-Sq = 57.0% R-Sq(adj) = 56.6%
Analysis of Variance
Source DF SS MS F PRegression 20 2617.58 130.88 144.34 0.000Residual Error 2177 1974.01 0.91Total 2197 4591.58
![Page 27: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/27.jpg)
Part 24: Multiple Regression – Part 424-27/45
A Cost “Function” Regression
The regression is “significant.” F is huge. Which variables are significant? Which variables are not significant?
![Page 28: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/28.jpg)
Part 24: Multiple Regression – Part 424-28/45
What About a Group of Variables?
Is Genre significant in the movie model? There are 12 genre variables Some are “significant” (fantasy, mystery,
horror) some are not. Can we conclude the group as a whole is?
Maybe. We need a test.
![Page 29: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/29.jpg)
Part 24: Multiple Regression – Part 424-29/45
Theory for the Test A larger model has a higher R2 than a smaller
one. (Larger model means it has all the variables in
the smaller one, plus some additional ones) Compute this statistic with a calculator
2 2Larger Model Smaller Model
2Larger Model
R RHow much larger = How many Variables
F1 R
N K 1 for the larger model
![Page 30: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/30.jpg)
Part 24: Multiple Regression – Part 424-30/45
Is Genre Significant?Calc -> Probability Distributions -> F…
The critical value shown by Minitab is 1.76
With the 12 Genre indicator variables:R-Squared = 57.0%Without the 12 Genre indicator variables:R-Squared = 55.4%The F statistic is 6.750.F is greater than the critical value.Reject the hypothesis that all the genre coefficients are zero.
(0.570 0.554) / 12F 6.750(1 .570) / (2198 20 1)
![Page 31: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/31.jpg)
Part 24: Multiple Regression – Part 424-31/45
Now What? If the value that Minitab shows you is
less than your F statistic, then your F statistic is large
I.e., conclude that the group of coefficients is “significant”
This means that at least one is nonzero, not that all necessarily are.
![Page 32: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/32.jpg)
Part 24: Multiple Regression – Part 424-32/45
Application: Part of a Regression Model Regression model includes variables x1, x2,
… I am sure of these variables. Maybe variables z1, z2,… I am not sure of
these. Model: y = α+β1x1+β2x2 + δ1z1+δ2z2 + ε Hypothesis: δ1=0 and δ2=0. Strategy: Start with model including x1 and
x2. Compute R2. Compute new model that also includes z1 and z2.
Rejection region: R2 increases a lot.
![Page 33: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/33.jpg)
Part 24: Multiple Regression – Part 424-33/45
Test Statistic
2 20
2 2 2 21 1 0
Model 0 contains x1, x2, ...Model 1 contains x1, x2, ... and additional variables z1, z2, ...
R = the R from Model 0
R = the R from Model 1. R will always be greater than R .
The test statisti2 21 0
21
(R R ) /(Number of z variables)c is F =
(1 - R ) /(N - total number of variables - 1)Critical F comes from the table of F[KZ, N - KX - KZ - 1].(Unfortunately, Minitab cannot do this kind of test aut
omatically.)
![Page 34: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/34.jpg)
Part 24: Multiple Regression – Part 424-34/45
Gasoline Market
![Page 35: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/35.jpg)
Part 24: Multiple Regression – Part 424-35/45
Gasoline MarketRegression Analysis: logG versus logIncome, logPG The regression equation islogG = - 0.468 + 0.966 logIncome - 0.169 logPGPredictor Coef SE Coef T PConstant -0.46772 0.08649 -5.41 0.000logIncome 0.96595 0.07529 12.83 0.000logPG -0.16949 0.03865 -4.38 0.000S = 0.0614287 R-Sq = 93.6% R-Sq(adj) = 93.4%Analysis of VarianceSource DF SS MS F PRegression 2 2.7237 1.3618 360.90 0.000Residual Error 49 0.1849 0.0038Total 51 2.9086
R2 = 2.7237/2.9086 = 0.93643
![Page 36: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/36.jpg)
Part 24: Multiple Regression – Part 424-36/45
Gasoline MarketRegression Analysis: logG versus logIncome, logPG, ...
The regression equation islogG = - 0.558 + 1.29 logIncome - 0.0280 logPG - 0.156 logPNC + 0.029 logPUC - 0.183 logPPTPredictor Coef SE Coef T PConstant -0.5579 0.5808 -0.96 0.342logIncome 1.2861 0.1457 8.83 0.000logPG -0.02797 0.04338 -0.64 0.522logPNC -0.1558 0.2100 -0.74 0.462logPUC 0.0285 0.1020 0.28 0.781logPPT -0.1828 0.1191 -1.54 0.132S = 0.0499953 R-Sq = 96.0% R-Sq(adj) = 95.6%Analysis of VarianceSource DF SS MS F PRegression 5 2.79360 0.55872 223.53 0.000Residual Error 46 0.11498 0.00250Total 51 2.90858
Now, R2 = 2.7936/2.90858 = 0.96047 Previously, R2 = 2.7237/2.90858 = 0.93643
![Page 37: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/37.jpg)
Part 24: Multiple Regression – Part 424-37/45
2R increased from 0.93643 to 0.96047when the 3 variables were added to the model.
(0.96047 - 0.93643)/3The F statistic is = 9.32482(1 - 0.96047)/(52 - 2 - 1 - 3)
![Page 38: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/38.jpg)
Part 24: Multiple Regression – Part 424-38/45
n1 = Number of predictors n2 = Sample size – number of predictors – 1
![Page 39: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/39.jpg)
Part 24: Multiple Regression – Part 424-39/45
Improvement in R2
2R increased from 0.93643 to 0.96047(0.96047 - 0.93643)/3The F statistic is = 9.32482
(1 - 0.96047)/(52 - 2 - 3 - 1)
Inverse Cumulative Distribution Function
F distribution with 3 DF in numerator and 46 DF in denominator
P( X <= x ) = 0.95 x = 2.80684
The null hypothesis is rejected.Notice that none of the three individual variables are “significant” but the three of them together are.
![Page 40: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/40.jpg)
Part 24: Multiple Regression – Part 424-40/45
Application Health satisfaction depends on many factors:
Age, Income, Children, Education, Marital Status Do these factors figure differently in a model for
women compared to one for men? Investigation: Multiple regression Null hypothesis: The regressions are the same. Rejection Region: Estimated regressions that are
very different.
![Page 41: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/41.jpg)
Part 24: Multiple Regression – Part 424-41/45
Equal Regressions
Setting: Two groups of observations (men/women, countries, two different periods, firms, etc.)
Regression Model: y = α+β1x1+β2x2 + … + ε
Hypothesis: The same model applies to both groups
Rejection region: Large values of F
![Page 42: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/42.jpg)
Part 24: Multiple Regression – Part 424-42/45
Procedure: Equal Regressions There are N1 observations in Group 1 and N2 in Group 2. There are K variables and the constant term in the model. This test requires you to compute three regressions and retain the sum of squared
residuals from each: SS1 = sum of squares from N1 observations in group 1 SS2 = sum of squares from N2 observations in group 2 SSALL = sum of squares from NALL=N1+N2 observations when the two groups
are pooled.
The hypothesis of equal regressions is rejected if F is larger than the critical value from the F table (K numerator and NALL-2K-2 denominator degrees of freedom)
(SSALL-SS1-SS2)/KF=(SS1+SS2)/(N1+N2-2K-2)
![Page 43: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/43.jpg)
Part 24: Multiple Regression – Part 424-43/45
+--------+--------------+----------------+--------+--------+----------+|Variable| Coefficient | Standard Error | T |P value]| Mean of X|+--------+--------------+----------------+--------+--------+----------+ Women===|=[NW = 13083]================================================ Constant| 7.05393353 .16608124 42.473 .0000 1.0000000 AGE | -.03902304 .00205786 -18.963 .0000 44.4759612 EDUC | .09171404 .01004869 9.127 .0000 10.8763811 HHNINC | .57391631 .11685639 4.911 .0000 .34449514 HHKIDS | .12048802 .04732176 2.546 .0109 .39157686 MARRIED | .09769266 .04961634 1.969 .0490 .75150959 Men=====|=[NM = 14243]================================================ Constant| 7.75524549 .12282189 63.142 .0000 1.0000000 AGE | -.04825978 .00186912 -25.820 .0000 42.6528119 EDUC | .07298478 .00785826 9.288 .0000 11.7286996 HHNINC | .73218094 .11046623 6.628 .0000 .35905406 HHKIDS | .14868970 .04313251 3.447 .0006 .41297479 MARRIED | .06171039 .05134870 1.202 .2294 .76514779 Both====|=[NALL = 27326]============================================== Constant| 7.43623310 .09821909 75.711 .0000 1.0000000 AGE | -.04440130 .00134963 -32.899 .0000 43.5256898 EDUC | .08405505 .00609020 13.802 .0000 11.3206310 HHNINC | .64217661 .08004124 8.023 .0000 .35208362 HHKIDS | .12315329 .03153428 3.905 .0001 .40273000 MARRIED | .07220008 .03511670 2.056 .0398 .75861817
German survey data over 7 years, 1984 to 1991 (with a gap). 27,326 observations on Health Satisfaction and several covariates.
Health Satisfaction Models: Men vs. Women
![Page 44: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/44.jpg)
Part 24: Multiple Regression – Part 424-44/45
Computing the F Statistic+--------------------------------------------------------------------------------+| Women Men All || HEALTH Mean = 6.634172 6.924362 6.785662 || Standard deviation = 2.329513 2.251479 2.293725 || Number of observs. = 13083 14243 27326 || Model size Parameters = 6 6 6 || Degrees of freedom = 13077 14237 27320 || Residuals Sum of squares = 66677.66 66705.75 133585.3 || Standard error of e = 2.258063 2.164574 2.211256 || Fit R-squared = 0.060762 0.076033 .070786 || Model test F (P value) = 169.20(.000) 234.31(.000) 416.24 (.0000) |+--------------------------------------------------------------------------------+
[133,585.3-(66,677.66+66,705.75)] / 6F= = 6.8904(66,677.66+66,705.75) / (27,326 - 6 - 6 - 2
The critical value for F[6, 23214] is 2.0989Even though the regressions look similar, the hypothesis ofequal regressions is rejected.
![Page 45: Statistics and Data Analysis](https://reader035.vdocuments.us/reader035/viewer/2022062810/56815b54550346895dc93445/html5/thumbnails/45.jpg)
Part 24: Multiple Regression – Part 424-45/45
Summary Simple regression: Test β = 0 Testing about individual coefficients
in a multiple regression R2 as the fit measure in a multiple
regression Testing R2 = 0 Testing about sets of coefficients Testing whether two groups have the
same model