man ova

33
MANOVA: Multivariate Analysis of Variance

Upload: adolf-velasco

Post on 06-Apr-2015

43 views

Category:

Documents


4 download

TRANSCRIPT

Page 1: Man Ova

MANOVA: Multivariate Analysis of Variance

Page 2: Man Ova

Review of ANOVA: Univariate Analysis of Variance• An univariate analysis of variance looks for the causal

impact of a nominal level independent variable (factor) on a single, interval or better level dependent variable

• The basic question you seek to answer is whether or not there is a difference in scores on the dependent variable attributable to membership in one or the other category of the independent variable• Analysis of Variance (ANOVA): Required when there are three

or more levels or conditions of the independent variable (but can be done when there are only two)• What is the impact of ethnicity (IV) (Hispanic, African-American,

Asian-Pacific Islander, Caucasian, etc) on annual salary (DV)?• What is the impact of three different methods of meeting a

potential mate (IV) (online dating service; speed dating; setup by friends) on likelihood of a second date (DV)

Page 3: Man Ova

Basic Analysis of Variance Concepts• We are going to make two estimates of the common population

variance, σ2

• The first estimate of the common variance σ2 is called the “between” (or “among”) estimate and it involves the variance of the IV category means about the grand mean

• The second is called the “within” estimate, which will be a weighted average of the variances within each of the IV categories. This is an unbiased estimate of σ2

• The ANOVA test, called the F test, involves comparing the between estimate to the within estimate

• If the null hypothesis, that the population means on the DV for the levels of the IV are equal to one another, is true, then the ratio of the between to the within estimate of σ2 should be equal to one (that is, the between and within estimates should be the same)

• If the null hypothesis is false, and the population means are not equal, then the F ratio will be significantly greater than unity (one).

Page 4: Man Ova

Basic ANOVA OutputTests of Between-Subjects Effects

Dependent Variable: Respondent Socioeconomic Index

29791.484b 4 7447.871 22.332 .000 .072 89.329 1.000

1006433.085 1 1006433.085 3017.774 .000 .724 3017.774 1.000

29791.484 4 7447.871 22.332 .000 .072 89.329 1.000

382860.051 1148 333.502

3073446.860 1153

412651.535 1152

SourceCorrected Model

Intercept

PADEG

Error

Total

Corrected Total

Type III Sumof Squares df Mean Square F Sig.

Partial EtaSquared

Noncent.Parameter

ObservedPower

a

Computed using alpha = .05a.

R Squared = .072 (Adjusted R Squared = .069)b.

Some of the things that we learned to look for on the ANOVA output:A. The value of the F ratio (same line as the IV or “factor”)B. The significance of that F ratio (same line)C. The partial eta squared (an estimate of the amount of the “effect size” attributable to between-group differences (differences in levels of the IV (ranges from 0 to 1 where 1 is strongest)D. The power to detect the effect (ranges from 0 to 1 where 1 is strongest)

The IV, father’s highest degree

A B C D

Page 5: Man Ova

More Review of ANOVA• Even if we have obtained a significant value of F and the overall

difference of means is significant, the F statistic isn’t telling us anything about how the mean scores varied among the levels of the IV.

• We can do some pairwise tests after the fact in which we compare the means of the levels of the IV

• The type of test we do depends on whether or not the variances of the groups (conditions or levels of the IV) are equal

• We test this using the Levene statistic. • If it is significant at p < .05 (group variances are significantly different)

we use an alternative post-hoc test like Tamhane• If it is not significant (groups variances are not significantly different)

we can use the Sheffé or similar test• In this example, variances are not significantly different (p > .05) so we

use the Sheffé testTest of Homogeneity of Variances

Self-disclosure

.000 2 9 1.000

LeveneStatistic df1 df2 Sig.

Page 6: Man Ova

Review of Factorial ANOVA Two-way ANOVA is applied to a situation in which you

have two independent nominal-level variables and one interval or better dependent variable

Each of the independent variables may have any number of levels or conditions (e.g., Treatment 1, Treatment 2, Treatment 3…… No Treatment)

In a two-way ANOVA you will obtain 3 F ratios One of these will tell you if your first independent

variable has a significant main effect on the DV A second will tell you if your second independent

variable has a significant main effect on the DV The third will tell you if the interaction of the two

independent variables has a significant effect on the DV, that is, if the impact of one IV depends on the level of the other

Page 7: Man Ova

Review: Factorial ANOVA Example

Tests of Hypotheses: (1) There is no significant main effect for education level (F(2, 58) = 1.685, p

= .194, partial eta squared = .055) (red dots)(2) There is no significant main effect for marital status (F (1, 58) = .441, p = .509,

partial eta squared = .008)(green dots)(3) There is a significant interaction effect of marital status and education level (F

(2, 58) = 3.586, p = .034, partial eta squared = .110) (blue dots)

Page 8: Man Ova

Plots of Interaction Effects

Estimated Marginal Means of TIMENET

CollegeorNot

CollegeorMoreSomePostHighHighSchool

Est

ima

ted

Ma

rgin

al M

ea

ns

9

8

7

6

5

4

3

2

MarriedorNot

Married/Partner

NotMarried/Partner

Education Level is plotted along the horizontal axis and hours spent on the net is plotted along the vertical axis. The red and green lines show how marital status interacts with education level. Here we note that spending time on the Internet is strongest among the Post High School group for single people, but lowest among this group for married people

Page 9: Man Ova

MANOVA: What Kinds of Hypotheses Can it Test? A MANOVA or multivariate analysis of variance is a

way to test the hypothesis that one or more independent variables, or factors, have an effect on a set of two or more dependent variables For example, you might wish to test the hypothesis

that sex and ethnicity interact to influence a set of job-related outcomes including attitudes toward co-workers, attitudes toward supervisors, feelings of belonging in the work environment, and identification with the corporate culture

As another example, you might want to test the hypothesis that three different methods of teaching writing result in significant differences in ratings of student creativity, student acquisition of grammar, and assessments of writing quality by an independent panel of judges

Page 10: Man Ova

Why Should You Do a MANOVA? You do a MANOVA instead of a series of one-at-a-time

ANOVAs for two main reasons Supposedly to reduce the experiment-wise level of Type I

error (8 F tests at .05 each means the experiment-wise probability of making a Type I error (rejecting the null hypothesis when it is in fact true) is 40%! The so-called overall test or omnibus test protects against this inflated error probability only when the null hypothesis is true. If you follow up a significant multivariate test with a bunch of ANOVAs on the individual variables without adjusting the error rates for the individual tests, there’s no “protection”

Another reasons to do MANOVA. None of the individual ANOVAs may produce a significant main effect on the DV, but in combination they might, which suggests that the variables are more meaningful taken together than considered separately

MANOVA takes into account the intercorrelations among the DVs

Page 11: Man Ova

Assumptions of MANOVA 1. Multivariate normality

All of the DVs must be distributed normally (can visualize this with histograms; tests are available for checking this out)

Any linear combination of the DVs must be distributed normally Check out pairwise relationships among the DVs for

nonlinear relationships using scatter plots All subsets of the variables must have a multivariate

normal distribution These requirements are rarely if ever tested in practice MANOVA is assumed to be a robust test that can stand up to

departures from multivariate normality in terms of Type I error rate

Statistical power (power to detect a main or interaction effect) may be reduced when distributions are very plateau-like (platykurtic)

Page 12: Man Ova

Assumptions of MANOVA, cont’d 2. Homogeneity of the covariance matrices

In ANOVA we talked about the need for the variances of the dependent variable to be equal across levels of the independent variable In MANOVA, the univariate requirement of equal variances

has to hold for each one of the dependent variables In MANOVA we extend this concept and require that the

“covariance matrices” be homogeneous Computations in MANOVA require the use of matrix algebra,

and each person’s “score” on the dependent variables is actually a “vector” of scores on DV1, DV2, DV3, …. DVn

The matrices of the covariances-the variance shared between any two variables-have to be equal across all levels of the independent variable

Page 13: Man Ova

Assumptions of MANOVA, cont’d This homogeneity assumption is tested with a test that is similar to

Levene’s test for the ANOVA case. It is called Box’s M, and it works the same way: it tests the hypothesis that the covariance matrices of the dependent variables are significantly different across levels of the independent variable Putting this in English, what you don’t want is the case where if

your IV, was, for example, ethnicity, all the people in the “other” category had scores on their 6 dependent variables clustered very tightly around their mean, whereas people in the “white” category had scores on the vector of 6 dependent variables clustered very loosely around the mean. You don’t want a leptokurtic set of distributions for one level of the IV and a platykurtic set for another level

If Box’s M is significant, it means you have violated an assumption of MANOVA. This is not much of a problem if you have equal cell sizes and large N; it is a much bigger issue with small sample sizes and/or unequal cell sizes (in factorial anova if there are unequal cell sizes the sums of squares for the three sources (two main effects and interaction effect) won’t add up to the Total SS)

Page 14: Man Ova

Assumptions of MANOVA, cont’d

3. Independence of observations Subjects’ scores on the dependent measures should not be

influenced by or related to scores of other subjects in the condition or level

Can be tested with an intraclass correlation coefficient if lack of independence of observations is suspected

Page 15: Man Ova

MANOVA Example Let’s test the hypothesis that region of the

country (IV) has a significant impact on three DVs, Percent of people who are Christian adherents, Divorces per 1000 population, and Abortions per 1000 populations. The hypothesis is that there will be a significant multivariate main effect for region. Another way to put this is that the vectors of means for the three DVs are different among regions of the country

This is done with the General Linear Model/ Multivariate procedure in SPSS (we will look first at an example where the analysis has already been done)

Computations are done using matrix algebra to find the ratio of the variability of B (Between-Groups sums of squares and cross-products (SSCP) matrix) to that of the W (Within-Groups SSCP matrix)

MY1

MY2

My3

Vectors of means on the three DVs (Y1, Y2, Y3) for Regions South and Midwest

MY1

MY2

My3

South Midwest

Page 16: Man Ova

MANOVA test of Our Hypothesis

First we will look at the overall F test (over all three dependent variables). What we are most interested in is a statistic called Wilks’ lambda (λ), and the F value associated with that. Lambda is a measure of the percent of variance in the DVs that is *not explained* by differences in the level of the independent variable. Lambda varies between 1 and zero, and we want it to be near zero (e.g, no variance that is not explained by the IV). In the case of our IV, REGION, Wilks’ lambda is .465, and has an associated F of 3.90, which is significant at p. <001. Lambda is the ratio of W to T (Total SSCP matrix)

Multivariate Testsd

.984 818.987b 3.000 39.000 .000 .984 2456.960 1.000

.016 818.987b 3.000 39.000 .000 .984 2456.960 1.000

62.999 818.987b 3.000 39.000 .000 .984 2456.960 1.000

62.999 818.987b 3.000 39.000 .000 .984 2456.960 1.000

.620 3.562 9.000 123.000 .001 .207 32.057 .986

.465 3.900 9.000 95.066 .000 .225 27.605 .964

.971 4.062 9.000 113.000 .000 .244 36.561 .994

.754 10.299c 3.000 41.000 .000 .430 30.897 .997

Pillai's Trace

Wilks' Lambda

Hotelling's Trace

Roy's Largest Root

Pillai's Trace

Wilks' Lambda

Hotelling's Trace

Roy's Largest Root

EffectIntercept

REGION

Value F Hypothesis df Error df Sig.Partial EtaSquared

Noncent.Parameter

ObservedPower

a

Computed using alpha = .05a.

Exact statisticb.

The statistic is an upper bound on F that yields a lower bound on the significance level.c.

Design: Intercept+REGIONd.

Page 17: Man Ova

MANOVA Test of our Hypothesis, cont’d

Continuing to examine our output, we find that the partial eta squared associated with the main effect of region is .225 and the power to detect the main effect is .964. These are very good results!

We would write this up in the following way: “A one-way MANOVA revealed a significant multivariate main effect for region, Wilks’ λ = .465, F (9, 95.066) = 3.9, p <. 001, partial eta squared = .225. Power to detect the effect was .964. Thus hypothesis 1 was confirmed.”

Multivariate Testsd

.984 818.987b 3.000 39.000 .000 .984 2456.960 1.000

.016 818.987b 3.000 39.000 .000 .984 2456.960 1.000

62.999 818.987b 3.000 39.000 .000 .984 2456.960 1.000

62.999 818.987b 3.000 39.000 .000 .984 2456.960 1.000

.620 3.562 9.000 123.000 .001 .207 32.057 .986

.465 3.900 9.000 95.066 .000 .225 27.605 .964

.971 4.062 9.000 113.000 .000 .244 36.561 .994

.754 10.299c 3.000 41.000 .000 .430 30.897 .997

Pillai's Trace

Wilks' Lambda

Hotelling's Trace

Roy's Largest Root

Pillai's Trace

Wilks' Lambda

Hotelling's Trace

Roy's Largest Root

EffectIntercept

REGION

Value F Hypothesis df Error df Sig.Partial EtaSquared

Noncent.Parameter

ObservedPower

a

Computed using alpha = .05a.

Exact statisticb.

The statistic is an upper bound on F that yields a lower bound on the significance level.c.

Design: Intercept+REGIONd.

Page 18: Man Ova

Box’s Test of Equality of Covariance Matrices

Box's Test of Equality of Covariance Matricesa

60.311

2.881

18

4805.078

.000

Box's M

F

df1

df2

Sig.

Tests the null hypothesis that the observed covariancematrices of the dependent variables are equal across groups.

Design: Intercept+REGIONa.

Checking out the Box’s M test we find that the test is significant (which means that there are significant differences among the regions in the covariance matrices). If we had low power that might be a problem, but we don’t have low power. However, when Box’s test finds that the covariance matrices are significantly different across levels of the IV that may indicate an increased possibility of Type I error, so you might want to make a smaller error region. If you redid the analysis with a confidence level of .001, you would still get a significant result, so it’s probably OK. You should report the results of the Box’s M, though.

Page 19: Man Ova

Looking at the Individual Dependent Variables If the overall F test is significant, then it’s common

practice to go ahead and look at the individual dependent variables with separate ANOVA tests The experimentwise alpha protection provided by the

overall or omnibus F test does not extend to the univariate tests. You should divide your confidence levels by the number of tests you intend to perform, so in this case if you expect to look at F tests for the three dependent variables you should require that p < .017 (.05/3)

This procedure ignores the fact the variables may be intercorrelated and that the separate ANOVAS do not take these intercorrelations into account You could get three significant F ratios but if the

variables are highly correlated you’re basically getting the same result over and over

Page 20: Man Ova

Univariate ANOVA tests of Three Dependent Variables

Above is a portion of the output table reporting the ANOVA tests on the three dependent variables, abortions per 1000, divorces per 1000, and % Christian adherents. Note that only the F values for %Christian adherents and Divorces per 1000 population are significant at your criterion of .017. (Note: the MANOVA procedure doesn’t seem to let you set different p levels for the overall test and the univariate tests, so the power here is higher than it would be if you did these tests separately in a ANOVA procedure and set p to .017 before you did the tests.)

Page 21: Man Ova

Writing up More of Your Results So far you have written the following:

“A one-way MANOVA revealed a significant multivariate main effect for region, Wilks’ λ = .465, F (9, 95.066) = 3.9, p <. 001, partial eta squared = .225. Power to detect the effect was .964. Thus hypothesis 1 was confirmed.”

You continue to write: “Given the significance of the overall test, the univariate

main effects were examined. Significant univariate main effects for region were obtained for percentage of Christian adherents, F (3, 41 ) = 3.944, p <.015 , partial eta square =.224, power = .794 ; and number of divorces per 1000 population, F (3,41 ) = 8.789 , p <.001 , partial eta square = .391, power = .991”

Page 22: Man Ova

Finally, Post-hoc Comparisons with Sheffé Test for the DVs that had Significant Univariate ANOVAs

Levene's Test of Equality of Error Variancesa

1.068 3 41 .373

1.015 3 41 .396

1.641 3 41 .195

Abortions per 1,000women

Percent of pop who areChristian adherents

Divorces per 1,000 pop

F df1 df2 Sig.

Tests the null hypothesis that the error variance of the dependent variable isequal across groups.

Design: Intercept+REGIONa.

The Levene’s statistics for the two DVs that had significant univariate ANOVAs are all non-significant, meaning that the group variances were equal, so you can use the Sheffé tests for comparing pairwise group means, e.g., do the South and the West differ significantly on % of Christian adherents and number of divorces.

2. Census region

23.333 3.188 16.895 29.772

14.136 2.884 8.312 19.960

17.229 2.556 12.066 22.391

18.118 2.884 12.294 23.942

53.389 3.907 45.498 61.280

60.182 3.534 53.044 67.320

55.921 3.133 49.594 62.248

43.718 3.534 36.580 50.856

3.600 .353 2.887 4.313

3.745 .319 3.101 4.390

4.964 .283 4.393 5.536

5.591 .319 4.946 6.236

Census regionNortheast

Midwest

South

West

Northeast

Midwest

South

West

Northeast

Midwest

South

West

Dependent VariableAbortions per 1,000women

Percent of pop who areChristian adherents

Divorces per 1,000 pop

Mean Std. Error Lower Bound Upper Bound

95% Confidence Interval

Page 23: Man Ova

Significant Pairwise Regional Differences on the Two Significant DVs

You might want to set your confidence level cutoff even lower since you are going to be doing 12 tests here (4(3)/2) for each variable

Page 24: Man Ova

Writing up All of Your MANOVA Results Your final paragraph will look like this “A one-way MANOVA revealed a significant multivariate main effect

for region, Wilks’ λ = .465, F (9, 95.066) = 3.9, p <. 001, partial eta squared = .225. Power to detect the effect was .964. Thus Hypothesis 1 was confirmed. Given the significance of the overall test, the univariate main effects were examined. Significant univariate main effects for region were obtained for percentage of Christian adherents, F (3, 41 ) = 3.944, p <.015 , partial eta square =.224, power = .794 ; and number of divorces per 1000 population, F (3,41 ) =8.789 , p <.001 , partial eta square = .391, power = .991. Significant regional pairwise differences were obtained in number of divorces per 1000 population between the West and both the Northeast and Midwest. The mean number of divorces per 1000 population were 5.59 in the West, 3.6 in the Northeast, and 3.74 in the Midwest.” You can present the pairwise results and the MANOVA overall F results and univariate F results in separate tables

Page 25: Man Ova

Now You Try It! Go here to download the file

statelevelmodified.sav Let’s test the hypothesis that region of the

country and availability of an educated workforce have an impact on three dependent variables: % union members, per capita income, and unemployment rate

Although a test will be performed for an interaction between region and workforce education level, no specific effect is hypothesized

Go to SPSS Data Editor

Page 26: Man Ova

Running a MANOVA in SPSS Go to Analyze/General Linear Model/ Multivariate Move Census Region and HS Educ into the Fixed Factors

category (this is where the IVs go) Move per capita income, unemployment rate, and % of

workers who are union members into the Dependent Variables category

Under Plots, create four plots, one for each of the two main effects (region, HS educ) and two for their interaction. Use the Add button to add each new plot Move region into the horizontal axis window and click the Add

button Move hscat4 (HS educ) into the horizontal axis window and

click the Add button Move region into the horizontal axis window and hscat 4 into

the separate lines window and click Add Move hscat4 (HS educ) into the horizontal axis window and

region into the separate lines window and click Add, then click Continue

Page 27: Man Ova

Setting up MANOVA in SPSS

Under Options, move all of the factors including the interactions into the Display Means for window

Select descriptive statistics, estimates of effect size, observed power, and homogeneity tests

Set the confidence level to .05 and click continue

Click OK Compare your output to the next several

slides

Page 28: Man Ova

Box's Test of Equality of Covariance Matricesa

55.398

2.191

18

704.185

.003

Box's M

F

df1

df2

Sig.

Tests the null hypothesis that the observed covariancematrices of the dependent variables are equal across groups.

Design: Intercept+REGION+HSCAT4+REGION *HSCAT4

a.

MANOVA Main and Interaction Effects

Note that there are significant main effects for both region (green) and hscat4 (red) but not for their interaction (blue). Note the values of Wilks’ lambda; only .237 of the variance is unexplained by region. That’s a very good result. Box’s M is significant which is not so good but we do have high power. If you redid the analysis with a lower significance level you would “lose” hscat4

Page 29: Man Ova

Univariate Tests: ANOVAs on each of the Three DVs for Region, HS Educ

Since we have obtained a significant multivariate main effect for each factor, we can go ahead and do the univariate F tests where we look at each DV in turn to see if the two IVS have a significant impact on them separately. Since we are doing six tests here we are going to reguire an experiment-wise alpha rate of .05, so we will divide it by six to get an acceptable confidence level for each of the six tests, so we will set the alpha level to p < .008. By that criterion, the only significant univariate result is for the effect of region on unemployment rate. With a more lenient criterion of .05 (and a greater probability of Type I error), three other univariate tests would have been significant

Page 30: Man Ova

Pairwise Comparisons on the Significant Univariate Tests We found that the only significant univariate main effect was for

the effect of region on unemployment rate. Now let’s ask the question, what are the differences between regions in unemployment rate, considered two at a time?

What does the Levene’s statistic say about the kind of post-hoc test we can do with respect to the region variable?

According to the output, the group variances on unemployment rate are not significantly different, so we can do a Sheffé test

Levene's Test of Equality of Error Variancesa

2.645 12 37 .012

1.281 12 37 .270

2.573 12 37 .014

Percent of workers whoare union members

Unemployment rate

percap income

F df1 df2 Sig.

Tests the null hypothesis that the error variance of the dependent variable isequal across groups.

Design: Intercept+REGION+HSCAT4+REGION * HSCAT4a.

Page 31: Man Ova

Pairwise Difference of Means

Since we are doing 6 significance tests (K(k-1)/2) looking at the pairwise tests comparing the employment rate by region, we can use the smaller confidence level again to protect against inflated alpha error, so let’s divide the .05 by 6 and set .008 as our error level. By this standard, the South and Midwest and the West and Midwest are significantly different in unemployment rate.

Page 32: Man Ova

Reporting the Differences

2. Census region

16.254 1.851 12.504 20.004

14.454a 1.656 11.098 17.810

9.447a 1.612 6.182 12.713

13.861a 1.584 10.651 17.070

5.108 .334 4.433 5.784

3.917a .299 3.312 4.521

5.076a .290 4.488 5.665

6.294a .285 5.716 6.872

23822.583 1010.336 21775.447 25869.719

20624.016a 904.224 18791.884 22456.147

21051.631a 879.836 19268.915 22834.348

21386.655a 864.768 19634.468 23138.841

Census regionNortheast

Midwest

South

West

Northeast

Midwest

South

West

Northeast

Midwest

South

West

Dependent VariablePercent of workers whoare union members

Unemployment rate

percap income

Mean Std. Error Lower Bound Upper Bound

95% Confidence Interval

Based on modified population marginal mean.a.

Significant mean differences in unemployment rate were obtained between the Midwest (M = 3.917) and the West (6.294) and Midwest and the South (M = 5.076)

Page 33: Man Ova

Lab # 9 Duplicate the preceding data analysis in SPSS. Write

up the results (the tests of the hypothesis about the main effects of region and HS Educ on the three dependent variables of per capita income, unemployment rate, and % union members, as if you were writing for publication. Put your paragraph in a Word document, and illustrate your results with tables from the output as appropriate (for example, the overall multivariate F table and the table of mean scores broken down by regions). You can also use plots to illustrate significant effects