Professional Documents
Culture Documents
MANOVA
MANOVA
Analysis of Variance
What is the impact of ethnicity (IV) (Hispanic, AfricanAmerican, Asian-Pacific Islander, Caucasian, etc) on annual
salary (DV)?
What is the impact of three different methods of meeting a
potential mate (IV) (online dating service; speed dating; setup
by friends) on likelihood of a second date (DV)
The first estimate of the common variance 2 is called the between (or
among) estimate and it involves the variance of the IV category
means about the grand mean
The second is called the within estimate, which will be a weighted
average of the variances within each of the IV categories. This is an
unbiased estimate of 2
The ANOVA test, called the F test, involves comparing the between
estimate to the within estimate
If the null hypothesis, that the population means on the DV for the
levels of the IV are equal to one another, is true, then the ratio of
the between to the within estimate of 2 should be equal to one
(that is, the between and within estimates should be the same)
If the null hypothesis is false, and the population means are not
equal, then the F ratio will be significantly greater than unity
(one).
The IV,
fathers
highest
degree
Source
Corrected Model
Intercept
PADEG
Error
Total
Corrected Total
df
4
1
4
1148
1153
1152
Mean Square
7447.871
1006433.085
7447.871
333.502
F
22.332
3017.774
22.332
Sig.
.000
.000
.000
Partial Eta
Squared
.072
.724
.072
Noncent.
Parameter
89.329
3017.774
89.329
Observed
a
Power
1.000
1.000
1.000
Some of the things that we learned to look for on the ANOVA output:
A. The value of the F ratio (same line as the IV or factor)
B. The significance of that F ratio (same line)
C. The partial eta squared (an estimate of the amount of the effect
size attributable to between-group differences (differences in levels of
the IV (ranges from 0 to 1 where 1 is strongest)
D. The power to detect the effect (ranges from 0 to 1 where 1 is
strongest)
df1
df2
2
Sig.
1.000
Tests of Hypotheses:
(1) There is no significant main effect for education level (F(2, 58) = 1.685, p = .
194, partial eta squared = .055) (red dots)
(2) There is no significant main effect for marital status (F (1, 58) = .441, p = .
509, partial eta squared = .008)(green dots)
(3) There is a significant interaction effect of marital status and education level (F
(2, 58) = 3.586, p = .034, partial eta squared = .110) (blue dots)
7
6
5
4
MarriedorNot
Married/Partner
NotMarried/Partner
HighSchool
CollegeorNot
SomePostHigh
CollegeorMore
Assumptions of MANOVA
1. Multivariate normality
Putting this in English, what you dont want is the case where if
your IV, was, for example, ethnicity, all the people in the other
category had scores on their 6 dependent variables clustered very
tightly around their mean, whereas people in the white category
had scores on the vector of 6 dependent variables clustered very
loosely around the mean. You dont want a leptokurtic set of
distributions for one level of the IV and a platykurtic set for another
level
If Boxs M is significant, it means you have violated an assumption
of MANOVA. This is not much of a problem if you have equal cell
sizes and large N; it is a much bigger issue with small sample sizes
and/or unequal cell sizes (in factorial anova if there are unequal cell
sizes the sums of squares for the three sources (two main effects
and interaction effect) wont add up to the Total SS)
3. Independence of observations
MANOVA Example
South
Midwest
MY1
MY1
MY2
My3
MY2
My3
Vectors of means
on the three DVs
(Y1, Y2, Y3) for
Regions South and
Midwest
REGION
Pillai's Trace
Wilks' Lambda
Hotelling's Trace
Roy's Largest Root
Pillai's Trace
Wilks' Lambda
Hotelling's Trace
Roy's Largest Root
Value
.984
.016
62.999
62.999
.620
.465
.971
.754
F
Hypothesis df
818.987b
3.000
818.987b
3.000
818.987b
3.000
b
818.987
3.000
3.562
9.000
3.900
9.000
4.062
9.000
c
10.299
3.000
Error df
39.000
39.000
39.000
39.000
123.000
95.066
113.000
41.000
Sig.
.000
.000
.000
.000
.001
.000
.000
.000
Partial Eta
Squared
.984
.984
.984
.984
.207
.225
.244
.430
Noncent.
Parameter
2456.960
2456.960
2456.960
2456.960
32.057
27.605
36.561
30.897
Observed
a
Power
1.000
1.000
1.000
1.000
.986
.964
.994
.997
First we will look at the overall F test (over all three dependent variables). What we
are most interested in is a statistic called Wilks lambda (), and the F value
associated with that. Lambda is a measure of the percent of variance in the DVs
that is *not explained* by differences in the level of the independent variable.
Lambda varies between 1 and zero, and we want it to be near zero (e.g, no variance
that is not explained by the IV). In the case of our IV, REGION, Wilks lambda is .
465, and has an associated F of 3.90, which is significant at p. <001. Lambda is the
ratio of W to T (Total SSCP matrix)
REGION
Pillai's Trace
Wilks' Lambda
Hotelling's Trace
Roy's Largest Root
Pillai's Trace
Wilks' Lambda
Hotelling's Trace
Roy's Largest Root
Value
.984
.016
62.999
62.999
.620
.465
.971
.754
F
Hypothesis df
818.987b
3.000
818.987b
3.000
818.987b
3.000
818.987b
3.000
3.562
9.000
3.900
9.000
4.062
9.000
10.299c
3.000
Error df
39.000
39.000
39.000
39.000
123.000
95.066
113.000
41.000
Sig.
.000
.000
.000
.000
.001
.000
.000
.000
Partial Eta
Squared
.984
.984
.984
.984
.207
.225
.244
.430
Noncent.
Parameter
2456.960
2456.960
2456.960
2456.960
32.057
27.605
36.561
30.897
Observed
a
Power
1.000
1.000
1.000
1.000
.986
.964
.994
.997
Box's M
F
df1
df2
Sig.
60.311
2.881
18
4805.078
.000
Checking out the Boxs M test we find that the test is significant (which means that there are
significant differences among the regions in the covariance matrices). If we had low power
that might be a problem, but we dont have low power. However, when Boxs test finds that
the covariance matrices are significantly different across levels of the IV that may indicate an
increased possibility of Type I error, so you might want to make a smaller error region. If you
redid the analysis with a confidence level of .001, you would still get a significant result, so
its probably OK. You should report the results of the Boxs M, though.
Above is a portion of the output table reporting the ANOVA tests on the three
dependent variables, abortions per 1000, divorces per 1000, and % Christian
adherents. Note that only the F values for %Christian adherents and Divorces per
1000 population are significant at your criterion of .017. (Note: the MANOVA
procedure doesnt seem to let you set different p levels for the overall test and the
univariate tests, so the power here is higher than it would be if you did these tests
separately in a ANOVA procedure and set p to .017 before you did the tests.)
2. Census region
Dependent Variable
Abortions per 1,000
women
Census region
Northeast
Midwest
South
West
Northeast
Midwest
South
West
Northeast
Midwest
South
West
Mean
23.333
14.136
17.229
18.118
53.389
60.182
55.921
43.718
3.600
3.745
4.964
5.591
Std. Error
3.188
2.884
2.556
2.884
3.907
3.534
3.133
3.534
.353
.319
.283
.319
F
Abortions per 1,000
women
Percent of pop who are
Christian adherents
Divorces per 1,000 pop
df1
df2
Sig.
1.068
41
.373
1.015
41
.396
1.641
41
.195
Tests the null hypothesis that the error variance of the dependent variable is
equal across groups.
a. Design: Intercept+REGION
Move region into the horizontal axis window and click the Add
button
Move hscat4 (HS educ) into the horizontal axis window and click
the Add button
Move region into the horizontal axis window and hscat 4 into the
separate lines window and click Add
Move hscat4 (HS educ) into the horizontal axis window and region
into the separate lines window and click Add, then click Continue
Box's M
F
df1
df2
Sig.
55.398
2.191
18
704.185
.003
Note that there are significant main effects for both region (green) and hscat4
(red) but not for their interaction (blue). Note the values of Wilks lambda;
only .237 of the variance is unexplained by region. Thats a very good result.
Boxs M is significant which is not so good but we do have high power. If you
redid the analysis with a lower significance level you would lose hscat4
We found that the only significant univariate main effect was for
the effect of region on unemployment rate. Now lets ask the
question, what are the differences between regions in
unemployment rate, considered two at a time?
What does the Levenes statistic say about the kind of post-hoc
test we can do with respect to the region variable?
According to the output, the group variances on unemployment
rate are not significantly different, so we can do a Sheff test
a
Levene's Test of Equality of Error Variances
F
Percent of workers who
are union members
Unemployment rate
percap income
df1
df2
Sig.
2.645
12
37
.012
1.281
2.573
12
12
37
37
.270
.014
Tests the null hypothesis that the error variance of the dependent variable is
equal across groups.
a. Design: Intercept+REGION+HSCAT4+REGION * HSCAT4
Dependent Variable
Percent of workers who
are union members
Unemployment rate
percap income
Census region
Northeast
Midwest
South
West
Northeast
Midwest
South
West
Northeast
Midwest
South
West
Mean
Std. Error
16.254
1.851
14.454a
1.656
9.447a
1.612
13.861a
1.584
5.108
.334
3.917a
.299
5.076a
.290
6.294a
.285
23822.583
1010.336
20624.016a
904.224
21051.631a
879.836
21386.655a
864.768
Lab # 9