Professional Documents
Culture Documents
Chapter 5 Analysis of Variance SPSS - Analysis of Variance: How To Get There: Analyze Compare Means
Chapter 5 Analysis of Variance SPSS - Analysis of Variance: How To Get There: Analyze Compare Means
Compare Means
One-way ANOVA
To test the null hypothesis that several population means are equal, based on the results of several
independent samples. The test variable is measured on an interval- or ratio scale (for example age), and is
grouped by a variable which can be measured on a nominal or discrete ordinal scale (for example life
existing of the categories Dull, Routine and Exciting).
An independent T test and a one-way ANOVA for two independent samples test the same hypothesis.
You must select the dependent variable, and specify the factor to define the different groups. You can move
more than one variable into the Dependent List to test all of them. See following figure.
Button Options
Here you can choose to get descriptives of the data (Descriptive), and to test for equal variances in the
groups (Homogeneity-of-variance).
We performed a one-way ANOVA, with age as dependent variable, and life as factor, which exists of the
groups: 0 = Not applicable 1 = Dull
2 = Routine 3 = Exciting 8 = Dont know
Oneway
Descriptives
Age of Respondent
Dull
Routine
Exciting
Total
Mean
52,62
47,28
44,54
46,33
65
457
471
993
Std. Deviation
20,059
18,191
16,106
17,479
Std. Error
2,488
,851
,742
,555
Minimum
19
19
18
18
Maximum
89
89
87
89
df1
2
df2
990
Sig.
,000
ANOVA
Age of Respondent
Between Groups
Within Groups
Total
Sum of
Squares
4492,439
298568,2
303060,6
df
2
990
992
Mean Square
2246,220
301,584
F
7,448
Sig.
,001
However, we are dealing with large samples, which reduces the problem, and the Bonferroni test can be
used and interpreted with care.
In the table ANOVA the variation (Sum Of Squares), the degrees of freedom (df), and the variance (Mean
Square) are given for the within and the between groups, as well as the F value (F) and the significance of
the F (Sig.). Sig. indicates whether the null hypothesis the population means are all equal has to be
rejected or not.
As you can see, there is much difference between the two Mean Squares (2246,220 and 301,584), resulting
in a significant difference (F = 7,448; Sig. = 0,001). This means that H0 must be rejected. Thus: the average
age of people who find life Dull, Routine, or Exciting are not all equal.
But we dont know yet which means differ from each other. Rejecting a null-hypothesis means that
NOT ALL population means differ. We dont know yet whether one or more means vary from each other!
Therefore, we perform the Bonferoni procedure. The output shows us:
Mean
Difference
(I-J)
5,34
8,08*
-5,34
2,74*
-8,08*
-2,74*
Std. Error
2,302
2,298
2,302
1,140
2,298
1,140
Sig.
,062
,001
,062
,049
,001
,049
The table Multiple Comparisons shows that two out of three groups vary:
Dull vs Routine Sig. = 0,062 which is higher than the Sig. level of 0.05. These groups do not vary.
Dull vs Exciting Sig. = 0,001 which is lower than the Sig. level of 0.05. These groups vary.
Exciting vs Routine
Sig. = 0,049 which is lower than the Sig. level of 0.05. These groups vary (although
only just).
Because the Bonferroni test assumes equal variances, which does not hold in this case, you can do a test
that does not assume equal variances, for example the Tamhanes T2 test. We you perform this test, you
will see that in this case it results in the same conclusions.
A General Linear Model is, as the name suggest, general in that it incorporates many different models, so
that many different tests can be performed. Among these, are the one- and two-way ANOVA, and
regression analyses. More difficult designs can be analysed as well.
Univariate
A univariate GLM is a test with only one dependent variable. There can be one or more independent
variable or factors and/or variables.
A one-way ANOVA is a univariate GLM with exactly one independent variable (e.g. fixed factor).
A two-way ANOVA is a univariate GLM with exactly two independent variables (e.g. fixed factors).
You can test null hypotheses about the effects of other variables on the means of various groupings of a
single dependent variable. You can investigate interactions between factors as well as the effects of
individual factors. Also, the effects of covariates and covariate interactions with factors can be included.
In a univariate GLM you must select in the source variable list the variable you want to test and move it
into the Dependent Variable Box. You can select only one dependent variable. Then, select the variables
(factors) whose values define the groups and move them into the Fixed Factor box (or Random Factor(s)
box if appropriate). If you have covariates, you move them into the Covariate(s) box. To obtain the default
univariate GLM that contains all main effects and interactions, click OK. See following figure.
Button Plots
To obtain plots so that you can examine interaction visually. In the Factors list box, you see all main factors
in your model. To plot the means for the values of a singe factor, move that factor into the Horizontal Axis
box and click Add. If you want to see the means for all combinations of values of two factors, move the
second factor into the Separate Lines Box and click Add. The horizontal-axis factor means will be plotted
separately for each value of the second factor. If you move a third factor into the Separate Plots box,
separate plots for each value of this factor will be produced. See following figure.
0
1
2
Respondent's
Sex
3
4
1
2
Value Label
Less than
HS
N
80
High school
480
Junior
college
Bachelor
Graduate
Male
Female
67
181
92
450
450
df
9
1
4
1
4
890
900
899
Mean Square
2467,925
918896,468
1852,278
9296,790
137,948
191,260
F
12,904
4804,440
9,685
48,608
,721
Sig.
,000
,000
,000
,000
,577
Profile Plots
Estimated Marginal Means of Number of Hours Worked Last Week
60
50
40
Respondent's Sex
Male
30
Less than HS High school Junior college Bachelor
Female
Graduate
RS Highest Degree
The Profile Plot is a line plot for the average hours worked. Is there an interaction between gender and
education? You see that the two lines dont cross. The shapes of the lines for males and females are quite
similar. That suggests that there is no interaction. You dont expect pairs of lines drawn from real data to
have exactly the same shape, even if there is no interaction between gender and education in the population.
Your goal is to determine whether the interaction observed in the sample is large enough to believe that it
also exists in the population.
The table Between-Subjects Factors speaks for itself.
The table Tests of Between-Subjects Effects is very similar to the one-way analysis of variance table
ANOVA. What has changed is the number of hypotheses you are testing. In one-way analysis of variance,
you tested a single hypotheses. Now you can test three hypotheses: one about the main effect of degree,
one about the main effect of gender, and one about the degree-by-gender interaction.
Mean Square for Degree = the variability of the sample means of the five degree groups
Mean Square for Gender = the variability of the sample means of the two gender groups
Error Mean Square = the variability of the observations within the 10 cell means, that is 5 (Degree) x 2
(Gender). It is a kind of Within Groups Mean Square.
Remember: If the null hypotheses for an effect is true, then the corresponding F ratio is expected to be 1.
You look at the observed significance level for each observed F ratio to see if you can reject the
corresponding null hypotheses.
Always, you first have to look at possible interaction effect, since it doesnt make sense to talk about main
effects if there is significant interaction between the factors.
In this example, Sig. = 0.577, so you validate the null hypotheses that there is no interaction between the
two variables. The effect of the type of degree on hours worked seems to be similar for males and females.
The absence of interaction tells you that its reasonable to believe that the difference in average hours
worked between males and females is the same for all degree categories.
Since you didnt find an interaction between degree and gender, it makes sense to test hypotheses about the
main effects of degree and gender.
The F statistic for the degree main effect is 9,685. The observed significance level is 0,000. This means that
H0 must be rejected. The variable Degree has influence on the average hours worked. A posthoc test will
reveal more about the differences in degree.
The F statistic for the Gender main effect is 48,608. The observed significance level is 0,000. This means
that H0 must be rejected. The variable Gender has influence on the average hours worked.